主催: 一般社団法人 日本機械学会
会議名: ロボティクス・メカトロニクス 講演会2019
開催日: 2019/06/05 - 2019/06/08
Learning of the robot movement with reinforcement learning (RL) has attracted attention, and improvements of various RL methods have been carried out intensively. With conventional RL methods, however, a complicated task takes a long learning process, which is problematic in the robotics domain. In this paper, we focused on the compositionality of policies of Soft Q-learning (SoftQL). With SoftQL, it is possible to compose multiple already-learned policies and execute compound tasks efficiently. However, in the SoftQL, the action-sampling procedure and learning algorithm are complex due to the continuous action space. In this paper, we applied the SoftQL to a maze-solving problem which has discrete space and investigated its performance and computational tractability for discrete-space problems.