Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
34th (2020)
Session ID : 1Q4-GS-11-02
Conference information

Learning Alignment Tasks Based on Residual Reinforcement Learning with Multiple Expert Policies
*Kazuki YAGINUMATomoaki NAKAMURAYusuke KATOTakayuki NAGAIJun OZAWA
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

Reinforcement learning (RL) enables robots to flexibly learn skills from the interaction with the environment. However, robots move randomly to explore valuable actions at an early stage, which can be unsafe. Moreover, it takes a long time to learn motion from scratch. To learn more efficiently and safely, residual reinforcement learning (RRL) has been proposed. In RRL, skills are learned by correcting expert policy that can be obtained from an expert's demonstration. However, conventional RRL assumes a single expert policy, whereas we consider multiple policies for more complex tasks. In this paper, we propose RRL with multiple expert policies, where a selection of a suitable expert policy in the current state is also learned based on RL. Experimentally, we show that the agent can learn more accurate skills in the object alignment task.

Content from these authors
© 2020 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top