Journal of the Japanese Society for Artificial Intelligence
Online ISSN : 2435-8614
Print ISSN : 2188-2266
Print ISSN:0912-8085 until 2013
Propagating Learned Behaviors from a Virtual Agent to a Physical Robot in Exploitation-Oriented Reinforcement Learning
Tomohiro YAMAGUCHIMoto'omi MASUBUCHIYasuhiro TANAKAMasahiko YACHIDA
Author information
MAGAZINE FREE ACCESS

1997 Volume 12 Issue 4 Pages 570-581

Details
Abstract

For a physical robot to acquire behaviors, it is important for it to learn in the physical environment. Since reinforcement learning requires large computation costs as well as a lot of time in the physical environment, most research has performed learning by simulation. However, this does not work well in the real world. Realizing reinforcement learning of a physical robot in a physical environment requires both an adaptation for the diversity of possible situations and a high-speed learning method that can learn from fewer trials. This paper describes cooperative reinforcement learning based on propagating the learned behaviors of a virtual agent to a physical robot in order to accelerate learning in a physical environment. The method consists of two parts: (1) preparation learning in a virtual environment to accelerate initial learning, which accounts for most of the learning cost ; and, (2) refinement learning in a physical environment by using the virtual learning results as an initial behavior set of a physical robot. Experimental results are given for a ball-pushing task with the physical robot and a virtual agent.

Content from these authors
© 1997 The Japaense Society for Artificial Intelligence
Previous article Next article
feedback
Top