Journal of the Robotics Society of Japan
Online ISSN : 1884-7145
Print ISSN : 0289-1824
ISSN-L : 0289-1824
Paper
Reinforcement Learning by Using Dual Q-table for the Whole Space and a Partial Space
Hirokazu MatsuiChieko NishizawaYoshihiko Nomura
Author information
JOURNAL FREE ACCESS

2019 Volume 37 Issue 7 Pages 620-631

Details
Abstract

In this paper, we propose a Q-learning method by using dual Q-table. Concretely, the proposed method has two Q-tables: “Whole Q-table'' is larger, based on the whole space (in enough detail for learning optimal actions) of the environment and “Partial Q-table'' is smaller, based on a subspace (rough for learning outline actions) of the whole space. The two Q-tables simultaneously learn the environment based on a selected action. The action is selected by using the more learned Q-table out of the two Q-tables by each step. We simulated the proposed method, comparing with conventional ones, under three conditions of the learning environments, where the partial Q-table can learn optimal actions at the highest rate, middle rate and the lowest rate of the situations in the environment. As a result, we indicated that the proposed method can learn the optimal actions at any rates. As the rate is higher, it converges earlier. Even if at the lowest rate, the proposed method is almost as effective as conventional one. And we indicated the proposed method to be effective by using mathematical analysis. Furthermore, we verified that the proposed method was effective under an actual environment.

Content from these authors
© 2018 The Robotics Society of Japan
Previous article Next article
feedback
Top