計測自動制御学会論文集
Online ISSN : 1883-8189
Print ISSN : 0453-4654
ISSN-L : 0453-4654
論文
学習進度に基づくマルチエージェントQ学習における競合回避
市川 嘉裕高玉 圭樹
著者情報
ジャーナル フリー

2012 年 48 巻 11 号 p. 764-772

詳細
抄録
This paper focuses on the conflict which occurs even in the cooperated task in multi-agent reinforcement learning domain and explores its avoidance methods from the viewpoint of the learning progress of agents. For this purpose, we propose the conflict avoidance method in multi-agent Q-learning which enables agents to learn according to the learning progress of other agents, and investigate its effectiveness in the lane crossing road problem where agents aim at going to their own directions but they cannot cross each other due to the narrow lane. The intensive simulations on the problem has the following implications: (1) the temporary decrease of the discount factor of the Q-learning agent with a high learning progress (i.e., the agents who learns well) contributes to avoiding conflict more successfully than the increase of the random degree of agent behaviors; (2) the increase of the discount factor of the Q-learning agent with a low learning progress (i.e., the agents who has not yet learned well) contributes to avoiding the conflict more successfully; and (3) a balance of learning progress among agents is critical to solve the conflict situation.
著者関連情報
© 2012 公益社団法人 計測自動制御学会
前の記事 次の記事
feedback
Top