Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
36th (2022)
Session ID : 2C5-GS-2-03
Conference information

Max-Min Off-Policy Actor-Critic with Robustness to Model Misspecification
*Takumi TANABERei SATOKazuto FUKUCHIJun SAKUMAYouhei AKIMOTO
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

In reinforcement learning, since it is costly and risky to training policies in the real-world, policies trained in a simulation environment are often transferred to the real-world. However, because the simulation environment does not perfectly mimic the real-world environment, modeling errors may occur. We focus on scenarios where a simulation environment including an uncertainty parameter and a set of its possible values are available. The objective is to optimize the worst-case performance on the uncertainty parameter set to guarantee the performance in the corresponding real-world environment, provided that it is included in the uncertainty parameter set. We propose the Max-min Twin Delayed Deep Deterministic Policy Gradient Algorithm (M2TD3) and its soft variant (SoftM2TD3) to solve the max-min optimization problem in order to obtain a policy that optimizes the worst-case performance. Experiments in the MuJoCo environments show that the proposed method exhibited better worst-case performance than some baseline approaches.

Content from these authors
© 2022 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top