Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
39th (2025)
Session ID : 3Win5-13
Conference information

Robust Offline-to-Online Reinforcement Learning against Perturbation in Joint Torque Signals
*Shingo AYABEHiroshi KERAKazuhiko KAWAMOTO
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

Offline reinforcement learning (RL) enables policy learning from pre-collected datasets without environmental interaction. This approach reduces the cost of data collection and mitigating safety risks in robotic control. However, real-world deployment requires robustness to control failures, which remain challenging due to the lack of exploration during training. To address this issue, we propose an offline-to-online RL method that improves robustness with minimal online fine-tuning. During fine-tuning, perturbations simulating control component failures are applied to joint torque signals, including random and adversarial perturbations. We conduct experiments using legged robot models in OpenAI Gym. The results demonstrate that offline RL does not improve robustness and remains highly vulnerable to perturbations. In contrast, our method significantly improves robustness against these perturbations.

Content from these authors
© 2025 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top