Host: The Japanese Society for Artificial Intelligence
Name : The 37th Annual Conference of the Japanese Society for Artificial Intelligence
Number : 37
Location : [in Japanese]
Date : June 06, 2023 - June 09, 2023
One of the model-based reinforcement learning techniques, known as the world model, predicts the environment's transitions resulting from the agent's actions. Using the world model is expected to improve sample efficiency and enable adaptation to unseen tasks. However, the world model is larger compared to other reinforcement learning models, which raises concerns about prolonged training and computational constraints when executing the model. To address these issues, we propose enhancing the world model's practicality by applying model compression and transfer learning. The objective of this study is to investigate the effect of these approaches on the world model's performance. Based on our results, we draw two conclusions: (1) the proposed method (model compression + transfer learning) has the potential to outperform learning only the target task without applying any model compression or transfer learning, and (2) the proposed method is robust to changes in hyperparameters.