Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
38th (2024)
Session ID : 3O1-OS-16b-05
Conference information

Improving Learning Efficiency in Compositional Robot Tasks Using Prior Knowledge of Large Language Model
*Shota TAKASHIROTatsuya MATSUSHIMAYusuke IWASAWAYutaka MATSUO
Author information
Keywords: RL, LLM, IL
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

Large language model have shown high general performance in various tasks, and their applications are expanding not only in natural language processing but also in various other fields. Although there are many existing studies that utilize large language model in robot control, most of them are used for action planning in compositional tasks, and fail if an action is selected that is not prepared in advance by the robot. In other words, the a priori knowledge in large-scale language models can be used for policy selection during inference, but it cannot be used during actual policy learning. In this paper, we aim to decompose a task using prior knowledge from a large language model and intensively reinforce learning only the failed steps, so that the robot can acquire a new strategy with minimal interaction with the environment.

Content from these authors
© 2024 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top