Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
37th (2023)
Session ID : 1T3-GS-6-05
Conference information

Prompt Optimization for Training Generalizable Language Models
*Masaru ISONUMAJunichiro MORIIchiro SAKATA
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

Recently, instruction tuning has been attracting significant attention as a method for training generalizable language models (e.g., ChatGPT). Although various prompts have been manually created for instruction tuning, it has not been clarified what kind of prompts are optimal for obtaining cross-task generalization ability. This study presents \emph{instruction optimization}, which optimizes training prompts by leveraging bilevel optimization, and we clarify what kind of prompts are optimal for instruction tuning. Experimental results demonstrate that instruction optimization enhances the diversity of prompts and improves the generalization performance in a zero-shot setting, whereas using the same examples rather than a variety of exemplars is more effective in a few-shot setting.

Content from these authors
© 2023 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top