Host: The Japanese Society for Artificial Intelligence
Name : The 38th Annual Conference of the Japanese Society for Artificial Intelligence
Number : 38
Location : [in Japanese]
Date : May 28, 2024 - May 31, 2024
In the development of Large Language Models for Code (Code LLM), it has been found that instruction tuning is effective in enhancing the performance of Code LLM. Instruction tuning is a method that improves generalization performance by additional learning of instructions. However, there is a variety of opinions on what form of instruction is optimal, and it has not been clarified. The purpose of this study is to investigate the impact of different instruction formats on code generation performance in order to enhance the effects of instruction tuning for Code LLM. In particular, we focused on the output formats used for code extraction and conducted experiments. We also visualized the experimental results. The results revealed the performance differences in code generation by the models due to different output formats, and it was clarified that the Markdown format was the most versatile. Moreover, it was revealed that specifying an output format resulted in a higher accuracy rate than not specifying an output format.