Host: The Japanese Society for Artificial Intelligence
Name : The 38th Annual Conference of the Japanese Society for Artificial Intelligence
Number : 38
Location : [in Japanese]
Date : May 28, 2024 - May 31, 2024
In recent years, the development of Large Language Models (LLMs) has rapidly progressed, playing a significant role in Natural Language Processing (NLP). However, there is currently no established standard for efficiently evaluating these LLMs, which often generate complex sentences. Existing evaluation methods using trained Language Models (LMs) are popular due to their cost-effectiveness, but they often fall short in accuracy when training data is scarce. I propose a data augmentation method using ChatGPT to improve the accuracy of LMs in situations of data scarcity. Results on the Japanese Question Answering (QA) task demonstrate that an LM, trained using questions and answers generated by the proposed method, surpassed ChatGPT3.5 and achieved 92% of the evaluation performance of ChatGPT4, even in scenarios where only documents were available.