Host: The Japanese Society for Artificial Intelligence
Name : The 38th Annual Conference of the Japanese Society for Artificial Intelligence
Number : 38
Location : [in Japanese]
Date : May 28, 2024 - May 31, 2024
Inaccurate responses, termed hallucinations, pose challenges in various Large Language Model (LLM) applications. Although a sampling-based method called SelfCheckGPT has been devised to detect hallucinations by using the model's input-output interface without external knowledge, the method requires an increase in the temperature parameter, which cannot be controlled in some LLM services, including ChatGPT. In LLM services designed for accurate responses, the temperature parameter is fixed at a low level, which can degrade the performance of SelfCheckGPT. We therefore propose a novel methodology that utilizes data augmentation (adding random strings or back-translation) during sampling to detect hallucinations in Japanese LLMs under the fixed-low-temperature constraint. Our experimental results reveal that the proposed methodology outperforms SelfCheckGPT under the fixed-low-temperature constraint.