Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
38th (2024)
Session ID : 3Xin2-53
Conference information

Evaluation of Instruction Tuning on Finance-Specific Large Language Models
*Masatsugu YAMADAToshiya IMOTO
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

It is beginning to be reported that small language models specialized for specific domains exceed the performance of general-purpose large language models. However, open-source language models specialized for the financial domain are limited, and language models need more evaluation with sufficient performance. Therefore, in this paper, we used benchmark sets containing various financial domain tasks such as sentiment analysis, classification, and question answering, and evaluated the performance change of a small chat model when subjected to multiple conditions of instructional tuning. We trained 7B and 13B models for this task by fine-tuning using low-rank adaptation. We empirically found that each model tended to improve the performance with both continuous pre-training and supervised fine-tuning despite over-fitting, and the generated results were affected by the instruction template.

Content from these authors
© 2024 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top