Host: The Japanese Society for Artificial Intelligence
Name : The 36th Annual Conference of the Japanese Society for Artificial Intelligence
Number : 36
Location : [in Japanese]
Date : June 14, 2022 - June 17, 2022
This paper compares three cross-lingual sentiment analysis methods using various Bidirectional Encoder Representations from Transformers (BERT) models to predict product review ratings for low-resource languages. Sentiment analysis involves the task of predicting review ratings. Cross-lingual approaches have been studied because the prediction accuracy tends to be low due to the lack of training data for low-resource languages. Meanwhile, recently, the approaches that fine-tune pre-trained models attract attention. In particular, BERT achieved high performances for various tasks. Therefore, we compared cross-lingual methods that use various BERT with a large amount of English data and a small amount of Japanese data to improve the prediction accuracy of Japanese review ratings. We compared three methods: method A, which uses English BERT fine-tuned with English reviews and Japanese test data translated into English, method B, which uses Japanese BERT fine-tuned with English reviews translated into Japanese, and method C, which uses English-Japanese parallel data to fine-tune the multi-lingual BERT. The translation was conducted using machine translation. Compared to the baseline model, which uses Japanese BERT fine-tuned with only a small amount of Japanese data, we found that method B improved performance, while methods A and C lost performances.