Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
37th (2023)
Session ID : 3M1-GS-10-02
Conference information

Scandalous Article Classification with Contrastive Learning BERT and Study of Sentence Embedded Representation
*Yuichiro TAKASUSeiichi OZAWATakehide HIROSEYoshihiro IKEDANoriyasu NAKAGAWAMasaaki IIZUKADaisuke NISHIDA
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

This research reports on an attempt to determine whether an economic article deals with a scandal or not, attributed to a binary classification problem. Since scandals can have a tremendous impact on the management of a company or entity, it is absolutely crucial to detect reported articles as early as possible, and overlooking them is absolutely unacceptable. This requires a high recall rate. In this study, we attempted to improve the recall rate by using a deep learning model called SimCSE, which is anisotropic in the sentence space of BERT, to suppress the oversight of scandalous articles. The results of experiments using Reuters articles showed that BERT with SimCSE applied improved the recall rate compared to BERT without SimCSE. Improvement was also observed in the index of sentence space uniformity, suggesting that this isotropic space contributed to the improvement in recall. The high level of uniformity was also found to be inherited before and after fine tuning. Translated with www.DeepL.com/Translator (free version)

Content from these authors
© 2023 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top