人工知能学会第二種研究会資料
Online ISSN : 2436-5556
FinMegatron: Large Financial Domain Language Models
Xianchao WU
著者情報
研究報告書・技術報告書 フリー

2021 年 2021 巻 FIN-026 号 p. 22-

詳細
抄録

General domain pretrained large-scale language models, such as BERT and GPT3, have achieved state-of-the-art results among numerous NLP classification and generation applications. This pretraining technology is also willing to be used in vertical domains, such as finance. The downstream applications include financial event extraction from news, summarization, and causal inferencing. In this paper, we propose large-scale pretrained BERT models for financial domain in English and Japanese languages. The original datasets come from professional financial news. We empirically study the factors of sub-word vocabulary set, model size and their impacts to the downstream financial NLP applications. The code and pretrained models are released from https://github.com/NVIDIA/Megatron-LM.

著者関連情報
© 2021 Authors
前の記事 次の記事
feedback
Top