主催: 人工知能学会
会議名: 第93回 言語・音声理解と対話処理研究会
回次: 93
開催地: オンライン
開催日: 2021/11/29 - 2021/11/30
p. 169-170
We have developed two types of pre-trained models, GPT-2 and RoBERTa, that are trained from a public corpus consisting of about 75-gigabyte texts. The models and its training code have been released under licenses that allow for commercial use. By fine-tuning the released models, users will be able to accomplish a variety of Japanese natural language processing tasks with high task accuracy.