自然言語処理
Online ISSN : 2185-8314
Print ISSN : 1340-7619
ISSN-L : 1340-7619
一般論文(査読有)
Exploring how well Large-scale Masked Language Models can Recognize Grammatical Errors
Manabu KimuraRyo NagataKazuaki Hanawa
著者情報
ジャーナル フリー

2023 年 30 巻 2 号 p. 689-712

詳細
抄録

In this paper, we explore the capacity of a language model-based method for grammatical error detection in detail. We first show that 5 to 10% of training data are enough for a BERT-based error detection method to achieve performance equivalent to what a non-language model-based method can achieve with the full training data; recall improves much faster with respect to training data size in the BERT-based method than in the non-language model method. This suggests that (i) the BERT-based method should have a good knowledge of the grammar required to recognize certain types of error and that (ii) it can use the knowledge to estimate whether the given word is correct or erroneous after fine-tuning with few training samples, which explains its high generalization ability in grammatical error detection. We further show with pseudo error data that it actually exhibits such nice properties for recognizing various types of error. Finally, based on these findings, we discuss a cost-effective method for detecting grammatical errors with feedback comments to learners.

著者関連情報
© 2023 The Association for Natural Language Processing
前の記事 次の記事
feedback
Top