Host: The Japanese Society for Artificial Intelligence
Name : The 35th Annual Conference of the Japanese Society for Artificial Intelligence
Number : 35
Location : [in Japanese]
Date : June 08, 2021 - June 11, 2021
In this study, we apply the method of knowledge distillation to the Japanese morphological analyzerrakkyoand evaluate if the method compresses its model size, and the training converges for smaller datasets. Recently,Japanese morphological analyzers have achieved high performance in both accuracy and speed. From the viewpointof practical uses, however, it is preferable to reduce the model size. The rakkyo model, among others, succeeded insignificantly reducing its model size by using only character unigrams and discard the dictionary, by the training onsilver data of 500 million sentences generated by Juman++. We tried to further compress rakkyo by constructinga neural morphological analyzer for Japanese using the outputs of rakkyo, namely the probabilistic distributions astraining data. The evaluation is done against the silver data generated by rakkyo, which suggests that our modelapproaches the accuracy of rakkyo with a smaller amount of data.