Journal of Natural Language Processing
Online ISSN : 2185-8314
Print ISSN : 1340-7619
ISSN-L : 1340-7619
Precise Language Models Requiring No Lexical Knowledge
HIROKI MORIHIROTOMO AsoSHOZO MAKINO
Author information
JOURNAL FREE ACCESS

1999 Volume 6 Issue 2 Pages 29-40

Details
Abstract
This paper proposes a novel, knowledge-free language model with great ability in reducing ambiguity. This model is defined as n-gram of string which is referred to “superword, ” and belongs to a superclass of traditional word or string n-gram models'class. The concept of superword is based on only one principle-repetitionality in training text. The probabilistic distribution of the model is learned through the forward-backward algorithm. Experimental results showed that the performance of superword model combined with character trigram model was superior to the traditional word model based on morphological analysis, and traditional string-based model.
Content from these authors
© The Association for Natural Language Processing
Previous article Next article
feedback
Top