Journal of Natural Language Processing
Online ISSN : 2185-8314
Print ISSN : 1340-7619
ISSN-L : 1340-7619
General Paper (Peer-Reviewed)
Self-Ensemble: Integration of N-best Hypotheses by Lexically Constrained Decoding
Ryota MiyanoTomoyuki KajiwaraYuki Arase
Author information
JOURNAL FREE ACCESS

2024 Volume 31 Issue 2 Pages 680-706

Details
Abstract

In language generation, a reranking method improves the quality of generated sentences by re-scoring the top N hypotheses. The reranking method assumes the existence of higher quality hypotheses in N-best. We expand this assumption to be more practical as partly higher quality hypotheses exist in the N-best; however, they may be imperfect as the entire sentence. We propose a method for generating high-quality outputs by integrating high-quality fragments in the N-best. Specifically, we first obtain the N-best hypotheses and estimate the quality of each token. We then perform decoding again applying lexical constraints, with the words predicted to be wrong as negative constraints and those predicted to be correct as positive constraints. This method produces sentences that contain the correct words included in the N-best output and do not contain the wrong words. Empirical experiments on paraphrase generation, summarization, translation, and constrained text generation confirmed that our method outperformed strong N-best reranking methods in paraphrase generation and summarization tasks.

Content from these authors
© 2024 The Association for Natural Language Processing
Previous article Next article
feedback
Top