Journal of Natural Language Processing
Online ISSN : 2185-8314
Print ISSN : 1340-7619
ISSN-L : 1340-7619
General Paper (Peer-Reviewed)
Divide-and-Conquer Neural Machine Translation Using Intra-Sentence Context
Ryuta IshikawaYasumasa KanoKatsuhito SudohSatoshi Nakamura
Author information
JOURNAL FREE ACCESS

2025 Volume 32 Issue 1 Pages 114-133

Details
Abstract

Although neural machine translation (NMT) usually produces high-quality translation through flexible word choice and fluency, its quality can decrease for long input sentences. A divide-and-conquer approach to this problem exists that splits a long input sentence into shorter segments and merges their translations, resulting in limited improvement in NMT. In this study, we propose a novel divide-and-conquer method for NMT that improves the translation of long sentences in an intra-sentence context. The proposed method (1) splits a sentence around coordinating conjunctions, connecting clauses labeled S by syntactic parsing, (2) translates these clauses using a clause-level translation model that utilizes an intra-sentence context, and (3) merges clause-level translations using another sequence-to-sequence model to obtain a sentence-level translation. In our English-to-Japanese translation experiments on ASPEC using a pre-trained multilingual BART model, the proposed method outperformed a baseline multilingual BART-based NMT for input sentences with over 40 words.

Content from these authors
© 2025 The Association for Natural Language Processing
Previous article Next article
feedback
Top