Top-down string-to-dependency Neural Machine Translation

arXiv cs.CL / 3/31/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses weaknesses of attention-based encoder-decoder neural machine translation models when translating long inputs that are rare or absent in training data.
  • It proposes a syntactic decoder that outputs the target-language dependency tree using a top-down, left-to-right generation order rather than standard sequence-to-sequence decoding.
  • Experimental results indicate the top-down string-to-dependency-tree approach generalizes better for long, out-of-training-distribution inputs.
  • The core idea is to incorporate target-side syntax explicitly to mitigate length and coverage gaps in NMT training.

Abstract

Most of modern neural machine translation (NMT) models are based on an encoder-decoder framework with an attention mechanism. While they perform well on standard datasets, they can have trouble in translation of long inputs that are rare or unseen during training. Incorporating target syntax is one approach to dealing with such length-related problems. We propose a novel syntactic decoder that generates a target-language dependency tree in a top-down, left-to-right order. Experiments show that the proposed top-down string-to-tree decoding generalizes better than conventional sequence-to-sequence decoding in translating long inputs that are not observed in the training data.