Language on Demand, Knowledge at Core: Composing LLMs with Encoder-Decoder Translation Models for Extensible Multilinguality

arXiv cs.CL / 4/1/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current LLMs hold substantial cross-lingual knowledge in a shared semantic space, but reliably using it for low-resource or unseen languages is still a major weakness.
  • It proposes XBridge, a compositional encoder–LLM–decoder architecture that uses pretrained translation models to handle multilingual understanding and generation while keeping the LLM as an English-centric reasoning core.
  • To fix representation misalignment between the LLM and translation models, XBridge adds lightweight cross-model mapping layers plus an optimal-transport-based alignment objective for semantic consistency.
  • Experiments across four LLMs on multiple tasks (multilingual understanding, reasoning, summarization, and generation) show XBridge improves over strong baselines, with the biggest gains for low-resource and previously unseen languages, and does not require retraining the LLM.
  • The work suggests a scalable pathway for extending LLM multilinguality by composing them with translation systems rather than treating multilingual capability as a monolithic model property.

Abstract

Large language models (LLMs) exhibit strong general intelligence, yet their multilingual performance remains highly imbalanced. Although LLMs encode substantial cross-lingual knowledge in a unified semantic space, they often struggle to reliably interface this knowledge with low-resource or unseen languages. Fortunately, pretrained encoder-decoder translation models already possess balanced multilingual capability, suggesting a natural complement to LLMs. In this work, we propose XBridge, a compositional encoder-LLM-decoder architecture that offloads multilingual understanding and generation to external pretrained translation models, while preserving the LLM as an English-centric core for general knowledge processing. To address the resulting representation misalignment across models, we introduce lightweight cross-model mapping layers and an optimal transport-based alignment objective, enabling fine-grained semantic consistency for multilingual generation. Experiments on four LLMs across multilingual understanding, reasoning, summarization, and generation indicate that XBridge outperforms strong baselines, especially on low-resource and previously unseen languages, without retraining the LLM.