One Model to Translate Them All? A Journey to Mount Doom for Multilingual Model Merging
arXiv cs.CL / 4/6/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies weight-space model merging for multilingual machine translation, aiming to understand why merging strategies that work in multitask settings can fail across languages.
- Through full fine-tuning on large bilingual corpora and evaluation of standard merging methods, the authors find that merging typically degrades performance, with the drop being especially severe when target languages differ.
- The analysis uses neuron-selectivity measures (span-conditioned) and layer-wise centered kernel alignment to show that language-specific neurons are concentrated in embedding layers and upper transformer blocks, while intermediate layers stay comparatively shared.
- Fine-tuning is shown to redistribute rather than sharpen language selectivity, making supervised/related language neurons less exclusive and pushing unsupervised-language neurons to become more isolated.
- The resulting representational divergence in higher layers undermines the geometric assumptions that make weight-space merging effective, providing a mechanistic explanation for multilingual merging failure.
Related Articles
How Bash Command Safety Analysis Works in AI Systems
Dev.to
How to Get Better Output from AI Tools (Without Burning Time and Tokens)
Dev.to
How I Added LangChain4j Without Letting It Take Over My Spring Boot App
Dev.to
The Future of Artificial Intelligence in Everyday Life
Dev.to
Teaching Your AI to Read: Automating Document Triage for Investigators
Dev.to