Merge and Conquer: Instructing Multilingual Models by Adding Target Language Weights

arXiv cs.CL / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses how to improve instruction-following and language performance of LLMs in low-resource languages that are typically underrepresented in English-centric models.
  • It proposes using model merging to transfer language knowledge by combining an instruction-tuned LLM with a language-specific base model, avoiding the need for new language-specific instruction datasets and repeated fine-tuning.
  • Experiments on Basque, Catalan, Galician, and Spanish across two model families show that merging can produce effective instruction-following in newly targeted languages.
  • The authors also demonstrate that merging multiple language-specific models can yield multilingual capability, suggesting a scalable way to compose strengths across languages.
  • Overall, the work concludes that model merging can be a computationally efficient alternative to continual pre-training for low-resource language adaptation while maintaining competitive results.

Abstract

Large Language Models (LLMs) remain heavily centered on English, with limited performance in low-resource languages. Existing adaptation approaches, such as continual pre-training, demand significant computational resources. In the case of instructed models, high-quality instruction data is also required, both of which are often inaccessible for low-resource language communities. Under these constraints, model merging offers a lightweight alternative, but its potential in low-resource contexts has not been systematically explored. In this work, we explore whether it is possible to transfer language knowledge to an instruction-tuned LLM by merging it with a language-specific base model, thereby eliminating the need of language-specific instructions and repeated fine-tuning processes whenever stronger instructed variants become available. Through experiments covering four Iberian languages (Basque, Catalan, Galician, and Spanish) and two model families, we show that merging enables effective instruction following behavior in new languages and even supports multilingual capability through the combination of multiple language-specific models. Our results indicate that model merging is a viable and efficient alternative to traditional adaptation methods for low-resource languages, achieving competitive performance while greatly reducing computational cost.

Merge and Conquer: Instructing Multilingual Models by Adding Target Language Weights | AI Navigate