Optimizing Multilingual LLMs via Federated Learning: A Study of Client Language Composition

arXiv cs.CL / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • This study extends the FederatedScope-LLM framework to run multilingual instruction-tuning experiments for LLMs using federated learning under heterogeneous client language distributions.
  • It proposes Local Dynamic Early Stopping (LDES-FL), a client-side validation-driven pause/resume mechanism meant to improve FL training efficiency and sustainability.
  • Experimental results show monolingual local fine-tuning is best for single-language specialization, while federated training is more suitable for learning a single balanced multilingual global model.
  • Increasing multilinguality within clients generally improves global model quality and fairness, reduces the performance gap versus centralized multilingual fine-tuning, and delivers the biggest benefits to lower-resource languages.
  • The gains from richer within-client multilinguality come with higher training cost, as the approach requires more optimization steps.

Abstract

Federated Learning (FL) of Large Language Models (LLMs) in multilingual environments presents significant challenges stemming from heterogeneous language distributions across clients and disparities in language resource availability. To address these challenges, we extended the FederatedScope-LLM framework to support multilingual instruction-tuning experiments with LLMs. We also introduced a novel client-specific early stopping mechanism, Local Dynamic Early Stopping (LDES-FL), which allows clients to pause and resume local training based on client-side validation performance, enhancing training efficiency and sustainability. Through a series of experiments, we studied how client language composition - from fully monolingual to increasingly multilingual clients - affects multilingual quality, fairness and training cost. Monolingual local fine-tuning remains the most effective for single-language specialization, whereas federated training is better suited to learning a single balanced multilingual model. In FL, increasing within-client multilinguality leads to stronger and fairer global models, narrows the gap to centralized multilingual fine-tuning, and yields the largest gains for lower-resource languages, albeit at the cost of more optimization steps. Overall, our results identify client language composition as a key design variable in multilingual FL, shaping performance, fairness and efficiency