Task-Centric Personalized Federated Fine-Tuning of Language Models

arXiv cs.AI / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses limitations of personalized federated learning for language models, focusing on robustness issues like generalization to unseen tasks and interference among multiple intra-client task distributions.
  • It proposes FedRouter, a task-centric personalized FL method that builds specialized models per task (rather than per client) using adapter-based personalization.
  • FedRouter employs two clustering mechanisms—local clustering to associate adapters with task samples and global clustering to match similar adapters across clients into task-centric personalized models.
  • An evaluation “router” selects the best adapter for each test sample by routing it according to the learned task clusters.
  • Experiments on a multitask dataset show FedRouter delivers notable gains, including up to 6.1% relative improvement under task interference and up to 136% relative improvement on generalization evaluations.

Abstract

Federated Learning (FL) has emerged as a promising technique for training language models on distributed and private datasets of diverse tasks. However, aggregating models trained on heterogeneous tasks often degrades the overall performance of individual clients. To address this issue, Personalized FL (pFL) aims to create models tailored for each client's data distribution. Although these approaches improve local performance, they usually lack robustness in two aspects: (i) generalization: when clients must make predictions on unseen tasks, or face changes in their data distributions, and (ii) intra-client tasks interference: when a single client's data contains multiple distributions that may interfere with each other during local training. To tackle these two challenges, we propose FedRouter, a clustering-based pFL that builds specialized models for each task rather than for each client. FedRouter uses adapters to personalize models by employing two clustering mechanisms to associate adapters with specific tasks. A local clustering that associate adapters with task data samples and a global one that associates similar adapters from different clients to construct task-centric personalized models. Additionally, we propose an evaluation router mechanism that routes test samples to the best adapter based on the created clusters. Experiments comparing our method with existing approaches across a multitask dataset, FedRouter demonstrate strong resilience in these challenging scenarios performing up to 6.1% relatively better under tasks interference and up to 136% relative improvement under generalization evaluation.