AI Navigate

Few-for-Many Personalized Federated Learning

arXiv cs.AI / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Personalized Federated Learning (PFL) is framed as a multi-objective problem across M clients with highly heterogeneous data, highlighting the scalability challenge of maintaining M separate models.
  • The work reformulates PFL as a few-for-many optimization that uses only K shared server models (K << M) to serve all clients, with the approximation error diminishing as K grows and each client's model converging to its optimum as data increases.
  • FedFew is a gradient-based algorithm that jointly optimizes the K server models and automatically discovers diverse models without manual client partitioning or heavy hyperparameter tuning.
  • Experiments on vision, NLP, and real-world medical imaging show FedFew with just 3 models consistently outperforming state-of-the-art approaches, and the code is released on GitHub.

Abstract

Personalized Federated Learning (PFL) aims to train customized models for clients with highly heterogeneous data distributions while preserving data privacy. Existing approaches often rely on heuristics like clustering or model interpolation, which lack principled mechanisms for balancing heterogeneous client objectives. Serving M clients with distinct data distributions is inherently a multi-objective optimization problem, where achieving optimal personalization ideally requires M distinct models on the Pareto front. However, maintaining M separate models poses significant scalability challenges in federated settings with hundreds or thousands of clients. To address this challenge, we reformulate PFL as a few-for-many optimization problem that maintains only K shared server models (K \ll M) to collectively serve all M clients. We prove that this framework achieves near-optimal personalization: the approximation error diminishes as K increases and each client's model converges to each client's optimum as data grows. Building on this reformulation, we propose FedFew, a practical algorithm that jointly optimizes the K server models through efficient gradient-based updates. Unlike clustering-based approaches that require manual client partitioning or interpolation-based methods that demand careful hyperparameter tuning, FedFew automatically discovers the optimal model diversity through its optimization process. Experiments across vision, NLP, and real-world medical imaging datasets demonstrate that FedFew, with just 3 models, consistently outperforms other state-of-the-art approaches. Code is available at https://github.com/pgg3/FedFew.