SubFLOT: Submodel Extraction for Efficient and Personalized Federated Learning via Optimal Transport

arXiv cs.LG / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • SubFLOT introduces a server-side personalized federated pruning framework that targets the tradeoff between non-personalized server pruning and expensive client-side pruning on edge devices.
  • It adds an Optimal Transport-enhanced Pruning (OTP) module that uses historical client models as proxies for local data distributions and formulates pruning as a Wasserstein distance minimization problem without needing access to raw data.
  • It proposes Scaling-based Adaptive Regularization (SAR) to reduce parametric divergence among heterogeneous submodels by adaptively penalizing deviations from the global model based on each client’s pruning rate.
  • Experiments reported in the paper indicate SubFLOT substantially outperforms prior state-of-the-art approaches, supporting its usefulness for efficient, personalized FL deployments under system and statistical heterogeneity.

Abstract

Federated Learning (FL) enables collaborative model training while preserving data privacy, but its practical deployment is hampered by system and statistical heterogeneity. While federated network pruning offers a path to mitigate these issues, existing methods face a critical dilemma: server-side pruning lacks personalization, whereas client-side pruning is computationally prohibitive for resource-constrained devices. Furthermore, the pruning process itself induces significant parametric divergence among heterogeneous submodels, destabilizing training and hindering global convergence. To address these challenges, we propose SubFLOT, a novel framework for server-side personalized federated pruning. SubFLOT introduces an Optimal Transport-enhanced Pruning (OTP) module that treats historical client models as proxies for local data distributions, formulating the pruning task as a Wasserstein distance minimization problem to generate customized submodels without accessing raw data. Concurrently, to counteract parametric divergence, our Scaling-based Adaptive Regularization (SAR) module adaptively penalizes a submodel's deviation from the global model, with the penalty's strength scaled by the client's pruning rate. Comprehensive experiments demonstrate that SubFLOT consistently and substantially outperforms state-of-the-art methods, underscoring its potential for deploying efficient and personalized models on resource-constrained edge devices.