AI Navigate

CA-HFP: Curvature-Aware Heterogeneous Federated Pruning with Model Reconstruction

arXiv cs.AI / 3/16/2026

💬 OpinionDeveloper Stack & InfrastructureModels & Research

Key Points

  • CA-HFP enables each client in federated learning to perform device-specific structured pruning guided by a curvature-informed significance score, enabling heterogeneous pruning while maintaining aggregation compatibility.
  • Each client reconstructs its compact submodel back into a shared global parameter space via a lightweight reconstruction step to preserve cross-client interoperability.
  • The authors derive a convergence bound for federated optimization with multiple local SGD steps that accounts for local computation, data heterogeneity, and pruning-induced perturbations, leading to a principled loss-based pruning criterion.
  • Experiments on FMNIST, CIFAR-10, and CIFAR-100 with VGG and ResNet architectures show CA-HFP preserves accuracy while reducing per-client computation and communication, outperforming standard federated training and existing pruning baselines.

Abstract

Federated learning on heterogeneous edge devices requires personalized compression while preserving aggregation compatibility and stable convergence. We present Curvature-Aware Heterogeneous Federated Pruning (CA-HFP), a practical framework that enables each client perform structured, device-specific pruning guided by a curvature-informed significance score, and subsequently maps its compact submodel back into a common global parameter space via a lightweight reconstruction. We derive a convergence bound for federated optimization with multiple local SGD steps that explicitly accounts for local computation, data heterogeneity, and pruning-induced perturbations; from which a principled loss-based pruning criterion is derived. Extensive experiments on FMNIST, CIFAR-10, and CIFAR-100 using VGG and ResNet architectures under varying degrees of data heterogeneity demonstrate that CA-HFP preserves model accuracy while significantly reducing per-client computation and communication costs, outperforming standard federated training and existing pruning-based baselines.