Differentially Private Clustered Federated Learning with Privacy-Preserving Initialization and Normality-Driven Aggregation

arXiv cs.LG / 4/23/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a key challenge in combining clustered federated learning (CFL) with differential privacy (DP), where DP noise makes client updates too noisy and prevents the server from initializing cluster centroids well.
  • It proposes PINA, a two-stage framework that has clients fine-tune a lightweight LoRA adapter and privately share compressed update sketches so the server can build robust cluster centroids.
  • In the second stage, PINA applies a normality-driven aggregation strategy to improve convergence and overall robustness while maintaining CFL’s benefits.
  • The authors report that PINA provides formal privacy guarantees even when the server is untrusted and improves accuracy over existing DP-FL approaches by an average of 2.9% for privacy budgets ε = 2 and 8.
  • Overall, the work targets cross-device, highly heterogeneous settings where vanilla FL struggles to converge and generalize due to non-IID data distributions.

Abstract

Federated learning (FL) enables training of a global model while keeping raw data on end-devices. Despite this, FL has shown to leak private user information and thus in practice, it is often coupled with methods such as differential privacy (DP) and secure vector sum to provide formal privacy guarantees to its participants. In realistic cross-device deployments, the data are highly heterogeneous, so vanilla federated learning converges slowly and generalizes poorly. Clustered federated learning (CFL) mitigates this by segregating users into clusters, leading to lower intra-cluster data heterogeneity. Nevertheless, coupling CFL with DP remains challenging: the injected DP noise makes individual client updates excessively noisy, and the server is unable to initialize cluster centroids with the less noisy aggregated updates. To address this challenge, we propose PINA, a two-stage framework that first lets each client fine-tune a lightweight low-rank adaptation (LoRA) adapter and privately share a compressed sketch of the update. The server leverages these sketches to construct robust cluster centroids. In the second stage, PINA introduces a normality-driven aggregation mechanism that improves convergence and robustness. Our method retains the benefits of clustered FL while providing formal privacy guarantees against an untrusted server. Extensive evaluations show that our proposed method outperforms state-of-the-art DP-FL algorithms by an average of 2.9% in accuracy for privacy budgets (epsilon in {2, 8}).