Enhanced Privacy and Communication Efficiency in Non-IID Federated Learning with Adaptive Quantization and Differential Privacy

arXiv cs.CV / 4/28/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper tackles two core problems in non-IID federated learning: communication overhead from heterogeneous device connectivity and privacy leakage risks from model/gradient analysis.
  • It combines differential privacy with adaptive quantization, using Laplacian-based DP in particular to provide privacy guarantees and addressing a comparatively underexplored DP choice in FL.
  • The authors introduce both a global bit-length scheduler (round-based cosine annealing) and a client-based scheduler that adapts bit-length according to each client’s estimated contribution via dataset entropy.
  • Experiments on CIFAR-10, MNIST, and medical imaging datasets under non-IID settings show substantial reductions in total communicated data (up to 52.64% on MNIST and 45.06% on CIFAR-10) while keeping accuracy competitive and maintaining differential-privacy protections.

Abstract

Federated learning (FL) is a distributed machine learning method where multiple devices collaboratively train a model under the management of a central server without sharing underlying data. One of the key challenges of FL is the communication bottleneck caused by variations in connection speed and bandwidth across devices. Therefore, it is essential to reduce the size of transmitted data during training. Additionally, there is a potential risk of exposing sensitive information through the model or gradient analysis during training. To address both privacy and communication efficiency, we combine differential privacy (DP) and adaptive quantization methods. We use Laplacian-based DP to preserve privacy, which is relatively underexplored in FL and offers tighter privacy guarantees than Gaussian-based DP. We propose a simple and efficient global bit-length scheduler using round-based cosine annealing, along with a client-based scheduler that dynamically adapts based on client contribution estimated through dataset entropy analysis. We evaluate our approach through extensive experiments on CIFAR10, MNIST, and medical imaging datasets, using non-IID data distributions across varying client counts, bit-length schedulers, and privacy budgets. The results show that our adaptive quantization methods reduce total communicated data by up to 52.64% for MNIST, 45.06% for CIFAR10, and 31% to 37% for medical imaging datasets compared to 32-bit float training while maintaining competitive model accuracy and ensuring robust privacy through differential privacy.