AI Navigate

DriftGuard: Mitigating Asynchronous Data Drift in Federated Learning

arXiv cs.LG / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses asynchronous data drift in federated learning where device distributions shift at different times, complicating training.
  • DriftGuard uses a Mixture-of-Experts inspired architecture that separates shared global parameters from local cluster-specific parameters to enable efficient adaptation.
  • It supports two retraining strategies: global retraining updates the shared parameters when system-wide drift is identified, and group retraining selectively updates local parameters for device clusters without sharing raw data.
  • Empirical results show comparable or better accuracy with up to 83% reduction in retraining cost and up to 2.3x higher accuracy per retraining unit.
  • The framework is open-source and available at https://github.com/blessonvar/DriftGuard.

Abstract

In real-world Federated Learning (FL) deployments, data distributions on devices that participate in training evolve over time. This leads to asynchronous data drift, where different devices shift at different times and toward different distributions. Mitigating such drift is challenging: frequent retraining incurs high computational cost on resource-constrained devices, while infrequent retraining degrades performance on drifting devices. We propose DriftGuard, a federated continual learning framework that efficiently adapts to asynchronous data drift. DriftGuard adopts a Mixture-of-Experts (MoE) inspired architecture that separates shared parameters, which capture globally transferable knowledge, from local parameters that adapt to group-specific distributions. This design enables two complementary retraining strategies: (i) global retraining, which updates the shared parameters when system-wide drift is identified, and (ii) group retraining, which selectively updates local parameters for clusters of devices identified via MoE gating patterns, without sharing raw data. Experiments across multiple datasets and models show that DriftGuard matches or exceeds state-of-the-art accuracy while reducing total retraining cost by up to 83%. As a result, it achieves the highest accuracy per unit retraining cost, improving over the strongest baseline by up to 2.3x. DriftGuard is available for download from https://github.com/blessonvar/DriftGuard.