AI Navigate

Probabilistic Federated Learning on Uncertain and Heterogeneous Data with Model Personalization

arXiv cs.LG / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Meta-BayFL addresses training degradation in federated learning caused by data uncertainty and non-IID heterogeneity by proposing a personalized probabilistic FL method that combines Bayesian neural networks with meta-learning.
  • It uses BNN-based client models that incorporate uncertainty across hidden layers to stabilize training on small and noisy local datasets.
  • It introduces meta-learning with adaptive learning rates to enable personalized updates under non-IID data, improving local training.
  • It presents a unified probabilistic and personalized design that enhances robustness of global model aggregation and provides a theoretical convergence analysis with an upper bound on the global model over communication rounds.
  • In experiments on CIFAR-10, CIFAR-100, and Tiny-ImageNet, Meta-BayFL outperforms state-of-the-art methods (e.g., pFedMe, Ditto, FedFomo) with up to 7.42% higher test accuracy, and discusses runtime, latency, and communication costs for edge/IoT deployment.

Abstract

Conventional federated learning (FL) frameworks often suffer from training degradation due to data uncertainty and heterogeneity across local clients. Probabilistic approaches such as Bayesian neural networks (BNNs) can mitigate this issue by explicitly modeling uncertainty, but they introduce additional runtime, latency, and bandwidth overhead that has rarely been studied in federated settings. To address these challenges, we propose Meta-BayFL, a personalized probabilistic FL method that combines meta-learning with BNNs to improve training under uncertain and heterogeneous data. The framework is characterized by three main features: (1) BNN-based client models incorporate uncertainty across hidden layers to stabilize training on small and noisy datasets, (2) meta-learning with adaptive learning rates enables personalized updates that enhance local training under non-IID conditions, and (3) a unified probabilistic and personalized design improves the robustness of global model aggregation. We provide a theoretical convergence analysis and characterize the upper bound of the global model over communication rounds. In addition, we evaluate computational costs (runtime, latency, and communication) and discuss the feasibility of deployment on resource-constrained devices such as edge nodes and IoT systems. Extensive experiments on CIFAR-10, CIFAR-100, and Tiny-ImageNet show that Meta-BayFL consistently outperforms state-of-the-art methods, including both standard and personalized FL approaches (e.g., pFedMe, Ditto, FedFomo), with up to 7.42\% higher test accuracy.