Fast convergence of a Federated Expectation-Maximization Algorithm

arXiv stat.ML / 2026/3/24

💬 オピニオンIdeas & Deep AnalysisModels & Research

要点

  • The paper analyzes how data heterogeneity affects convergence by studying the Expectation-Maximization (EM) algorithm applied to a Federated Mixture of K Linear Regressions (FMLR) model.
  • It provides a complete characterization of EM’s convergence rate across different regimes of client counts and per-client data sizes, including partial asymptotic limits for the number of clients.
  • The authors show that when the signal-to-noise ratio (SNR) is at least on the order of sqrt(K) and EM is well-initialized, EM converges to the true ground truth across all considered regimes.
  • Experiments on synthetic data support the theory, finding that data heterogeneity may accelerate convergence rather than inherently slowing federated iterative algorithms.

Abstract

Data heterogeneity has been a long-standing bottleneck in studying the convergence rates of Federated Learning algorithms. In order to better understand the issue of data heterogeneity, we study the convergence rate of the Expectation-Maximization (EM) algorithm for the Federated Mixture of K Linear Regressions model (FMLR). We completely characterize the convergence rate of the EM algorithm under all regimes of number of clients and number of data points per client, with partial limits in the number of clients. We show that with a signal-to-noise-ratio (SNR) that is atleast of order \sqrt{K}, the well-initialized EM algorithm converges to the ground truth under all regimes. We perform experiments on synthetic data to illustrate our results. In line with our theoretical findings, the simulations show that rather than being a bottleneck, data heterogeneity can accelerate the convergence of iterative federated algorithms.