Stability of Sequential and Parallel Coordinate Ascent Variational Inference

arXiv stat.ML / 3/24/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper compares sequential versus parallel coordinate ascent variational inference and shows they can behave very differently in practice.
  • It studies a moderately high-dimensional linear regression setting to isolate how each algorithm’s update scheme affects convergence.
  • The authors find that the sequential variant, while often slower, converges under more relaxed conditions than the parallel variant.
  • The parallel variant is commonly used for block-wise updates and computational efficiency, but its convergence guarantees are tighter/more restrictive in the paper’s analysis.
  • The work extends known ideas from numerical analysis to the variational inference optimization literature, where such differences were previously underexplored.

Abstract

We highlight a striking difference in behavior between two widely used variants of coordinate ascent variational inference: the sequential and parallel algorithms. While such differences were known in the numerical analysis literature in simpler settings, they remain largely unexplored in the optimization-focused literature on variational inference in more complex models. Focusing on the moderately high-dimensional linear regression problem, we show that the sequential algorithm, although typically slower, enjoys convergence guarantees under more relaxed conditions than the parallel variant, which is often employed to facilitate block-wise updates and improve computational efficiency.