Random Coordinate Descent on the Wasserstein Space of Probability Measures

arXiv stat.ML / 4/3/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies optimization on the space of probability measures using the Wasserstein-2 geometry, focusing on reducing the computational burden of full Wasserstein-gradient methods in hard high-dimensional or ill-conditioned settings.
  • It proposes two randomized coordinate descent frameworks on the Wasserstein manifold—RWCD for standard objectives and RWCP for composite objectives that fit proximal-gradient-type formulations.
  • The approach leverages coordinate-wise structure to better handle anisotropic objective landscapes where full-gradient optimization can be inefficient.
  • The authors provide convergence guarantees under multiple conditions, including non-convex, Polyak–Łojasiewicz, and geodesically convex regimes, and relate the results to known Euclidean coordinate descent behavior.
  • Numerical experiments on ill-conditioned energies suggest that the randomized coordinate methods can deliver substantial speedups versus conventional full-gradient techniques.

Abstract

Optimization over the space of probability measures endowed with the Wasserstein-2 geometry is central to modern machine learning and mean-field modeling. However, traditional methods relying on full Wasserstein gradients often suffer from high computational overhead in high-dimensional or ill-conditioned settings. We propose a randomized coordinate descent framework specifically designed for the Wasserstein manifold, introducing both Random Wasserstein Coordinate Descent (RWCD) and Random Wasserstein Coordinate Proximal{-Gradient} (RWCP) for composite objectives. By exploiting coordinate-wise structures, our methods adapt to anisotropic objective landscapes where full-gradient approaches typically struggle. We provide a rigorous convergence analysis across various landscape geometries, establishing guarantees under non-convex, Polyak-{\L}ojasiewicz, and geodesically convex conditions. Our theoretical results mirror the classic convergence properties found in Euclidean space, revealing a compelling symmetry between coordinate descent on vectors and on probability measures. The developed techniques are inherently adaptive to the Wasserstein geometry and offer a robust analytical template that can be extended to other optimization solvers within the space of measures. Numerical experiments on ill-conditioned energies demonstrate that our framework offers significant speedups over conventional full-gradient methods.