Distributionally Robust K-Means Clustering

arXiv stat.ML / 4/14/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses K-means’ brittleness to outliers, distribution shift, and small samples by reframing it as Lloyd–Max quantization of the empirical distribution.
  • It introduces a distributionally robust K-means objective by assuming the true population distribution lies within a Wasserstein-2 ball around the empirical distribution and solving a minimax problem for worst-case expected squared distances.
  • The authors derive a tractable dual that yields a soft-clustering method with smoothly weighted assignments instead of hard assignments.
  • They present an efficient block coordinate descent algorithm with proven monotonic decrease of the objective and local linear convergence guarantees.
  • Experiments on benchmarks and large synthetic datasets show improved robustness and better outlier detection under noise and distributional perturbations.

Abstract

K-means clustering is a workhorse of unsupervised learning, but it is notoriously brittle to outliers, distribution shifts, and limited sample sizes. Viewing k-means as Lloyd--Max quantization of the empirical distribution, we develop a distributionally robust variant that protects against such pathologies. We posit that the unknown population distribution lies within a Wasserstein-2 ball around the empirical distribution. In this setting, one seeks cluster centers that minimize the worst-case expected squared distance over this ambiguity set, leading to a minimax formulation. A tractable dual yields a soft-clustering scheme that replaces hard assignments with smoothly weighted ones. We propose an efficient block coordinate descent algorithm with provable monotonic decrease and local linear convergence. Experiments on standard benchmarks and large-scale synthetic data demonstrate substantial gains in outlier detection and robustness to noise.