Last-Iterate Convergence of Randomized Kaczmarz and SGD with Greedy Step Size

arXiv cs.LG / 4/14/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes last-iterate convergence for SGD using a greedy step size on smooth quadratic problems in the interpolation regime, which includes Randomized Kaczmarz as a special case.
  • It establishes an improved per-iterate convergence rate of O(1/t^{3/4}), improving upon a prior O(1/t^{1/2}) guarantee posed and studied in earlier work.
  • The authors introduce a framework of stochastic contraction processes whose dynamics are characterized through an associated deterministic eigenvalue equation.
  • They develop a proof strategy using a discrete-to-continuous reduction to study the eigenvalue equation and derive the convergence bounds.
  • Overall, the work answers an open question from Attia, Schliserman, Sherman, and Koren about achieving faster last-iterate convergence in this setting.

Abstract

We study last-iterate convergence of SGD with greedy step size over smooth quadratics in the interpolation regime, a setting which captures the classical Randomized Kaczmarz algorithm as well as other popular iterative linear system solvers. For these methods, we show that the t-th iterate attains an O(1/t^{3/4}) convergence rate, addressing a question posed by Attia, Schliserman, Sherman, and Koren, who gave an O(1/t^{1/2}) guarantee for this setting. In the proof, we introduce the family of stochastic contraction processes, whose behavior can be described by the evolution of a certain deterministic eigenvalue equation, which we analyze via a careful discrete-to-continuous reduction.