Exact Risk Curves of signSGD in High-Dimensions: Quantifying Preconditioning and Noise-Compression Effects

arXiv stat.ML / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper analyzes signSGD in a high-dimensional limit, deriving limiting SDE/ODE dynamics that describe how training risk evolves over time.
  • It provides a quantitative breakdown of four mechanisms attributed to signSGD: effective learning-rate adjustment, noise compression, diagonal preconditioning, and reshaping of gradient noise.
  • The authors’ results align with existing experimental observations while extending them by showing how the effects depend on the underlying data and noise distributions.
  • The work ends with a conjecture for extending the framework to Adam, aiming to connect signSGD’s behavior to more complex adaptive optimizers.

Abstract

In recent years, signSGD has garnered interest as both a practical optimizer as well as a simple model to understand adaptive optimizers like Adam. Though there is a general consensus that signSGD acts to precondition optimization and reshapes noise, quantitatively understanding these effects in theoretically solvable settings remains difficult. We present an analysis of signSGD in a high dimensional limit, and derive a limiting SDE and ODE to describe the risk. Using this framework we quantify four effects of signSGD: effective learning rate, noise compression, diagonal preconditioning, and gradient noise reshaping. Our analysis is consistent with experimental observations but moves beyond that by quantifying the dependence of these effects on the data and noise distributions. We conclude with a conjecture on how these results might be extended to Adam.