Limit Theorems for Stochastic Gradient Descent in High-Dimensional Single-Layer Networks
arXiv stat.ML / 5/1/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper analyzes high-dimensional scaling limits of online stochastic gradient descent (SGD), focusing on how the step size governs the effective dynamics.
- It identifies a critical step-size scaling regime for single-layer networks, where the behavior transitions from deterministic (ballistic) dynamics to a qualitatively new regime.
- At the critical scaling, the authors show that an additional correction term appears, altering the system’s phase diagram compared with purely deterministic limits.
- Near fixed points, under certain conditions, the diffusive (SDE) limits of the effective dynamics simplify to an Ornstein–Uhlenbeck process.
- The results connect the “information exponent” to sample complexity and argue that deterministic scaling limits cannot fully capture stochastic fluctuations in high-dimensional learning.
Related Articles

Why Autonomous Coding Agents Keep Failing — And What Actually Works
Dev.to

Why Enterprise AI Pilots Fail
Dev.to

The PDF Feature Nobody Asked For (That I Use Every Day)
Dev.to

How to Fix OpenClaw Tool Calling Issues
Dev.to

Mistral's new flagship Medium 3.5 folds chat, reasoning, and code into one model
THE DECODER