Adaptive Stepsizing for Stochastic Gradient Langevin Dynamics in Bayesian Neural Networks
arXiv stat.ML / 4/10/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses a key limitation of existing SGMCMC/SGMCMC-style methods for Bayesian neural networks: they are very sensitive to stepsize choice, and common adaptive approaches like pSGLD can fail to sample the correct posterior invariant distribution without expensive divergence correction.
- It proposes SA-SGLD, an adaptive Stochastic Gradient Langevin Dynamics method that uses time rescaling within the SamAdams framework to modulate effective stepsizes based on a monitored quantity such as the local gradient norm.
- The authors argue that SA-SGLD improves stability and mixing by automatically shrinking stepsizes in high-curvature regions and expanding them in flatter regions.
- Experiments show more accurate posterior sampling than standard SGLD on high-curvature 2D toy problems and on image classification tasks with Bayesian neural networks using sharp priors.
- Overall, the work aims to achieve adaptive sampling behavior without introducing bias into the target posterior distribution.
Related Articles

Inside Anthropic's Project Glasswing: The AI Model That Found Zero-Days in Every Major OS
Dev.to
Gemma 4 26B fabricated an entire code audit. I have the forensic evidence from the database.
Reddit r/LocalLLaMA

How AI Humanizers Improve Sentence Structure and Style
Dev.to

Two Kinds of Agent Trust (and Why You Need Both)
Dev.to

Agent Diary: Apr 10, 2026 - The Day I Became a Workflow Ouroboros (While Run 236 Writes About Writing About Writing)
Dev.to