A ghost mechanism: An analytical model of abrupt learning in recurrent networks
arXiv stat.ML / 4/16/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes the “ghost mechanism” as an analytical dynamical-systems model for abrupt learning in recurrent neural networks (RNNs), attributing sudden performance gains to transient slowdowns near remnants of a saddle-node bifurcation.
- By reducing high-dimensional dynamics to a one-dimensional canonical form with a single scale parameter, the authors derive how abrupt learning behavior depends on learning rate and the timescale of the learned computation.
- The study identifies a critical learning-rate threshold (scaling as an inverse power law with the computation timescale), beyond which learning breaks down via two interacting issues: vanishing gradients and oscillatory gradients near minima.
- The authors show that these effects can trap training in “no-learning zones” where gradients vanish, leading the system to make high-confidence but incorrect predictions, and validate the theory in both low-rank and full-rank RNNs on working-memory tasks.
- Two mitigation strategies are suggested: increasing trainable ranks to stabilize learning trajectories and reducing output confidence to prevent entrapment in no-learning zones.
Related Articles

Introducing Claude Opus 4.7
Anthropic News

Who Audits the Auditors? Building an LLM-as-a-Judge for Agentic Reliability
Dev.to

"Enterprise AI Cost Optimization: How Companies Are Cutting AI Infrastructure Sp
Dev.to

Config-first code generator to replace repetitive AI boilerplate — looking for feedback and collaborators
Dev.to

The US Government Fired 40% of an Agency, Then Asked AI to Do Their Jobs
Dev.to