Grokking From Abstraction to Intelligence

arXiv cs.AI / 4/1/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper examines “grokking” in modular arithmetic as a key testbed for understanding how neural models generalize after initially memorizing training data.
  • It argues that prior work has focused too narrowly on local circuitry or optimization details, and instead proposes a global structural explanation for the grokking transition.
  • The authors claim grokking arises from a spontaneous simplification of internal model structures driven by a parsimony principle.
  • They use causal, spectral, and algorithmic complexity metrics, combined with Singular Learning Theory, to link the memorization-to-generalization shift with the collapse of redundant manifolds and “deep information compression.”
  • The proposed framework reframes model overfitting and generalization as physically grounded changes in internal representations rather than only changes in training dynamics.

Abstract

Grokking in modular arithmetic has established itself as the quintessential fruit fly experiment, serving as a critical domain for investigating the mechanistic origins of model generalization. Despite its significance, existing research remains narrowly focused on specific local circuits or optimization tuning, largely overlooking the global structural evolution that fundamentally drives this phenomenon. We propose that grokking originates from a spontaneous simplification of internal model structures governed by the principle of parsimony. We integrate causal, spectral, and algorithmic complexity measures alongside Singular Learning Theory to reveal that the transition from memorization to generalization corresponds to the physical collapse of redundant manifolds and deep information compression, offering a novel perspective for understanding the mechanisms of model overfitting and generalization.