Information as Structural Alignment: A Dynamical Theory of Continual Learning

arXiv cs.LG / 4/9/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that catastrophic forgetting arises mathematically from storing knowledge as a global parameter superposition, rather than being solely an engineering shortcoming.
  • It proposes the Informational Buildup Framework (IBF), where information is realized through structural alignment, and learning is driven by (1) a Law of Motion increasing coherence and (2) Modification Dynamics that reshape the coherence landscape based on localized errors.
  • Unlike prior continual-learning methods that add external mechanisms (e.g., replay, regularization, or frozen subnetworks), IBF aims to produce memory and self-correction directly from learning dynamics.
  • The authors demonstrate a full lifecycle in a 2D toy setting and then evaluate IBF on three benchmarks, including non-stationary control, chess scored independently by Stockfish, and Split-CIFAR-100 using a frozen ViT encoder.
  • Results reportedly show replay-superior retention without storing raw data, including near-zero forgetting on CIFAR-100 (BT = -0.004), positive backward transfer in chess (+38.5 cp), and reduced forgetting versus replay in the controlled domain, with strong independent chess advantages over baselines.

Abstract

Catastrophic forgetting is not an engineering failure. It is a mathematical consequence of storing knowledge as global parameter superposition. Existing methods, such as regularization, replay, and frozen subnetworks, add external mechanisms to a shared-parameter substrate. None derives retention from the learning dynamics themselves. This paper introduces the Informational Buildup Framework (IBF), an alternative substrate for continual learning, based on the premise that information is the achievement of structural alignment rather than stored content. In IBF, two equations govern the dynamics: a Law of Motion that drives configuration toward higher coherence, and Modification Dynamics that persistently deform the coherence landscape in response to localized discrepancies. Memory, agency, and self-correction arise from these dynamics rather than being added as separate modules. We first demonstrate the full lifecycle in a transparent two-dimensional toy model, then validate across three domains: a controlled non-stationary world, chess evaluated independently by Stockfish, and Split-CIFAR-100 with a frozen ViT encoder. Across all three, IBF achieves replay-superior retention without storing raw data. We observe near-zero forgetting on CIFAR-100 (BT = -0.004), positive backward transfer in chess (+38.5 cp), and 43% less forgetting than replay in the controlled domain. In chess, the framework achieves a mean behavioral advantage of +88.9 +/- 2.8 cp under independent evaluation, exceeding MLP and replay baselines.