Dynamical Systems Theory Behind a Hierarchical Reasoning Model

arXiv cs.AI / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that existing LLM-style sequence generation and even prior reasoning models (HRM/TRM) can fail on complex algorithmic reasoning due to unstable or insufficiently justified training dynamics.
  • It introduces the Contraction Mapping Model (CMM), reformulating hierarchical/recursive reasoning into continuous NODE/NSDE latent dynamics with explicit convergence toward a stable equilibrium.
  • To prevent representational failure, the method uses a hyperspherical repulsion loss intended to mitigate feature collapse during training.
  • On the Sudoku-Extreme benchmark, a 5M-parameter CMM reaches 93.7% accuracy, greatly outperforming larger or prior hierarchical models, while still performing strongly under extreme compression (0.26M parameters).
  • The authors claim the results demonstrate a direction for replacing brute-force parameter scaling with mathematically grounded latent dynamics for robust reasoning engines.

Abstract

Current large language models (LLMs) primarily rely on linear sequence generation and massive parameter counts, yet they severely struggle with complex algorithmic reasoning. While recent reasoning architectures, such as the Hierarchical Reasoning Model (HRM) and Tiny Recursive Model (TRM), demonstrate that compact recursive networks can tackle these tasks, their training dynamics often lack rigorous mathematical guarantees, leading to instability and representational collapse. We propose the Contraction Mapping Model (CMM), a novel architecture that reformulates discrete recursive reasoning into continuous Neural Ordinary and Stochastic Differential Equations (NODEs/NSDEs). By explicitly enforcing the convergence of the latent phase point to a stable equilibrium state and mitigating feature collapse with a hyperspherical repulsion loss, the CMM provides a mathematically grounded and highly stable reasoning engine. On the Sudoku-Extreme benchmark, a 5M-parameter CMM achieves a state-of-the-art accuracy of 93.7 %, outperforming the 27M-parameter HRM (55.0 %) and 5M-parameter TRM (87.4 %). Remarkably, even when aggressively compressed to an ultra-tiny footprint of just 0.26M parameters, the CMM retains robust predictive power, achieving 85.4 % on Sudoku-Extreme and 82.2 % on the Maze benchmark. These results establish a new frontier for extreme parameter efficiency, proving that mathematically rigorous latent dynamics can effectively replace brute-force scaling in artificial reasoning.