The Geometric Price of Discrete Logic: Context-driven Manifold Dynamics of Number Representations

arXiv cs.LG / 2026/3/26

💬 オピニオンSignals & Early TrendsIdeas & Deep AnalysisModels & Research

要点

  • The paper argues that discrete logical reasoning in LLMs requires context-driven, non-isometric geometric distortions rather than linear isometric projections.
  • Using Gram-Schmidt decomposition on residual-stream activations, it proposes a dual mechanism: class-agnostic topological preservation to maintain global structure and an algebraic divergence component to separate cross-class concepts into decision boundaries.
  • Experiments across tasks from simple mappings to primality testing show that this geometric evolution occurs in practice, and targeted vector ablation causally links the divergence component to function quality.
  • Removing the divergence component sharply degrades parity classification performance from 100% to chance (38.57%), supporting a causal “topological deformation” cost for forming discrete logic.
  • The authors also identify a three-phase, layer-wise geometric dynamic and suggest that social pressure reduces sufficient divergence, leading to “manifold entanglement” that may underlie sycophancy and hallucination.

Abstract

Large language models (LLMs) generalize smoothly across continuous semantic spaces, yet strict logical reasoning demands the formation of discrete decision boundaries. Prevailing theories relying on linear isometric projections fail to resolve this fundamental tension. In this work, we argue that task context operates as a non-isometric dynamical operator that enforces a necessary "topological distortion." By applying Gram-Schmidt decomposition to residual-stream activations , we reveal a dual-modulation mechanism driving this process: a class-agnostic topological preservation that anchors global structure to prevent semantic collapse, and a specific algebraic divergence that directionally tears apart cross-class concepts to forge logical boundaries. We validate this geometric evolution across a gradient of tasks, from simple mapping to complex primality testing. Crucially, targeted specific vector ablation establishes a strict causal binding between this topology and model function: algebraically erasing the divergence component collapses parity classification accuracy from 100% to chance levels (38.57%). Furthermore, we uncover a three-phase layer-wise geometric dynamic and demonstrate that under social pressure prompts, models fail to generate sufficient divergence. This results in a "manifold entanglement" that geometrically explains sycophancy and hallucination. Ultimately, our findings revise the linear-isometric presumption, demonstrating that the emergence of discrete logic in LLMs is purchased at an irreducible cost of topological deformation.