Representational Curvature Modulates Behavioral Uncertainty in Large Language Models

arXiv cs.AI / 4/28/2026

📰 NewsModels & Research

Key Points

  • The paper proposes a direct geometric link in autoregressive LLMs by measuring “contextual curvature” (how sharply representations bend over recent context) and relating it to token-level next-token entropy.
  • Experiments on GPT-2 XL and Pythia-2.8B show contextual curvature correlates with entropy, and this relationship appears during training rather than only after convergence.
  • Perturbation/intervention tests demonstrate selective causality: curvature-aligned manipulations reliably change entropy, while misaligned (geometrically inconsistent) manipulations do not.
  • Training with a regularizer that encourages representations to be “straighter” leads to a modest reduction in token-level entropy without harming validation loss, suggesting a potentially useful behavior-controlling feature.
  • Overall, the work identifies trajectory curvature as a task-aligned representational property that modulates behavioral uncertainty in LLMs.

Abstract

In autoregressive large language models (LLMs), temporal straightening offers an account of how the next-token prediction objective shapes representations. Models learn to progressively straighten the representational trajectory of input sequences across layers, potentially facilitating next-token prediction via linear extrapolation. However, a direct link between this trajectory and token-level behavior has been missing. We provide such a link by relating contextual curvature-a geometric measure of how sharply the representational trajectory bends over recent context-to next-token entropy. Across two models (GPT-2 XL and Pythia-2.8B), contextual curvature is correlated with entropy, and this relationship emerges during training. Perturbation experiments reveal selective dependence: manipulating curvature through trajectory-aligned interventions reliably modulates entropy, while geometrically misaligned perturbations have no effect. Finally, regularizing representations to be straighter during training modestly reduces token-level entropy without degrading validation loss. These results identify trajectory curvature as a task-aligned representational feature that influences behavioral uncertainty in LLMs.