Representational Curvature Modulates Behavioral Uncertainty in Large Language Models
arXiv cs.AI / 4/28/2026
📰 NewsModels & Research
Key Points
- The paper proposes a direct geometric link in autoregressive LLMs by measuring “contextual curvature” (how sharply representations bend over recent context) and relating it to token-level next-token entropy.
- Experiments on GPT-2 XL and Pythia-2.8B show contextual curvature correlates with entropy, and this relationship appears during training rather than only after convergence.
- Perturbation/intervention tests demonstrate selective causality: curvature-aligned manipulations reliably change entropy, while misaligned (geometrically inconsistent) manipulations do not.
- Training with a regularizer that encourages representations to be “straighter” leads to a modest reduction in token-level entropy without harming validation loss, suggesting a potentially useful behavior-controlling feature.
- Overall, the work identifies trajectory curvature as a task-aligned representational property that modulates behavioral uncertainty in LLMs.
Related Articles
LLMs will be a commodity
Reddit r/artificial

What it feels like to have to have Qwen 3.6 or Gemma 4 running locally
Reddit r/LocalLLaMA

Dex lands $5.3M to grow its AI-driven talent matching platform
Tech.eu

AI Voice Agents in Production: What Actually Works in 2026
Dev.to

How we built a browser-based AI Pathology platform
Dev.to