An Information-Geometric Framework for Stability Analysis of Large Language Models under Entropic Stress
arXiv cs.AI / 4/28/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that LLM reliability in high-stakes deployments cannot be fully captured by aggregate accuracy and proposes a new evaluation approach based on thermodynamic/ information-geometric ideas.
- It introduces a composite “stability score” that combines task utility, entropy (external uncertainty), and two internal structural proxies (internal integration and aligned reflective capacity) to model how disorder affects behavior.
- Using the IST-20 benchmarking protocol and metadata, the authors analyze 80 model-scenario observations across four contemporary LLMs and find that the full formulation yields higher stability scores than a reduced utility–entropy baseline.
- The reported average improvement is 0.0299 (95% CI: 0.0247–0.0351), and the benefit is larger under higher-entropy conditions, indicating a stronger, nonlinear-like attenuation of uncertainty.
- The work is positioned as an interpretable abstraction for connecting uncertainty, performance, and internal structure, intended to complement existing safety, reliability, and governance discussions rather than claim physical law or a complete theory of ethics.
Related Articles
LLMs will be a commodity
Reddit r/artificial
Indian Developers: How to Build AI Side Income with $0 Capital in 2026
Dev.to

What it feels like to have to have Qwen 3.6 or Gemma 4 running locally
Reddit r/LocalLLaMA

Dex lands $5.3M to grow its AI-driven talent matching platform
Tech.eu
AI Citation Registry: Why Daily Updates Leave No Time for Data Structuring
Dev.to