The Anatomy of Uncertainty in LLMs

arXiv cs.AI / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that existing LLM uncertainty methods (single scores or aleatoric/epistemic splits) do not provide actionable guidance for improving generative models.
  • It proposes decomposing LLM uncertainty into three semantic components: input ambiguity, knowledge gaps, and decoding randomness.
  • Experiments suggest the dominant uncertainty component can vary depending on model size and task type.
  • The framework is presented as a way to better audit LLM reliability and detect hallucinations, enabling more targeted interventions for trustworthy deployments.

Abstract

Understanding why a large language model (LLM) is uncertain about the response is important for their reliable deployment. Current approaches, which either provide a single uncertainty score or rely on the classical aleatoric-epistemic dichotomy, fail to offer actionable insights for improving the generative model. Recent studies have also shown that such methods are not enough for understanding uncertainty in LLMs. In this work, we advocate for an uncertainty decomposition framework that dissects LLM uncertainty into three distinct semantic components: (i) input ambiguity, arising from ambiguous prompts; (ii) knowledge gaps, caused by insufficient parametric evidence; and (iii) decoding randomness, stemming from stochastic sampling. Through a series of experiments we demonstrate that the dominance of these components can shift across model size and task. Our framework provides a better understanding to audit LLM reliability and detect hallucinations, paving the way for targeted interventions and more trustworthy systems.
広告