Towards Reliable Truth-Aligned Uncertainty Estimation in Large Language Models

arXiv cs.AI / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies why LLM uncertainty estimation (UE) metrics can fail to reliably detect hallucinations, attributing the instability to “proxy failure” where metrics reflect model behavior rather than factual correctness.
  • It shows UE metrics can become non-discriminative in low-information settings, limiting their usefulness for trustworthy reliability assessment.
  • To address this, the authors introduce Truth AnChoring (TAC), a post-hoc calibration method that maps raw UE scores to “truth-aligned” uncertainty scores.
  • Experiments indicate TAC can produce well-calibrated uncertainty estimates even with noisy and few-shot supervision, along with a practical calibration protocol.
  • The work argues that heuristic UE metrics should not be treated as direct indicators of truth uncertainty and positions TAC as a step toward more reliable UE for LLMs, with an accompanying code repository.

Abstract

Uncertainty estimation (UE) aims to detect hallucinated outputs of large language models (LLMs) to improve their reliability. However, UE metrics often exhibit unstable performance across configurations, which significantly limits their applicability. In this work, we formalise this phenomenon as proxy failure, since most UE metrics originate from model behaviour, rather than being explicitly grounded in the factual correctness of LLM outputs. With this, we show that UE metrics become non-discriminative precisely in low-information regimes. To alleviate this, we propose Truth AnChoring (TAC), a post-hoc calibration method to remedy UE metrics, by mapping the raw scores to truth-aligned scores. Even with noisy and few-shot supervision, our TAC can support the learning of well-calibrated uncertainty estimates, and presents a practical calibration protocol. Our findings highlight the limitations of treating heuristic UE metrics as direct indicators of truth uncertainty, and position our TAC as a necessary step toward more reliable uncertainty estimation for LLMs. The code repository is available at https://github.com/ponhvoan/TruthAnchor/.