Learning Uncertainty from Sequential Internal Dispersion in Large Language Models

arXiv cs.CL / 4/20/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research

Key Points

  • The paper introduces SIVR (Sequential Internal Variance Representation), a supervised hallucination-detection framework for LLMs that estimates uncertainty using variance/dispersion of internal hidden representations across layers.
  • Unlike prior methods that assume specific evolution patterns of hidden states or rely only on last/mean tokens (which can lose information), SIVR uses token-wise, layer-wise features to capture richer uncertainty signals.
  • SIVR aggregates the full sequence of per-token variance features, enabling it to learn temporal patterns linked to factual errors.
  • Experiments on hallucination detection show SIVR consistently beats strong baselines, with improved generalization and reduced dependence on very large training datasets.
  • The authors provide an open-source code repository to support practical adoption: https://github.com/ponhvoan/internal-variance.

Abstract

Uncertainty estimation is a promising approach to detect hallucinations in large language models (LLMs). Recent approaches commonly depend on model internal states to estimate uncertainty. However, they suffer from strict assumptions on how hidden states should evolve across layers, and from information loss by solely focusing on last or mean tokens. To address these issues, we present Sequential Internal Variance Representation (SIVR), a supervised hallucination detection framework that leverages token-wise, layer-wise features derived from hidden states. SIVR adopts a more basic assumption that uncertainty manifests in the degree of dispersion or variance of internal representations across layers, rather than relying on specific assumptions, which makes the method model and task agnostic. It additionally aggregates the full sequence of per-token variance features, learning temporal patterns indicative of factual errors and thereby preventing information loss. Experimental results demonstrate SIVR consistently outperforms strong baselines. Most importantly, SIVR enjoys stronger generalisation and avoids relying on large training sets, highlighting the potential for practical deployment. Our code repository is available online at https://github.com/ponhvoan/internal-variance.