Learning Uncertainty from Sequential Internal Dispersion in Large Language Models
arXiv cs.CL / 4/20/2026
📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research
Key Points
- The paper introduces SIVR (Sequential Internal Variance Representation), a supervised hallucination-detection framework for LLMs that estimates uncertainty using variance/dispersion of internal hidden representations across layers.
- Unlike prior methods that assume specific evolution patterns of hidden states or rely only on last/mean tokens (which can lose information), SIVR uses token-wise, layer-wise features to capture richer uncertainty signals.
- SIVR aggregates the full sequence of per-token variance features, enabling it to learn temporal patterns linked to factual errors.
- Experiments on hallucination detection show SIVR consistently beats strong baselines, with improved generalization and reduced dependence on very large training datasets.
- The authors provide an open-source code repository to support practical adoption: https://github.com/ponhvoan/internal-variance.



