Efficient Hallucination Detection: Adaptive Bayesian Estimation of Semantic Entropy with Guided Semantic Exploration

arXiv cs.CL / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes an Adaptive Bayesian Estimation framework for hallucination detection that estimates “semantic entropy” using a hierarchical Bayesian model rather than fixed sampling budgets.
  • It dynamically adjusts the number of LLM samples based on observed uncertainty and stops generation early once variance-based thresholds indicate sufficient certainty, improving compute efficiency.
  • To explore the semantic space more effectively, the method adds a perturbation-based importance sampling strategy for systematic guided semantic exploration.
  • Experiments on four QA datasets show better hallucination detection performance with efficiency gains, including about 50% fewer samples in low-budget settings and an average AUROC improvement of 12.6% under the same budget.
  • The approach is positioned as more computationally scalable for practical use, especially when query complexity varies and fixed re-sampling is wasteful.

Abstract

Large language models (LLMs) have achieved remarkable success in various natural language processing tasks, yet they remain prone to generating factually incorrect outputs known as hallucinations. While recent approaches have shown promise for hallucination detection by repeatedly sampling from LLMs and quantifying the semantic inconsistency among the generated responses, they rely on fixed sampling budgets that fail to adapt to query complexity, resulting in computational inefficiency. We propose an Adaptive Bayesian Estimation framework for Semantic Entropy with Guided Semantic Exploration, which dynamically adjusts sampling requirements based on observed uncertainty. Our approach employs a hierarchical Bayesian framework to model the semantic distribution, enabling dynamic control of sampling iterations through variance-based thresholds that terminate generation once sufficient certainty is achieved. We also develop a perturbation-based importance sampling strategy to systematically explore the semantic space. Extensive experiments on four QA datasets demonstrate that our method achieves superior hallucination detection performance with significant efficiency gains. In low-budget scenarios, our approach requires about 50% fewer samples to achieve comparable detection performance to existing methods, while delivers an average AUROC improvement of 12.6% under the same sampling budget.