SELFDOUBT: Uncertainty Quantification for Reasoning LLMs via the Hedge-to-Verify Ratio

arXiv cs.AI / 4/10/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SELFDOUBT, a single-pass uncertainty estimation method for reasoning LLMs that works even when proprietary APIs hide logits and intermediate probabilities.
  • SELFDOUBT derives an uncertainty score from behavioral cues in the model’s reasoning trace using the Hedge-to-Verify Ratio (HVR), distinguishing hedging/uncertainty markers from explicit self-checking.
  • Across seven models on BBH, GPQA-Diamond, and MMLU-Pro, traces without hedging markers are correct 96% of the time, enabling a high-precision “confidence gate” at no extra inference cost.
  • In the hedging-marker cases, SELFDOUBT outperforms sampling-based semantic entropy while requiring about 10x lower inference cost.
  • A two-stage deployment cascade combining the zero-marker gate and the full SELFDOUBT score achieves 90% accuracy at 71% coverage without task-specific labels, suggesting a production-ready uncertainty foundation.

Abstract

Uncertainty estimation for reasoning language models remains difficult to deploy in practice: sampling-based methods are computationally expensive, while common single-pass proxies such as verbalized confidence or trace length are often inconsistent across models. This problem is compounded for proprietary reasoning APIs that expose neither logits nor intermediate token probabilities, leaving practitioners with no reliable uncertainty signal at inference time. We propose SELFDOUBT, a single-pass uncertainty framework that resolves this impasse by extracting behavioral signals directly from the reasoning trace itself. Our key signal, the Hedge-to-Verify Ratio (HVR), detects whether a reasoning trace contains uncertainty markers and, if so, whether they are offset by explicit selfchecking behavior. Unlike methods that require multiple sampled traces or model internals, SELFDOUBT operates on a single observed reasoning trajectory, making it suitable for latency- and cost-constrained deployment over any proprietary API. We evaluate SELFDOUBT across seven models and three multi-step reasoning benchmarks (BBH, GPQA-Diamond, and MMLU-Pro). Most notably, traces containing no hedging markers are correct 96% of the time, revealing an emergent high-precision confidence gate at zero additional cost. For the remaining cases, the full SELFDOUBT score significantly outperforms sampling-based semantic entropy at 10x lower inference cost. A deployment cascade combining both stages attains 90% accuracy at 71% coverage without any task-specific labels. These results establish SELFDOUBT as a scalable, production-ready foundation for uncertainty estimation over proprietary reasoning models.