Complementing Self-Consistency with Cross-Model Disagreement for Uncertainty Quantification

arXiv cs.AI / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper shows that self-consistency-based aleatoric uncertainty (AU) can fail when LLMs are overconfident and repeatedly produce the same incorrect answer across samples.
  • It finds that cross-model semantic disagreement is higher on incorrect answers specifically when AU is low, suggesting a complementary signal for uncertainty.
  • The authors propose an epistemic uncertainty (EU) method for black-box settings that uses only generated text from a small, scale-matched model ensemble and measures a similarity gap between inter-model and intra-model semantic scores.
  • By defining total uncertainty (TU) as AU + EU, the method improves ranking calibration and selective abstention across multiple instruction-tuned models and long-form tasks, and better flags confident failures.
  • The study also analyzes when EU is most effective using agreement and complementarity diagnostics, indicating practical conditions for applying the approach.

Abstract

Large language models (LLMs) often produce confident yet incorrect responses, and uncertainty quantification is one potential solution to more robust usage. Recent works routinely rely on self-consistency to estimate aleatoric uncertainty (AU), yet this proxy collapses when models are overconfident and produce the same incorrect answer across samples. We analyze this regime and show that cross-model semantic disagreement is higher on incorrect answers precisely when AU is low. Motivated by this, we introduce an epistemic uncertainty (EU) term that operates in the black-box access setting: EU uses only generated text from a small, scale-matched ensemble and is computed as the gap between inter-model and intra-model sequence-semantic similarity. We then define total uncertainty (TU) as the sum of AU and EU. In a comprehensive study across five 7-9B instruction-tuned models and ten long-form tasks, TU improves ranking calibration and selective abstention relative to AU, and EU reliably flags confident failures where AU is low. We further characterize when EU is most useful via agreement and complementarity diagnostics.