INTRYGUE: Induction-Aware Entropy Gating for Reliable RAG Uncertainty Estimation
arXiv cs.AI / 2026/3/24
💬 オピニオンIdeas & Deep AnalysisModels & Research
要点
- The paper argues that common entropy-based uncertainty quantification methods can misbehave in retrieval-augmented generation (RAG) because induction mechanisms interact with internal components that inflate predictive entropy.
- It identifies a “tug-of-war” effect: induction heads help produce grounded answers by copying relevant content, but they also activate previously established “entropy neurons,” leading the model to report false uncertainty even when outputs are correct.
- The proposed method, INTRYGUE, gates predictive entropy using induction-head activation patterns to better reflect true uncertainty in RAG scenarios.
- Experiments across four RAG benchmarks and six open-source LLMs (4B–13B) show INTRYGUE consistently matches or outperforms multiple uncertainty quantification baselines.
- The work concludes that more reliable hallucination detection in RAG can come from combining predictive uncertainty with mechanistically interpretable signals tied to context utilization.

