The Cost of Reasoning: Chain-of-Thought Induces Overconfidence in Vision-Language Models
arXiv cs.LG / 3/18/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The authors show that extended reasoning through chain-of-thought prompting in vision-language models reduces the reliability of uncertainty estimates, even if it improves task accuracy.
- The primary mechanism is implicit answer conditioning: as reasoning traces converge on a conclusion, token probabilities reflect consistency with the model's own reasoning rather than true uncertainty about correctness, leading to overconfidence.
- In contrast, agreement-based consistency remains robust under reasoning and often improves, making it a practical uncertainty estimator in reasoning-enabled VLMs.
- These findings have important implications for deploying VLMs in high-stakes settings and for designing reliable uncertainty quantification methods in such systems.




