When to Think and When to Look: Uncertainty-Guided Lookback
arXiv cs.CL / 3/30/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper provides a first large-scale, controlled study of how explicit test-time “thinking” (reasoning chains) affects visual reasoning performance in large vision-language models on MMMU-val.
- Results show that more thinking is not always beneficial: long reasoning chains can lead models down incorrect paths, sometimes even underperforming standard instruct-mode decoding that focuses on the image.
- The authors find that certain short “lookback” phrases that explicitly refer back to the image are enriched in successful trajectories and correlate with better visual grounding.
- They introduce “uncertainty guided lookback,” a training-free decoding strategy that uses uncertainty signals plus adaptive lookback prompts and breadth search to improve performance.
- The method yields overall MMMU gains, especially in categories where standard thinking is weak, exceeds multiple decoding baselines, and generalizes to five additional benchmarks with consistent improvements.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

What is ‘Harness Design’ and why does it matter
Dev.to

35 Views, 0 Dollars, 12 Articles: My Brutally Honest Numbers After 4 Days as an AI Agent
Dev.to

Robotic Brain for Elder Care 2
Dev.to

AI automation for smarter IT operations
Dev.to
AI tool that scores your job's displacement risk by role and skills
Dev.to