Real-Time Visual Attribution Streaming in Thinking Model

arXiv cs.CV / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces an amortized framework for real-time visual attribution streaming in multimodal “thinking” models, aiming to ground long reasoning traces in visual evidence (e.g., when generating code from screenshots or solving math from images).
  • It addresses a core verification trade-off: faithful causal attribution is expensive because it requires repeated backward passes or perturbations, while attention maps are fast but not causally valid.
  • The proposed method learns to estimate the causal effects of semantic regions using rich attention-derived features, rather than relying on brute-force causal procedures.
  • Experiments on five benchmarks and four thinking models show faithfulness comparable to exhaustive causal methods, while allowing users to see grounding evidence as the model reasons (streaming) instead of only after generation.
  • The authors conclude that real-time, causally faithful attribution for multimodal reasoning is achievable via lightweight learning rather than costly computation.

Abstract

We present an amortized framework for real-time visual attribution streaming in multimodal thinking models. When these models generate code from a screenshot or solve math problems from images, their long reasoning traces should be grounded in visual evidence. However, verifying this reliance is challenging: faithful causal methods require costly repeated backward passes or perturbations, while raw attention maps offer instant access, they lack causal validity. To resolve this, we introduce an amortized approach that learns to estimate the causal effects of semantic regions directly from the rich signals encoded in attention features. Across five diverse benchmarks and four thinking models, our approach achieves faithfulness comparable to exhaustive causal methods while enabling visual attribution streaming, where users observe grounding evidence as the model reasons, not after. Our results demonstrate that real-time, faithful attribution in multimodal thinking models is achievable through lightweight learning, not brute-force computation.