Act2See: Emergent Active Visual Perception for Video Reasoning

arXiv cs.CV / 5/5/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Vision-Language Models (VLMs) often use only static initial frames for video reasoning, which limits their ability to incorporate dynamic evidence as reasoning progresses.
  • The proposed Act-to-See (Act2See) framework lets VLMs actively interleave video frames into text-based Chain-of-Thought (CoT), improving visual synthesis and enabling hypothetical/counterfactual reasoning.
  • Act2See is trained via supervised fine-tuning (SFT) on high-quality reasoning-trace data generated by a frontier VLM, where the traces include verified active frame retrieval or synthesis steps.
  • At inference time, the model dynamically decides when to retrieve existing frames or generate/synthesize new ones to obtain the needed visual evidence.
  • Experiments report new state-of-the-art performance on VideoEspresso and ViTIB, and improvements over comparable or larger models on several other video reasoning benchmarks.

Abstract

Vision-Language Models (VLMs) typically rely on static initial frames for video reasoning, restricting their ability to incorporate essential dynamic information as the reasoning process evolves. Existing methods that augment Chain-of-Thought (CoT) with additional frame information often exhibit suboptimal CoT quality and lack the crucial ability to synthesize visual information for hypothetical or counterfactual scenarios. We introduce Act-to-See (Act2See), a novel framework that enables active visual perception by empowering VLMs to actively interleave video frames within text CoTs. Act2See is developed via Supervised Fine-Tuning (SFT) on a high-quality dataset of reasoning traces generated by a frontier VLM. These traces integrate active calls to either retrieve existing frames or generate new ones, and are rigorously verified against human-annotated CoTs to ensure quality. This approach cultivates an emergent capability: at inference time, the model actively determines when to search for or synthesize the necessary visual evidence. Act2See establishes new state-of-the-art results on challenging benchmarks, including VideoEspresso and ViTIB, and outperforms comparable or larger models on Video-MME, EgoNormia, and VCR-Bench, demonstrating an advancement in enabling VLMs with active visual perception for video reasoning.