Don't Show Pixels, Show Cues: Unlocking Visual Tool Reasoning in Language Models via Perception Programs

arXiv cs.CV / 4/15/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that multimodal language models often underuse vision-tool outputs because feeding dense pixel-level (raw) tool representations misaligns with how LLMs do language-native reasoning.
  • It introduces “Perception Programs (P$^2$),” a training-free, model-agnostic method that rewrites tool outputs into compact, structured, cue-focused summaries that LLMs can more directly parse and reason over.
  • Experiments on six perception-centric tasks in BLINK show P$^2$ provides large gains over both base models and raw tool-augmented baselines.
  • Using GPT-5 Mini, P$^2$ boosts multi-view reasoning accuracy from 41.35% to 86.47% and relative depth from 52.42% to 81.45%, with about a 22% average improvement across tasks.
  • The method also delivers strong absolute gains (15–40%) on smaller MLLMs like InternVL3.5-4B and Qwen3VL-4B, outperforming prior tool-use approaches without training or model changes.

Abstract

Multimodal language models (MLLMs) are increasingly paired with vision tools (e.g., depth, flow, correspondence) to enhance visual reasoning. However, despite access to these tool-generated visual cues, MLLMs often fail to benefit from them. Existing approaches typically feed raw tool outputs into the model, but these dense, pixel-level representations are misaligned with the language-native reasoning strengths of LLMs, leading to weak perception and reliance on language priors. We argue that, in problems where vision tools can provide the necessary visual cues, the bottleneck is not more tool calls or larger MLLMs, it is how tool outputs are represented. We introduce Perception Programs (P^2), a training-free, model-agnostic method that rewrites tool outputs into compact, structured, language-native summaries that MLLMs can directly parse and reason over. Across six perception-centric tasks in BLINK, P^2 consistently yields large improvements over base models and raw tool-augmented baselines. With GPT-5 Mini as the base model, P^2 raises its accuracy from 41.35\% to 86.47\% on multi-view reasoning, from 52.42\% to 81.45\% on relative depth, and achieves a 22\% average gain across tasks, setting new state-of-the-art results. Even on smaller MLLMs, e.g., InternVL3.5-4B and Qwen3VL-4B, we observe 15-40\% absolute gains from P^2, surpassing prior agentic, supervised, and RL-based tool-use methods-without any training or model modifications.