RVLM: Recursive Vision-Language Models with Adaptive Depth

arXiv cs.CV / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • RVLM (Recursive Vision-Language Models) is proposed as a unified framework for medical vision-language AI that improves auditability and explanation by grounding each diagnostic claim in executable Python code within an iterative generate-execute loop.
  • The method uses vision sub-agents at each step to manipulate images and accumulate evidence, replacing conventional single-pass VLM inference with a more transparent reasoning process.
  • RRouter introduces adaptive iteration depth, using a lightweight controller to predict the optimal reasoning budget from task-complexity features and to terminate early when progress stalls, reducing wasted compute.
  • Experiments on BraTS 2023 Meningioma (brain MRI) and MIMIC-CXR (chest X-ray) with Gemini 2.5 Flash (no fine-tuning) show consistent detection of key findings and the ability to identify cross-modal discrepancies, while generating structured reports for radiology tasks.
  • The authors provide code in a public repository, enabling reproducibility and further evaluation of the audit-friendly, adaptive-depth medical VLM approach.

Abstract

Medical AI systems face two fundamental limitations. First, conventional vision-language models (VLMs) perform single-pass inference, yielding black-box predictions that cannot be audited or explained in clinical terms. Second, iterative reasoning systems that expose intermediate steps rely on fixed iteration budgets wasting compute on simple cases while providing insufficient depth for complex ones. We address both limitations with a unified framework. RVLM replaces single-pass inference with an iterative generate-execute loop: at each step, the model writes Python code, invokes vision sub-agents, manipulates images, and accumulates evidence. Every diagnostic claim is grounded in executable code, satisfying auditability requirements of clinical AI governance frameworks. RRouter makes iteration depth adaptive: a lightweight controller predicts the optimal budget from task-complexity features, then monitors progress and terminates early when reasoning stalls. We evaluate on BraTS 2023 Meningioma (brain MRI) and MIMIC-CXR (chest X-ray) using Gemini 2.5 Flash without fine-tuning. Across repeated runs, RVLM shows high consistency on salient findings (e.g., mass presence and enhancement) and can detect cross-modal discrepancies between Fluid-Attenuated Inversion Recovery (FLAIR) signal characteristics and segmentation boundaries. On MIMIC-CXR, it generates structured reports and correctly recognises view-specific artefacts. Code: https://github.com/nican2018/rvlm.