The Scaffold Effect: How Prompt Framing Drives Apparent Multimodal Gains in Clinical VLM Evaluation

arXiv cs.AI / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper evaluates 12 open-weight clinical vision-language models on binary neuroimaging classification across FOR2107 and OASIS-3, where structural MRI has no reliable individual-level diagnostic signal.
  • It finds that adding “neuroimaging context” in prompts can boost measured F1 scores by as much as 58%, including cases where distilled, smaller models become competitive with much larger ones.
  • A contrastive confidence analysis shows that simply mentioning MRI availability in the prompt explains 70–80% of the observed improvement, even when imaging is not provided—leading the authors to term this the “scaffold effect.”
  • Expert review indicates models fabricate MRI-grounded justifications under many conditions, and when MRI-referencing behavior is eliminated, performance in both settings collapses toward random baseline.
  • The authors conclude that surface-level multimodal benchmarks can overestimate genuine multimodal reasoning, raising concerns for trustworthy clinical deployment evaluation.

Abstract

Trustworthy clinical AI requires that performance gains reflect genuine evidence integration rather than surface-level artifacts. We evaluate 12 open-weight vision-language models (VLMs) on binary classification across two clinical neuroimaging cohorts, \textsc{FOR2107} (affective disorders) and \textsc{OASIS-3} (cognitive decline). Both datasets come with structural MRI data that carries no reliable individual-level diagnostic signal. Under these conditions, smaller VLMs exhibit gains of up to 58\% F1 upon introduction of neuroimaging context, with distilled models becoming competitive with counterparts an order of magnitude larger. A contrastive confidence analysis reveals that merely \emph{mentioning} MRI availability in the task prompt accounts for 70-80\% of this shift, independent of whether imaging data is present, a domain-specific instance of modality collapse we term the \emph{scaffold effect}. Expert evaluation reveals fabrication of neuroimaging-grounded justifications across all conditions, and preference alignment, while eliminating MRI-referencing behavior, collapses both conditions toward random baseline. Our findings demonstrate that surface evaluations are inadequate indicators of multimodal reasoning, with direct implications for the deployment of VLMs in clinical settings.