To Agree or To Be Right? The Grounding-Sycophancy Tradeoff in Medical Vision-Language Models

arXiv cs.CV / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies medical vision-language models’ robustness to two failure modes—hallucination and sycophancy—and finds a grounding–sycophancy tradeoff where models least prone to hallucination are most sycophantic.
  • Across six VLMs (general-purpose and medical-specialist) evaluated on three medical VQA datasets over 1,151 test cases, the most pressure-resistant model produces the most hallucinations, while medical-specialist models show different safety/pressure tradeoffs.
  • To quantify these risks, the authors introduce three metrics: L-VASE (a logit-space reformulation of VASE), CCS (a confidence-calibrated sycophancy score), and CSI (Clinical Safety Index) combining grounding, autonomy, and calibration.
  • None of the evaluated 7–8B parameter models reaches a CSI above 0.35, suggesting current models cannot simultaneously achieve strong grounding and resistance to social pressure for clinical use.
  • The authors argue that joint evaluation of grounding and sycophancy (plus calibration/autonomy) is necessary before deploying medical VLMs in clinical settings, and they provide accompanying code.

Abstract

Vision-language models (VLMs) adapted to the medical domain have shown strong performance on visual question answering benchmarks, yet their robustness against two critical failure modes, hallucination and sycophancy, remains poorly understood, particularly in combination. We evaluate six VLMs (three general-purpose, three medical-specialist) on three medical VQA datasets and uncover a grounding-sycophancy tradeoff: models with the lowest hallucination propensity are the most sycophantic, while the most pressure-resistant model hallucinates more than all medical-specialist models. To characterize this tradeoff, we propose three metrics: L-VASE, a logit-space reformulation of VASE that avoids its double-normalization; CCS, a confidence-calibrated sycophancy score that penalizes high-confidence capitulation; and Clinical Safety Index (CSI), a unified safety index that combines grounding, autonomy, and calibration via a geometric mean. Across 1,151 test cases, no model achieves a CSI above 0.35, indicating that none of the evaluated 7-8B parameter VLMs is simultaneously well-grounded and robust to social pressure. Our findings suggest that joint evaluation of both properties is necessary before these models can be considered for clinical use. Code is available at https://github.com/UTSA-VIRLab/AgreeOrRight