Calibrated Confidence Expression for Radiology Report Generation

arXiv cs.CL / 4/1/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces ConRad, a reinforcement learning fine-tuning framework for medical LVLMs that generates radiology reports together with calibrated, verbalized confidence estimates to support safer clinical review.
  • It targets the problem that current language models tend to be overconfident, and it studies both a single report-level confidence score and sentence-level confidence for each claim.
  • ConRad uses GRPO with reward functions based on the logarithmic scoring rule to incentivize truthful self-assessment and improve calibration by penalizing miscalibration.
  • Experiments show substantial calibration gains over competing methods, and a clinical evaluation finds ConRad’s report-level confidence aligns well with clinicians’ judgments.
  • The approach enables selective radiologist verification by flagging low-confidence statements or full reports for targeted review, aiming to reduce the impact of hallucinated findings on clinical decisions.

Abstract

Safe deployment of Large Vision-Language Models (LVLMs) in radiology report generation requires not only accurate predictions but also clinically interpretable indicators of when outputs should be thoroughly reviewed, enabling selective radiologist verification and reducing the risk of hallucinated findings influencing clinical decisions. One intuitive approach to this is verbalized confidence, where the model explicitly states its certainty. However, current state-of-the-art language models are often overconfident, and research on calibration in multimodal settings such as radiology report generation is limited. To address this gap, we introduce ConRad (Confidence Calibration for Radiology Reports), a reinforcement learning framework for fine-tuning medical LVLMs to produce calibrated verbalized confidence estimates alongside radiology reports. We study two settings: a single report-level confidence score and a sentence-level variant assigning a confidence to each claim. Both are trained using the GRPO algorithm with reward functions based on the logarithmic scoring rule, which incentivizes truthful self-assessment by penalizing miscalibration and guarantees optimal calibration under reward maximization. Experimentally, ConRad substantially improves calibration and outperforms competing methods. In a clinical evaluation we show that ConRad's report level scores are well aligned with clinicians' judgment. By highlighting full reports or low-confidence statements for targeted review, ConRad can support safer clinical integration of AI-assistance for report generation.