AI Navigate

VERI-DPO: Evidence-Aware Alignment for Clinical Summarization via Claim Verification and Direct Preference Optimization

arXiv cs.CL / 3/12/2026

💬 OpinionModels & Research

Key Points

  • VERI-DPO combines claim verification with Direct Preference Optimization to train a summarizer that stays faithful to fragmented EHR evidence using a retrieval-augmented verifier.
  • It labels claim-evidence pairs as Supported, Not Supported, or Not Addressed and uses these signals to derive length-controlled, contradiction-anchored preference pairs for learning.
  • On held-out ICU patients in MIMIC-III-Ext-VeriFact-BHC, Not Supported rates drop from 10.7% to 1.9% (local verifier) and 11.6% to 6.4% (GPT-4o), and validity rises from 76.7% to 82.5%.
  • The approach aims to reduce omissions and unsupported statements in LLM-based clinical summarization, improving reliability without sacrificing informative length.

Abstract

Brief Hospital Course (BHC) narratives must be clinically useful yet faithful to fragmented EHR evidence. LLM-based clinical summarizers still introduce unsupported statements, and alignment can encourage omissions ("say-less" degeneration). We introduce VERI-DPO, which uses claim verification to mine preferences and distill them into the summarizer with Direct Preference Optimization (DPO). On MIMIC-III-Ext-VeriFact-BHC (100 ICU patients; patient-level splits), we train a retrieval-augmented verifier to label claim-evidence pairs as Supported, Not Supported, or Not Addressed via a single-token format. The verifier scores sentence-level claims from sampled BHC candidates and aggregates margins into a coverage-aware utility to mine length-controlled, contradiction-anchored preference pairs. On held-out patients, verifier-mined preferences separate candidates by contradiction density, and VERI-DPO reduces Not Supported claim rates from 10.7% to 1.9% (local verifier judge) and from 11.6% to 6.4% (GPT-4o judge), while improving validity from 76.7% to 82.5% and maintaining informative length.