AI Navigate

DeCode: Decoupling Content and Delivery for Medical QA

arXiv cs.CL / 3/16/2026

📰 NewsModels & Research

Key Points

  • DeCode is a training-free, model-agnostic framework that decouples content and delivery to tailor LLM answers to individual clinical contexts.
  • It evaluates on OpenAI HealthBench and reports a zero-shot performance rise from 28.4% to 49.8%, achieving new state-of-the-art among existing methods.
  • The approach enables contextualized clinical QA without additional fine-tuning, facilitating deployment across existing LLMs in healthcare settings.
  • Experimental results suggest DeCode improves clinical relevance and validity of LLM responses, with practical benefits for patient-centered care.

Abstract

Large language models (LLMs) exhibit strong medical knowledge and can generate factually accurate responses. However, existing models often fail to account for individual patient contexts, producing answers that are clinically correct yet poorly aligned with patients' needs. In this work, we introduce DeCode (Decoupling Content and Delivery), a training-free, model-agnostic framework that adapts existing LLMs to produce contextualized answers in clinical settings. We evaluate DeCode on OpenAI HealthBench, a comprehensive and challenging benchmark designed to assess clinical relevance and validity of LLM responses. DeCode boosts zero-shot performance from 28.4% to 49.8% and achieves new state-of-the-art compared to existing methods. Experimental results suggest the effectiveness of DeCode in improving clinical question answering of LLMs.