MedPRMBench: A Fine-grained Benchmark for Process Reward Models in Medical Reasoning

arXiv cs.CL / 4/21/2026

📰 NewsSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The paper introduces MedPRMBench, the first fine-grained benchmark specifically designed to evaluate Process-Level Reward Models (PRMs) for medical reasoning rather than general domains.
  • MedPRMBench is built from Clinical Reasoning Blueprints using a three-phase pipeline, generating evaluation data from seven medical QA sources with 14 error types grouped into Simplicity, Soundness, and Sensitivity.
  • It includes a four-level severity grading system to quantify how clinically significant different reasoning failures are, addressing the safety-critical nature of healthcare use.
  • The benchmark contains 6,500 questions (13,000 reasoning chains and 113,910 step-level labels) plus 6,879 training questions, and the authors report a medical PRM baseline achieving an 87.1% overall PRMScore.
  • Using MedPRMBench as a plug-and-play verifier improves downstream medical QA accuracy by 3.2–6.7 percentage points, and evaluations across multiple model types expose common weaknesses in error detection.

Abstract

Process-Level Reward Models (PRMs) are essential for guiding complex reasoning in large language models, yet existing PRM benchmarks cover only general domains such as mathematics, failing to address medical reasoning -- which is uniquely characterized by safety criticality, knowledge intensity, and diverse error patterns. Without a reliable medical PRM evaluation framework, we cannot quantify models' error detection capabilities in clinical reasoning, leaving their safety in real-world healthcare applications unverified. We propose MedPRMBench, the first process-level reward model benchmark for the medical domain. Built through a three-phase pipeline based on Clinical Reasoning Blueprints (CRBs), MedPRMBench systematically generates high-quality evaluation data from seven medical QA sources, covering 14 fine-grained error types across three categories (Simplicity, Soundness, and Sensitivity) with the first 4-level severity grading system to quantify clinical impact. The benchmark comprises 6{,}500 questions with 13{,}000 reasoning chains and 113{,}910 step-level labels, plus 6{,}879 questions for training. Our medical PRM baseline achieves an 87.1\% overall PRMScore -- substantially surpassing all baselines -- and serves as a plug-and-play verifier that improves downstream medical QA accuracy by 3.2--6.7 percentage points. Systematic evaluation spanning proprietary frontier models, open-source reasoning models, and medical-specialized models reveals critical weaknesses in current models' medical reasoning error detection capabilities, providing clear directions for future PRM improvement.