Can LLM Agents Generate Real-World Evidence? Evaluating Observational Studies in Medical Databases

arXiv cs.AI / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper evaluates whether LLM agents can generate real-world evidence end-to-end in medical databases by reproducing observational studies using database execution plus coherent reporting rather than isolated QA steps.
  • It introduces RWE-bench, a benchmark built from MIMIC-IV and peer-reviewed observational studies, where agents receive the study protocol as the reference standard and must produce tree-structured evidence bundles.
  • Across 162 tasks with six LLMs and three different agent scaffolds, overall task success is low, with the best agent at 39.9% and the best open-source model at 30.4%.
  • Agent scaffold selection significantly affects outcomes, driving over 30% variation in performance, indicating that workflow design is a key determinant of results.
  • The authors also add an automated cohort evaluation method to pinpoint error locations and characterize agent failure modes, and conclude that efficient validation is a major open direction.

Abstract

Observational studies can yield clinically actionable evidence at scale, but executing them on real-world databases is open-ended and requires coherent decisions across cohort construction, analysis, and reporting. Prior evaluations of LLM agents emphasize isolated steps or single answers, missing the integrity and internal structure of the resulting evidence bundle. To address this gap, we introduce RWE-bench, a benchmark grounded in MIMIC-IV and derived from peer-reviewed observational studies. Each task provides the corresponding study protocol as the reference standard, requiring agents to execute experiments in a real database and iteratively generate tree-structured evidence bundles. We evaluate six LLMs (three open-source, three closed-source) under three agent scaffolds using both question-level correctness and end-to-end task metrics. Across 162 tasks, task success is low: the best agent reaches 39.9%, and the best open-source model reaches 30.4%. Agent scaffolds also matter substantially, causing over 30% variation in performance metrics. Furthermore, we implement an automated cohort evaluation method to rapidly localize errors and identify agent failure modes. Overall, the results highlight persistent limitations in agents' ability to produce end-to-end evidence bundles, and efficient validation remains an important direction for future work. Code and data are available at https://github.com/somewordstoolate/RWE-bench.