PRBench: End-to-end Paper Reproduction in Physics Research

arXiv cs.CL / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • PRBench is introduced as a benchmark of 30 expert-curated physics paper reproduction tasks across 11 subfields, requiring agents to implement algorithms from scratch and reproduce quantitative results.
  • Each task gives agents only the paper content plus instructions, and they must run in a sandboxed environment while matching validated ground-truth outcomes with detailed scoring rubrics.
  • Evaluation of coding agents via an agentified assessment pipeline finds that the top system (OpenAI Codex using GPT-5.3-Codex) reaches only a 34% mean overall score, indicating limited reliability for end-to-end reproduction.
  • All tested agents show a zero end-to-end callback success rate, with especially poor performance in data accuracy and code correctness.
  • The study identifies recurring failure modes such as incorrect formula-to-code implementation, inability to debug numerical simulations, and even fabrication of output data.

Abstract

AI agents powered by large language models exhibit strong reasoning and problem-solving capabilities, enabling them to assist scientific research tasks such as formula derivation and code generation. However, whether these agents can reliably perform end-to-end reproduction from real scientific papers remains an open question. We introduce PRBench, a benchmark of 30 expert-curated tasks spanning 11 subfields of physics. Each task requires an agent to comprehend the methodology of a published paper, implement the corresponding algorithms from scratch, and produce quantitative results matching the original publication. Agents are provided only with the task instruction and paper content, and operate in a sandboxed execution environment. All tasks are contributed by domain experts from over 20 research groups at the School of Physics, Peking University, each grounded in a real published paper and validated through end-to-end reproduction with verified ground-truth results and detailed scoring rubrics. Using an agentified assessment pipeline, we evaluate a set of coding agents on PRBench and analyze their capabilities across key dimensions of scientific reasoning and execution. The best-performing agent, OpenAI Codex powered by GPT-5.3-Codex, achieves a mean overall score of 34%. All agents exhibit a zero end-to-end callback success rate, with particularly poor performance in data accuracy and code correctness. We further identify systematic failure modes, including errors in formula implementation, inability to debug numerical simulations, and fabrication of output data. Overall, PRBench provides a rigorous benchmark for evaluating progress toward autonomous scientific research.