Read the Paper, Write the Code: Agentic Reproduction of Social-Science Results
arXiv cs.AI / 4/27/2026
📰 NewsModels & Research
Key Points
- The paper explores whether LLM-based agents can reproduce social-science findings using only a paper’s methods description and the original data, without access to the original code or the paper itself beyond the extracted methods.
- It introduces an agentic reproduction system that converts methods text into structured instructions, runs reimplementations under strict information isolation, and performs deterministic, cell-level comparisons between reproduced outputs and the published results.
- The system includes an error-attribution step that traces discrepancies across the agent’s pipeline to identify likely root causes of reproduction failures.
- Experiments across four agent scaffolds and four LLMs on 48 human-verified reproducible papers show that agents can often recover published results, but success rates vary widely by model, scaffold, and paper.
- Root-cause analysis indicates that failures arise from both agent-specific mistakes and from missing or ambiguous details (underspecification) in the papers’ methods descriptions.
Related Articles

Subagents: The Building Block of Agentic AI
Dev.to

DeepSeek-V4 Models Could Change Global AI Race
AI Business

Got OpenAI's privacy filter model running on-device via ExecuTorch
Reddit r/LocalLLaMA

The Agent-Skill Illusion: Why Prompt-Based Control Fails in Multi-Agent Business Consulting Systems
Dev.to
We Built a Voice AI Receptionist in 8 Weeks — Every Decision We Made and Why
Dev.to