EpiBench: Benchmarking Multi-turn Research Workflows for Multimodal Agents

arXiv cs.CL / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces EpiBench, a new episodic multi-turn multimodal benchmark designed to evaluate research agents that conduct proactive literature search and sustained evidence use over multiple turns.
  • Tasks require agents to navigate across multiple papers, extract and align evidence from figures and tables, and then use accumulated memory to answer objective questions involving cross-paper comparisons and multi-figure integration.
  • The authors propose a process-level evaluation framework aimed at fine-grained testing and diagnosis of how research agents perform throughout the workflow (not just final answers).
  • Experimental results show even leading models achieve only 29.23% accuracy on the hard split, highlighting significant gaps in current capabilities for multi-step, multi-evidence scientific reasoning.

Abstract

Scientific research follows multi-turn, multi-step workflows that require proactively searching the literature, consulting figures and tables, and integrating evidence across papers to align experimental settings and support reproducible conclusions. This joint capability is not systematically assessed in existing benchmarks, which largely under-evaluate proactive search, multi-evidence integration and sustained evidence use over time. In this work, we introduce EpiBench, an episodic multi-turn multimodal benchmark that instantiates short research workflows. Given a research task, agents must navigate across papers over multiple turns, align evidence from figures and tables, and use the accumulated evidence in the memory to answer objective questions that require cross paper comparisons and multi-figure integration. EpiBench introduces a process-level evaluation framework for fine-grained testing and diagnosis of research agents. Our experiments show that even the leading model achieves an accuracy of only 29.23% on the hard split, indicating substantial room for improvement in multi-turn, multi-evidence research workflows, providing an evaluation platform for verifiable and reproducible research agents.