Not Search, But Scan: Benchmarking MLLMs on Scan-Oriented Academic Paper Reasoning

arXiv cs.AI / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current academic-paper reasoning benchmarks are mostly search-oriented and therefore do not capture researcher-style full-document understanding, cross-checking, and verification.
  • It introduces ScholScan, a new scan-oriented benchmark that tasks multimodal LLMs with reading entire papers and identifying consistency issues.
  • ScholScan includes 1,800 annotated questions across nine error categories, covering 13 natural-science domains and 715 papers, with evidence localization and reasoning traces plus a unified evaluation protocol.
  • Experiments with 15 models across 24 input settings show that retrieval-augmented generation (RAG) does not yield significant gains, highlighting systematic weaknesses on scan-oriented tasks.
  • The authors position ScholScan as the leading representative benchmark for the proposed scan-oriented paradigm in academic paper reasoning.

Abstract

With the rapid progress of multimodal large language models (MLLMs), AI already performs well at literature retrieval and certain reasoning tasks, serving as a capable assistant to human researchers, yet it remains far from autonomous research. The fundamental reason is that current work on academic paper reasoning is largely confined to a search-oriented paradigm centered on pre-specified targets, with reasoning grounded in relevance retrieval, which struggles to support researcher-style full-document understanding, reasoning, and verification. To bridge this gap, we propose \textbf{ScholScan}, a new benchmark for academic paper reasoning. ScholScan introduces a scan-oriented task setting that asks models to read and cross-check entire papers like human researchers, scanning the document to identify consistency issues. The benchmark comprises 1,800 carefully annotated questions drawn from nine error categories across 13 natural-science domains and 715 papers, and provides detailed annotations for evidence localization and reasoning traces, together with a unified evaluation protocol. We assessed 15 models across 24 input configurations and conducted a fine-grained analysis of MLLM capabilities for all error categories. Across the board, retrieval-augmented generation (RAG) methods yield no significant improvements, revealing systematic deficiencies of current MLLMs on scan-oriented tasks and underscoring the challenge posed by ScholScan. We expect ScholScan to be the leading and representative work of the scan-oriented task paradigm.