SciMDR: Benchmarking and Advancing Scientific Multimodal Document Reasoning
arXiv cs.CL / 3/13/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The authors propose a two-stage synthesize-and-reground framework to generate faithful reasoning data for scientific multimodal documents.
- They build SciMDR, a large-scale dataset with 300K QA pairs across 20K papers, plus SciMDR-Eval for expert-annotated benchmarks.
- Experiments show models fine-tuned on SciMDR achieve significant gains on scientific QA benchmarks, especially for complex document-level reasoning.
- The work addresses the trade-off among scale, faithfulness, and realism in creating datasets for foundation-model training.
Related Articles

報告:LLMにおける「自己言及的再帰」と「ステートフル・エミュレーション」の観測
note

フリーランスの泥臭い経験を資産に変える。AIの文章に「あなたの魂」を注入する技術。【コピペOK】
note

Hey dev.to community – sharing my journey with Prompt Builder, Insta Posts, and practical SEO
Dev.to

How to Build Passive Income with AI in 2026: A Developer's Practical Guide
Dev.to

The Research That Doesn't Exist
Dev.to