ReactBench: A Benchmark for Topological Reasoning in MLLMs on Chemical Reaction Diagrams
arXiv cs.AI / 4/20/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces ReactBench, a new benchmark designed to test structural (topological) reasoning in multimodal LLMs using chemical reaction diagrams rather than only semantic understanding of visuals.
- It targets model weaknesses on complex graph structures such as branching paths, converging flows, and cyclic dependencies, including simple counting endpoint tasks.
- The benchmark contains 1,618 expert-annotated QA pairs across four hierarchical task dimensions, enabling evaluation from localized recognition to holistic structural reasoning.
- Experiments across 17 MLLMs show a performance drop of over 30% from anchor-based tasks to holistic structural reasoning tasks, indicating a bottleneck in reasoning rather than perception.
- Ablation studies support the conclusion that the limitation is fundamentally about structural understanding, and the results suggest directions for improving visual/topological reasoning.
Related Articles
Awesome Open-Weight Models: The Practitioner's Guide to Open-Source LLMs (2026 Edition) [P]
Reddit r/MachineLearning
The Mythos vs GPT-5.4-Cyber debate is missing the benchmark
Dev.to
Beyond the Crop: Automating "Ghost Mannequin" Effects with Depth-Aware Inpainting
Dev.to
The $20/month AI subscription is gaslighting developers in emerging markets
Dev.to
A Claude Code hook that warns you before calling a low-trust MCP server
Dev.to