Looking for OCR for AI papers (math-heavy PDFs) — FireRed-OCR vs DeepSeek-OCR vs MonkeyOCR?

Reddit r/LocalLLaMA / 3/29/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The post asks for recommendations on OCR tools to extract structured content from math-heavy AI research PDFs (especially arXiv), where dense formulas, multi-column layouts, and tables make plain OCR insufficient.
  • The author is comparing FireRed-OCR, DeepSeek-OCR, and MonkeyOCR and wants to know which ones best preserve structure and mathematical content rather than just producing rough transcription.
  • The writer is considering building a small benchmark using ~20 recent arXiv papers with varied layouts, evaluating extraction accuracy for text, formulas, and tables as well as the required post-processing effort.
  • The discussion emphasizes practical workflow needs like consistent reading order and formula/layout robustness, not just image-to-text conversion quality.
  • Overall, the request is a crowdsourced search for OCR systems suited to academic, layout-complex technical documents for faster reading, indexing, and note-taking.

Right now I’m trying to build a workflow for extracting content from recent AI research papers (mostly arXiv PDFs) so I can speed up reading, indexing, and note-taking.

The catch is: these papers are not “clean text” documents. They usually include:

  • Dense mathematical formulas (often LaTeX-heavy)
  • Multi-column layouts
  • Complex tables
  • Figures/diagrams embedded with captions
  • Mixed reading order issues

So for me, plain OCR accuracy is not enough—I care a lot about structure + formulas + layout consistency.

I’ve been experimenting and reading about some projects, such as:

FireRed-OCR

Looks promising for document-level OCR with better structure awareness. I’ve seen people mention it performs reasonably well on complex layouts, though I’m still unclear how robust it is on heavy math-heavy papers.

DeepSeek-OCR

Interesting direction, especially with the broader DeepSeek ecosystem pushing multimodal understanding. Curious if anyone has used it specifically for academic PDFs with formulas—does it actually preserve LaTeX-quality output or is it more “semantic transcription”?

MonkeyOCR

This one caught my attention because it seems lightweight and relatively easy to deploy. But I’m not sure how it performs on scientific papers vs more general document OCR.

I’m thinking of running a small benchmark myself by selecting around 20 recent arXiv papers with different layouts and comparing how well each model extracts plain text, formulas, and tables, while also measuring both accuracy and the amount of post-processing effort required.

Could you guys take a look at the models below and let me know which ones are actually worth testing?

submitted by /u/still_debugging_note
[link] [comments]