RenoBench: A Citation Parsing Benchmark
arXiv cs.CL / 3/27/2026
💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- RenoBench is introduced as a public-domain benchmark for citation parsing, designed to address limitations of prior evaluations (lack of generalizability, reliance on synthetic data, or limited availability).
- The dataset is built from 161,000 annotated citations extracted from PDFs across four publishing ecosystems (SciELO, Redalyc, Public Knowledge Project, and Open Research Europe), producing 10,000 citations with multilingual, multi–publication-type coverage.
- The authors apply automated validation and feature-based sampling to improve dataset quality and representativeness across languages, platforms, and citation formats.
- Experiments evaluate multiple citation parsing systems and report field-level precision/recall, finding that language models perform strongly, especially when fine-tuned.
- RenoBench aims to enable reproducible and standardized evaluation for citation parsing and to support downstream automated citation infrastructure and metascientific research.
広告
Related Articles
![[Boost]](/_next/image?url=https%3A%2F%2Fmedia2.dev.to%2Fdynamic%2Fimage%2Fwidth%3D800%252Cheight%3D%252Cfit%3Dscale-down%252Cgravity%3Dauto%252Cformat%3Dauto%2Fhttps%253A%252F%252Fdev-to-uploads.s3.amazonaws.com%252Fuploads%252Fuser%252Fprofile_image%252F3618325%252F470cf6d0-e54c-4ddf-8d83-e3db9f829f2b.jpg&w=3840&q=75)
[Boost]
Dev.to

Managing LLM context in a real application
Dev.to

Got My 39-Agent System Audited Live. Here's What the Maturity Scorecard Revealed.
Dev.to

OpenAI Killed Sora — Here's Your 10-Minute Migration Guide (Free API)
Dev.to

Switching my AI voice agent from WebSocket to WebRTC — what broke and what I learned
Dev.to