UNIKIE-BENCH: Benchmarking Large Multimodal Models for Key Information Extraction in Visual Documents
arXiv cs.CV / 4/27/2026
💬 OpinionSignals & Early TrendsModels & Research
Key Points
- UNIKIE-BENCH is a new unified benchmark on arXiv for evaluating Large Multimodal Models (LMMs) on Key Information Extraction (KIE) from real-world visual documents where layout, quality, and task requirements vary widely.
- The benchmark includes two tracks—one with constrained, scenario-defined schemas to match practical use cases, and another open-category track that extracts any explicitly stated key information.
- Experiments across 15 state-of-the-art LMMs show notable performance drops when schema definitions change, when key fields are rare/long-tail, and when document layouts become complex.
- Results also indicate large performance differences across document types and scenarios, highlighting ongoing difficulties in reliable grounding accuracy and layout-aware reasoning for LMM-based KIE.
- The paper provides access to the code and datasets via the project’s GitHub repository to support systematic, repeatable evaluation.
Related Articles

Subagents: The Building Block of Agentic AI
Dev.to
GET Serves Cache, POST Runs Inference: Cost Safety for a Public LLM Endpoint
Dev.to

DeepSeek-V4 Models Could Change Global AI Race
AI Business

Got OpenAI's privacy filter model running on-device via ExecuTorch
Reddit r/LocalLLaMA

The Agent-Skill Illusion: Why Prompt-Based Control Fails in Multi-Agent Business Consulting Systems
Dev.to