Universal NER v2: Towards a Massively Multilingual Named Entity Recognition Benchmark

arXiv cs.CL / 4/15/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Universal NER v2 aims to expand and refine gold-standard, massively multilingual Named Entity Recognition (NER) benchmark datasets to better evaluate multilingual language models across many languages.
  • The project builds standardized cross-lingual NER annotations using a general tagset and detailed annotation guidelines, inspired by similar efforts such as Universal Dependencies.
  • Universal NER has been running for multiple years, with an initial release (UNER v1) in 2024 and continued community contributions from organizers, annotators, and collaborators.
  • The work targets a key gap: the scarcity of high-quality evaluation benchmarks for most languages that can test assumptions behind multilingual LLM benefits.

Abstract

While multilingual language models promise to bring the benefits of LLMs to speakers of many languages, gold-standard evaluation benchmarks in most languages to interrogate these assumptions remain scarce. The Universal NER project, now entering its fourth year, is dedicated to building gold-standard multilingual Named Entity Recognition (NER) benchmark datasets. Inspired by existing massively multilingual efforts for other core NLP tasks (e.g., Universal Dependencies), the project uses a general tagset and thorough annotation guidelines to collect standardized, cross-lingual annotations of named entity spans. The first installment (UNER v1) was released in 2024, and the project has continued and expanded since then, with various organizers, annotators, and collaborators in an active community.