Search papers, labs, and topics across Lattice.
The Universal NER v2 project expands upon the original UNER v1 dataset to create a massively multilingual Named Entity Recognition benchmark, addressing the scarcity of gold-standard evaluation resources for non-English languages. This benchmark employs a general tagset and annotation guidelines to ensure standardized, cross-lingual annotations of named entity spans. The project aims to facilitate more robust evaluation of multilingual language models and promote research in low-resource languages.
Massively multilingual NER just got easier: UNER v2 offers a standardized benchmark for evaluating LLMs across diverse languages.
While multilingual language models promise to bring the benefits of LLMs to speakers of many languages, gold-standard evaluation benchmarks in most languages to interrogate these assumptions remain scarce. The Universal NER project, now entering its fourth year, is dedicated to building gold-standard multilingual Named Entity Recognition (NER) benchmark datasets. Inspired by existing massively multilingual efforts for other core NLP tasks (e.g., Universal Dependencies), the project uses a general tagset and thorough annotation guidelines to collect standardized, cross-lingual annotations of named entity spans. The first installment (UNER v1) was released in 2024, and the project has continued and expanded since then, with various organizers, annotators, and collaborators in an active community.