The dataset viewer is not available for this dataset.
Cannot get the config names for the dataset.
Error code:   ConfigNamesError
Exception:    ImportError
Message:      To be able to use universalner/universal_ner, you need to install the following dependency: conllu.
Please install it using 'pip install conllu' for instance.
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/dataset/config_names.py", line 66, in compute_config_names_response
                  config_names = get_dataset_config_names(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 347, in get_dataset_config_names
                  dataset_module = dataset_module_factory(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1914, in dataset_module_factory
                  raise e1 from None
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1880, in dataset_module_factory
                  return HubDatasetModuleFactoryWithScript(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 1504, in get_module
                  local_imports = _download_additional_modules(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/load.py", line 354, in _download_additional_modules
                  raise ImportError(
              ImportError: To be able to use universalner/universal_ner, you need to install the following dependency: conllu.
              Please install it using 'pip install conllu' for instance.

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

Dataset Card for Universal NER

Dataset Summary

Universal NER (UNER) is an open, community-driven initiative aimed at creating gold-standard benchmarks for Named Entity Recognition (NER) across multiple languages. The primary objective of UNER is to offer high-quality, cross-lingually consistent annotations, thereby standardizing and advancing multilingual NER research. UNER v1 includes 19 datasets with named entity annotations, uniformly structured across 13 diverse languages.

Supported Tasks and Leaderboards

  • token-classification: The dataset can be used to train token classification models of the NER variety. Some pre-trained models released as part of the UNER v1 release can be found at https://huggingface.co/universalner

Languages

The dataset contains data in the following languages:

  • Cebuano (ceb)
  • Danish (da)
  • German (de)
  • English (en)
  • Croatian (hr)
  • Portuguese (pt)
  • Russian (ru)
  • Slovak (sk)
  • Serbian (sr)
  • Swedish (sv)
  • Tagalog (tl)
  • Chinese (zh)

Dataset Structure

Data Instances

An example from the UNER_English-PUD test set looks as follows

{
  "idx": "n01016-0002",
  "text": "Several analysts have suggested Huawei is best placed to benefit from Samsung's setback.",
  "tokens": [
    "Several", "analysts", "have", "suggested", "Huawei", 
    "is", "best", "placed", "to", "benefit", 
    "from", "Samsung", "'s", "setback", "."
  ],
  "ner_tags": [
    "O", "O", "O", "O", "B-ORG", 
    "O", "O", "O", "O", "O", 
    "O", "B-ORG", "O", "O", "O"
  ],
  "annotator": "blvns"
}

Data Fields

  • idx: the ID uniquely identifying the sentence (instance), if available.
  • text: the full text of the sentence (instance)
  • tokens: the text of the sentence (instance) split into tokens. Note that this split is inhereted from Universal Dependencies
  • ner_tags: the NER tags associated with each one of the tokens
  • annotator: the annotator who provided the ner_tags for this particular instance

Data Splits

TBD

Dataset Creation

Curation Rationale

TBD

Source Data

Initial Data Collection and Normalization

We selected the Universal Dependency (UD) corpora as the default base texts for annotation due to their extensive language coverage, pre-existing data collection, cleaning, tokenization, and permissive licensing. This choice accelerates our process by providing a robust foundation. By adding another annotation layer to the already detailed UD annotations, we facilitate verification within our project and enable comprehensive multilingual research across the entire NLP pipeline. Given that UD annotations operate at the word level, we adopted the BIO annotation schema (specifically IOB2). In this schema, words forming the beginning (B) or inside (I) part of an entity (X ∈ {PER, LOC, ORG}) are annotated accordingly, while all other words receive an O tag. To maintain consistency, we preserve UD's original tokenization.

Although UD serves as the default data source for UNER, the project is not restricted to UD corpora, particularly for languages not currently represented in UD. The primary requirement for inclusion in the UNER corpus is adherence to the UNER tagging guidelines. Additionally, we are open to converting existing NER efforts on UD treebanks to align with UNER. In this initial release, we have included four datasets transferred from other manual annotation efforts on UD sources (for DA, HR, ARABIZI, and SR).

Who are the source language producers?

This information can be found on per-dataset basis for each of the source Universal Dependencies datasets.

Annotations

Annotation process

The data has been annotated by

Who are the annotators?

For the initial UNER annotation effort, we recruited volunteers from the multilingual NLP community via academic networks and social media. The annotators were coordinated through a Slack workspace, with all contributors working on a voluntary basis. We assume that annotators are either native speakers of the language they annotate or possess a high level of proficiency, although no formal language tests were conducted. The selection of the 13 dataset languages in the first UNER release was driven by the availability of annotators. As the project evolves, we anticipate the inclusion of additional languages and datasets as more annotators become available.

Personal and Sensitive Information

TBD

Considerations for Using the Data

Social Impact of Dataset

TBD

Discussion of Biases

TBD

Other Known Limitations

TBD

Additional Information

Dataset Curators

List the people involved in collecting the dataset and their affiliation(s). If funding information is known, include it here.

Licensing Information

The UNER v1 is released under the terms of the Creative Commons Attribution-ShareAlike 4.0 International license

Citation Information

If you use this dataset, please cite the corresponding paper:

@inproceedings{
  mayhew2024universal,
  title={Universal NER: A Gold-Standard Multilingual Named Entity Recognition Benchmark},
  author={Stephen Mayhew and Terra Blevins and Shuheng Liu and Marek Šuppa and Hila Gonen and Joseph Marvin Imperial and Börje F. Karlsson and Peiqin Lin and Nikola Ljubešić and LJ Miranda and Barbara Plank and Arij Riab and Yuval Pinter}
  booktitle={Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL)},
  year={2024},
  url={https://aclanthology.org/2024.naacl-long.243/}
}
Downloads last month
588