Davlan's picture
Update README.md
license: afl-3.0
  - am
  - bm
  - obj
  - ee
  - fon
  - ha
  - ig
  - rw
  - lg
  - luo
  - mos
  - ny
  - pcm
  - sn
  - sw
  - tn
  - tw
  - wo
  - xh
  - yo
  - zu
  - multilingual
  - masakhaner2


Model description

masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0 is a Named Entity Recognition (NER) model for 21 African languages. Specifically, this model is a Davlan/afro-xlmr-large model that was fine-tuned on an aggregation of African language datasets obtained from two versions of MasakhaNER dataset i.e. MasakhaNER 1.0 and MasakhaNER 2.0. The languages covered are:

  • Amharic (Amharic)
  • Bambara (bam)
  • Ghomala (bbj)
  • Ewe (ewe)
  • Fon (fon)
  • Hausa (hau)
  • Igbo (ibo)
  • Kinyarwanda (kin)
  • Luganda (lug)
  • Dholuo (luo) -Mossi (mos)
  • Chichewa (nya)
  • Nigerian Pidgin
  • chShona (sna)
  • Kiswahili (swą)
  • Setswana (tsn)
  • Twi (twi)
  • Wolof (wol)
  • isiXhosa (xho)
  • Yorùbá (yor)
  • isiZulu (zul)

It has been trained to recognize four types of entities: dates & times (DATE), location (LOC), organization (ORG), and person (PER).

Intended uses & limitations

How to use

You can use this model with Transformers pipeline for NER.

from transformers import AutoTokenizer, AutoModelForTokenClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0")
model = AutoModelForTokenClassification.from_pretrained("masakhane/afroxlmr-large-ner-masakhaner-1.0_2.0")
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
example = "Emir of Kano turban Zhang wey don spend 18 years for Nigeria"
ner_results = nlp(example)

Eval results on MasakhaNER (F-score)

Model evaluated on MasakhaNER 1.0 and MasakhaNER 2.0 test sets

language MasakhaNER 1.0 MasakhaNER 2.0
amh 80.5
bam 83.1
bbj 76.6
ewe 89.6
fon 83.8
hau 90.3 87.5
ibo 89.5 93.5
kin 82.0 87.6
lug 87.1 89.7
luo 80.8 82.5
mos 75.5
nya 92.7
pcm 91.1 90.9
sna 96.5
swa 88.5 93.4
tsn 90.3
twi 81.3
wol 72.7 87.3
xho 90.0
yor 88.1 90.5
zul 91.3
avg 85.1 87.7

Limitations and bias

This model is limited by its training dataset of entity-annotated news articles from a specific span of time. This may not generalize well for all use cases in different domains.

Training data

This model was fine-tuned on the aggregation of MasakhaNER 1.0 and MasakhaNER 2.0 datasets

The training dataset distinguishes between the beginning and continuation of an entity so that if there are back-to-back entities of the same type, the model can output where the second entity begins. As in the dataset, each token will be classified as one of the following classes:

Abbreviation Description
O Outside of a named entity
B-DATE Beginning of a DATE entity right after another DATE entity
I-DATE DATE entity
B-PER Beginning of a person’s name right after another person’s name
I-PER Person’s name
B-ORG Beginning of an organisation right after another organisation
I-ORG Organisation
B-LOC Beginning of a location right after another location
I-LOC Location

Training procedure

This model was trained on a single NVIDIA V100 GPU with recommended hyperparameters from the original MasakhaNER paper which trained & evaluated the model on MasakhaNER corpus.

BibTeX entry and citation info

  title={MasakhaNER 2.0: Africa-centric Transfer Learning for Named Entity Recognition},
  author={David Ifeoluwa Adelani and Graham Neubig and Sebastian Ruder and Shruti Rijhwani and Michael Beukman and Chester Palen-Michel and Constantine Lignos and Jesujoba Oluwadara Alabi and Shamsuddeen Hassan Muhammad and Peter Nabende and Cheikh M. Bamba Dione and Andiswa Bukula and Rooweither Mabuya and Bonaventure F. P. Dossou and Blessing K. Sibanda and Happy Buzaaba and Jonathan Mukiibi and Godson Kalipe and Derguene Mbaye and Amelia Taylor and Fatoumata Kabore and Chris C. Emezue and Anuoluwapo Aremu and Perez Ogayo and Catherine W. Gitau and Edwin Munkoh-Buabeng and Victoire Memdjokam Koagne and Allahsera Auguste Tapo and Tebogo Macucwa and Vukosi Marivate and Elvis Mboning and Tajuddeen R. Gwadabe and Tosin P. Adewumi and Orevaoghene Ahia and Joyce Nakatumba-Nabende and Neo L. Mokono and Ignatius M Ezeani and Chiamaka Ijeoma Chukwuneke and Mofetoluwa Adeyemi and Gilles Hacheme and Idris Abdulmumin and Odunayo Ogundepo and Oreen Yousuf and Tatiana Moteu Ngoli and Dietrich Klakow},