Michael Beukman commited on
Commit
da81d7b
1 Parent(s): 364741c

Fixed a typo.

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -21,7 +21,7 @@ More information, and other similar models can be found in the [main Github repo
21
 
22
  ## About
23
  This models is transformer based and was fine-tuned on the MasakhaNER dataset. It is a named entity recognition dataset, containing mostly news articles in 10 different African languages.
24
- The model was fine-tuned for 50 epochs, with a maximum sequence length of 200, 32 batch size, 5e-5 learning rate. This process was repeated 5 times (with different random seeds), and this uploaded model performed the best out of those 5 seeds (aggregate F1 on on test set).
25
 
26
  This model was fine-tuned by me, Michael Beukman while doing a project at the University of the Witwatersrand, Johannesburg. This is version 1, as of 20 November 2021.
27
  This models is licensed under the [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).
@@ -110,7 +110,7 @@ tokenizer = AutoTokenizer.from_pretrained(model_name)
110
  model = AutoModelForTokenClassification.from_pretrained(model_name)
111
 
112
  nlp = pipeline("ner", model=model, tokenizer=tokenizer)
113
- example = "A (Swahili) sentence that may contain entities"
114
 
115
  ner_results = nlp(example)
116
  print(ner_results)
 
21
 
22
  ## About
23
  This models is transformer based and was fine-tuned on the MasakhaNER dataset. It is a named entity recognition dataset, containing mostly news articles in 10 different African languages.
24
+ The model was fine-tuned for 50 epochs, with a maximum sequence length of 200, 32 batch size, 5e-5 learning rate. This process was repeated 5 times (with different random seeds), and this uploaded model performed the best out of those 5 seeds (aggregate F1 on test set).
25
 
26
  This model was fine-tuned by me, Michael Beukman while doing a project at the University of the Witwatersrand, Johannesburg. This is version 1, as of 20 November 2021.
27
  This models is licensed under the [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).
 
110
  model = AutoModelForTokenClassification.from_pretrained(model_name)
111
 
112
  nlp = pipeline("ner", model=model, tokenizer=tokenizer)
113
+ example = "Wizara ya afya ya Tanzania imeripoti Jumatatu kuwa , watu takriban 14 zaidi wamepata maambukizi ya Covid - 19 ."
114
 
115
  ner_results = nlp(example)
116
  print(ner_results)