Michael Beukman
commited on
Commit
•
da81d7b
1
Parent(s):
364741c
Fixed a typo.
Browse files
README.md
CHANGED
@@ -21,7 +21,7 @@ More information, and other similar models can be found in the [main Github repo
|
|
21 |
|
22 |
## About
|
23 |
This models is transformer based and was fine-tuned on the MasakhaNER dataset. It is a named entity recognition dataset, containing mostly news articles in 10 different African languages.
|
24 |
-
The model was fine-tuned for 50 epochs, with a maximum sequence length of 200, 32 batch size, 5e-5 learning rate. This process was repeated 5 times (with different random seeds), and this uploaded model performed the best out of those 5 seeds (aggregate F1 on
|
25 |
|
26 |
This model was fine-tuned by me, Michael Beukman while doing a project at the University of the Witwatersrand, Johannesburg. This is version 1, as of 20 November 2021.
|
27 |
This models is licensed under the [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).
|
@@ -110,7 +110,7 @@ tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
110 |
model = AutoModelForTokenClassification.from_pretrained(model_name)
|
111 |
|
112 |
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
|
113 |
-
example = "
|
114 |
|
115 |
ner_results = nlp(example)
|
116 |
print(ner_results)
|
|
|
21 |
|
22 |
## About
|
23 |
This models is transformer based and was fine-tuned on the MasakhaNER dataset. It is a named entity recognition dataset, containing mostly news articles in 10 different African languages.
|
24 |
+
The model was fine-tuned for 50 epochs, with a maximum sequence length of 200, 32 batch size, 5e-5 learning rate. This process was repeated 5 times (with different random seeds), and this uploaded model performed the best out of those 5 seeds (aggregate F1 on test set).
|
25 |
|
26 |
This model was fine-tuned by me, Michael Beukman while doing a project at the University of the Witwatersrand, Johannesburg. This is version 1, as of 20 November 2021.
|
27 |
This models is licensed under the [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).
|
|
|
110 |
model = AutoModelForTokenClassification.from_pretrained(model_name)
|
111 |
|
112 |
nlp = pipeline("ner", model=model, tokenizer=tokenizer)
|
113 |
+
example = "Wizara ya afya ya Tanzania imeripoti Jumatatu kuwa , watu takriban 14 zaidi wamepata maambukizi ya Covid - 19 ."
|
114 |
|
115 |
ner_results = nlp(example)
|
116 |
print(ner_results)
|