GorkaUrbizu
commited on
Commit
•
13256a9
1
Parent(s):
cf4177e
Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,14 @@ BERT medium (cased) model trained on a subset of 125M tokens of cc100-Swahili fo
|
|
10 |
The model has 51M parameters (8L), and a vocab size of 50K.
|
11 |
It was trained for 500K steps with a sequence length of 512 tokens.
|
12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
|
14 |
Authors
|
15 |
-----------
|
|
|
10 |
The model has 51M parameters (8L), and a vocab size of 50K.
|
11 |
It was trained for 500K steps with a sequence length of 512 tokens.
|
12 |
|
13 |
+
Results
|
14 |
+
-----------
|
15 |
+
| | [bert-base-sw](https://huggingface.co/orai-nlp/bert-base-sw) | [bert-medium-sw](https://huggingface.co/orai-nlp/bert-medium-sw) | Flair | [mBERT](https://huggingface.co/bert-base-multilingual-cased) | [swahBERT](https://github.com/gatimartin/SwahBERT#pre-trained-models) (Martin et al., 2022b) |
|
16 |
+
|-----------|--------------|----------------|-------|-------|---------------------------------|
|
17 |
+
| NERC | **92.09** | 91.63 | 92.04 | 91.17 | 88.60 |
|
18 |
+
| Topic | **93.07** | 92.88 | 91.83 | 91.52 | 90.90 |
|
19 |
+
| Sentiment | **79.04** | 77.07 | 73.60 | 69.17 | 71.12 |
|
20 |
+
| QNLI | 63.34 | 63.87 | 52.82 | 63.48 | **64.72** |
|
21 |
|
22 |
Authors
|
23 |
-----------
|