Update README.md
Browse files
README.md
CHANGED
@@ -1,20 +1,22 @@
|
|
1 |
-
Hugging Face's logo
|
2 |
---
|
3 |
-
language:
|
4 |
- om
|
5 |
- am
|
6 |
- rw
|
7 |
- rn
|
8 |
- ha
|
9 |
- ig
|
10 |
-
- pcm
|
11 |
- so
|
12 |
- sw
|
13 |
- ti
|
14 |
- yo
|
|
|
15 |
- multilingual
|
16 |
-
|
|
|
|
|
17 |
---
|
|
|
18 |
# afriberta_large
|
19 |
## Model description
|
20 |
AfriBERTa large is a pretrained multilingual language model with around 126 million parameters.
|
@@ -63,6 +65,4 @@ For information on training procedures, please refer to the AfriBERTa [paper]()
|
|
63 |
url = "https://aclanthology.org/2021.mrl-1.11",
|
64 |
pages = "116--126",
|
65 |
}
|
66 |
-
```
|
67 |
-
|
68 |
-
|
|
|
|
|
1 |
---
|
2 |
+
language:
|
3 |
- om
|
4 |
- am
|
5 |
- rw
|
6 |
- rn
|
7 |
- ha
|
8 |
- ig
|
|
|
9 |
- so
|
10 |
- sw
|
11 |
- ti
|
12 |
- yo
|
13 |
+
- pcm
|
14 |
- multilingual
|
15 |
+
license: mit
|
16 |
+
datasets:
|
17 |
+
- castorini/afriberta-corpus
|
18 |
---
|
19 |
+
|
20 |
# afriberta_large
|
21 |
## Model description
|
22 |
AfriBERTa large is a pretrained multilingual language model with around 126 million parameters.
|
|
|
65 |
url = "https://aclanthology.org/2021.mrl-1.11",
|
66 |
pages = "116--126",
|
67 |
}
|
68 |
+
```
|
|
|
|