File size: 3,421 Bytes
efaccae
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d83da96
9382ba6
 
d83da96
9382ba6
 
efaccae
ad7c2a0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a9e9f2d
ad7c2a0
a9e9f2d
c7ebed6
7312548
c7ebed6
 
ad7c2a0
0def4f8
 
 
 
ad7c2a0
 
 
e601542
ad7c2a0
efaccae
 
ad7c2a0
d1de3cd
 
 
 
 
 
 
 
 
 
 
 
ad7c2a0
efaccae
 
2f32017
efaccae
 
 
 
2f32017
efaccae
 
ad7c2a0
 
 
 
 
 
fb79dee
ad7c2a0
 
 
efaccae
 
 
 
 
 
ad7c2a0
efaccae
 
ad7c2a0
efaccae
ad7c2a0
efaccae
 
 
 
 
ad7c2a0
 
efaccae
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ad7c2a0
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
---
language: 
- om
- am
- rw
- rn
- ha
- ig
- pcm
- so
- sw
- ti
- yo
- multilingual
    
license: apache-2.0

task_categories:
- text-generation
task_ids:
- language-modeling
---

# Dataset Card for AfriBERTa's Corpus

## Table of Contents
- [Dataset Description](#dataset-description)
  - [Dataset Summary](#dataset-summary)
  - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
  - [Languages](#languages)
  - [Loading Dataset](#loading-dataset)
- [Dataset Structure](#dataset-structure)
  - [Data Instances](#data-instances)
  - [Data Fields](#data-fields)
  - [Data Splits](#data-splits)
- [Considerations for Using the Data](#considerations-for-using-the-data)
  - [Discussion of Biases](#discussion-of-biases)
- [Additional Information](#additional-information)
  - [Citation Information](#citation-information)
  - [Contributions](#contributions)

## Dataset Description

### Dataset Summary
This is the corpus on which AfriBERTa was trained on.
The dataset is mostly from the BBC news website, but some languages also have data from Common Crawl. 

- **Homepage:** https://github.com/keleog/afriberta
- **Models:** 
    - https://huggingface.co/castorini/afriberta_small
    - https://huggingface.co/castorini/afriberta_base
    - https://huggingface.co/castorini/afriberta_large
- **Paper:** https://aclanthology.org/2021.mrl-1.11/
- **Point of Contact:** kelechi.ogueji@uwaterloo.ca


### Supported Tasks and Leaderboards
The AfriBERTa corpus was mostly intended to pre-train language models.

### Languages
```
afaanoromoo
amharic
gahuza
hausa
igbo
pidgin
somali
swahili
tigrinya
yoruba
```
### Loading Dataset
An example to load the train split of the Somali corpus:
```
dataset = load_dataset("castorini/afriberta-corpus", "somali", split="train")
```

An example to load the test split of the Pidgin corpus:
```
dataset = load_dataset("castorini/afriberta-corpus", "pidgin", split="test")
```

## Dataset Structure

### Data Instances
Each data point is a line of text. 
An example from the `igbo` dataset:
```
{"id": "6", "text": "Ngwá ọrụ na-echebe ma na-ebuli gị na kọmputa."}
```

### Data Fields

The data fields are:

- id: id of the example
- text: content as a string

### Data Splits
Each language has a train and test split, with varying sizes. 

## Considerations for Using the Data

### Discussion of Biases
Since majority of the data is obtained from the BBC's news website, models trained on this dataset are likely going to 
be biased towards the news domain.

Also, since some of the data is obtained from Common Crawl, care should be taken (especially for text generation models) since personal and sensitive information might be present.

## Additional Information
### Citation Information
```
@inproceedings{ogueji-etal-2021-small,
    title = "Small Data? No Problem! Exploring the Viability of Pretrained Multilingual Language Models for Low-resourced Languages",
    author = "Ogueji, Kelechi  and
      Zhu, Yuxin  and
      Lin, Jimmy",
    booktitle = "Proceedings of the 1st Workshop on Multilingual Representation Learning",
    month = nov,
    year = "2021",
    address = "Punta Cana, Dominican Republic",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2021.mrl-1.11",
    pages = "116--126",
}
```

### Contributions
Thanks to [Kelechi Ogueji](https://github.com/keleog) for adding this dataset.