File size: 4,275 Bytes
e34d040
 
 
 
 
f3a6d99
e34d040
f3a6d99
e34d040
 
 
 
 
 
 
 
 
 
 
 
 
 
751abc7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e34d040
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
382b6d2
e34d040
 
382b6d2
e34d040
3170039
 
 
e34d040
 
 
382b6d2
e34d040
 
 
382b6d2
e34d040
 
 
 
 
 
 
 
 
6c7216b
 
3170039
e34d040
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
382b6d2
e34d040
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
751abc7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
---
annotations_creators:
- no-annotation
language_creators:
- found
language:
- en
license:
- mit
multilinguality:
- monolingual
pretty_name: enwik8
size_categories:
- 10K<n<100K
source_datasets:
- original
task_categories:
- fill-mask
- text-generation
task_ids:
- language-modeling
- masked-language-modeling
dataset_info:
- config_name: enwik8
  features:
  - name: text
    dtype: string
  splits:
  - name: train
    num_bytes: 104299244
    num_examples: 1128024
  download_size: 36445475
  dataset_size: 102383126
- config_name: enwik8-raw
  features:
  - name: text
    dtype: string
  splits:
  - name: train
    num_bytes: 100000008
    num_examples: 1
  download_size: 36445475
  dataset_size: 100000008
---

# Dataset Card for enwik8

## Table of Contents
- [Dataset Description](#dataset-description)
  - [Dataset Summary](#dataset-summary)
  - [Supported Tasks](#supported-tasks-and-leaderboards)
  - [Languages](#languages)
- [Dataset Structure](#dataset-structure)
  - [Data Instances](#data-instances)
  - [Data Fields](#data-instances)
  - [Data Splits](#data-instances)
- [Dataset Creation](#dataset-creation)
  - [Curation Rationale](#curation-rationale)
  - [Source Data](#source-data)
  - [Annotations](#annotations)
  - [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
  - [Social Impact of Dataset](#social-impact-of-dataset)
  - [Discussion of Biases](#discussion-of-biases)
  - [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
  - [Dataset Curators](#dataset-curators)
  - [Licensing Information](#licensing-information)
  - [Citation Information](#citation-information)

## Dataset Description

- **Homepage:** http://mattmahoney.net/dc/textdata.html
- **Repository:** [Needs More Information]
- **Paper:** [Needs More Information]
- **Leaderboard:** https://paperswithcode.com/sota/language-modelling-on-enwiki8
- **Point of Contact:** [Needs More Information]
- **Size of downloaded dataset files:** 36.45 MB
- **Size of the generated dataset:** 102.38 MB
- **Total amount of disk used:** 138.83 MB

### Dataset Summary

The enwik8 dataset is the first 100,000,000 (100M) bytes of the English Wikipedia XML dump on Mar. 3, 2006 and is typically used to measure a model's ability to compress data.

### Supported Tasks and Leaderboards

A leaderboard for byte-level causal language modelling can be found on [paperswithcode](https://paperswithcode.com/sota/language-modelling-on-enwiki8)

### Languages

en

## Dataset Structure

### Data Instances

- **Size of downloaded dataset files:** 36.45 MB
- **Size of the generated dataset:** 102.38 MB
- **Total amount of disk used:** 138.83 MB

```
{
   "text": "In [[Denmark]], the [[Freetown Christiania]] was created in downtown [[Copenhagen]]....",
}
```

### Data Fields

The data fields are the same among all sets.

#### enwik8

- `text`: a `string` feature.

#### enwik8-raw

- `text`: a `string` feature.

### Data Splits

| dataset     | train   |
| ---         | ---     |
| enwik8      | 1128024	|
| enwik8- raw | 1       |

## Dataset Creation

### Curation Rationale

[Needs More Information]

### Source Data

#### Initial Data Collection and Normalization

The data is just English Wikipedia XML dump on Mar. 3, 2006 split by line for enwik8 and not split by line for enwik8-raw.

#### Who are the source language producers?

[Needs More Information]

### Annotations

#### Annotation process

[Needs More Information]

#### Who are the annotators?

[Needs More Information]

### Personal and Sensitive Information

[Needs More Information]

## Considerations for Using the Data

### Social Impact of Dataset

[Needs More Information]

### Discussion of Biases

[Needs More Information]

### Other Known Limitations

[Needs More Information]

## Additional Information

### Dataset Curators

[Needs More Information]

### Licensing Information

[Needs More Information]

### Citation Information

Dataset is not part of a publication, and can therefore not be cited.

### Contributions

Thanks to [@HallerPatrick](https://github.com/HallerPatrick) for adding this dataset and [@mtanghu](https://github.com/mtanghu) for updating it.