Datasets:

Languages:
English
Multilinguality:
monolingual
Size Categories:
1M<n<10M
Language Creators:
crowdsourced
Annotations Creators:
no-annotation
Source Datasets:
original
ArXiv:
Tags:
License:
albertvillanova HF staff commited on
Commit
5fddba4
1 Parent(s): c8a1c90

Add wikitext-103-raw-v1 data files

Browse files
README.md CHANGED
@@ -29,16 +29,16 @@ dataset_info:
29
  dtype: string
30
  splits:
31
  - name: test
32
- num_bytes: 1305092
33
  num_examples: 4358
34
  - name: train
35
- num_bytes: 546501673
36
  num_examples: 1801350
37
  - name: validation
38
- num_bytes: 1159292
39
  num_examples: 3760
40
- download_size: 191984949
41
- dataset_size: 548966057
42
  - config_name: wikitext-103-v1
43
  features:
44
  - name: text
@@ -88,6 +88,14 @@ dataset_info:
88
  download_size: 7371282
89
  dataset_size: 13323188
90
  configs:
 
 
 
 
 
 
 
 
91
  - config_name: wikitext-103-v1
92
  data_files:
93
  - split: test
 
29
  dtype: string
30
  splits:
31
  - name: test
32
+ num_bytes: 1305088
33
  num_examples: 4358
34
  - name: train
35
+ num_bytes: 546500949
36
  num_examples: 1801350
37
  - name: validation
38
+ num_bytes: 1159288
39
  num_examples: 3760
40
+ download_size: 315466397
41
+ dataset_size: 548965325
42
  - config_name: wikitext-103-v1
43
  features:
44
  - name: text
 
88
  download_size: 7371282
89
  dataset_size: 13323188
90
  configs:
91
+ - config_name: wikitext-103-raw-v1
92
+ data_files:
93
+ - split: test
94
+ path: wikitext-103-raw-v1/test-*
95
+ - split: train
96
+ path: wikitext-103-raw-v1/train-*
97
+ - split: validation
98
+ path: wikitext-103-raw-v1/validation-*
99
  - config_name: wikitext-103-v1
100
  data_files:
101
  - split: test
dataset_infos.json CHANGED
@@ -95,18 +95,14 @@
95
  "features": {
96
  "text": {
97
  "dtype": "string",
98
- "id": null,
99
  "_type": "Value"
100
  }
101
  },
102
- "post_processed": null,
103
- "supervised_keys": null,
104
- "task_templates": null,
105
  "builder_name": "wikitext",
 
106
  "config_name": "wikitext-103-raw-v1",
107
  "version": {
108
  "version_str": "1.0.0",
109
- "description": null,
110
  "major": 1,
111
  "minor": 0,
112
  "patch": 0
@@ -114,33 +110,26 @@
114
  "splits": {
115
  "test": {
116
  "name": "test",
117
- "num_bytes": 1305092,
118
  "num_examples": 4358,
119
- "dataset_name": "wikitext"
120
  },
121
  "train": {
122
  "name": "train",
123
- "num_bytes": 546501673,
124
  "num_examples": 1801350,
125
- "dataset_name": "wikitext"
126
  },
127
  "validation": {
128
  "name": "validation",
129
- "num_bytes": 1159292,
130
  "num_examples": 3760,
131
- "dataset_name": "wikitext"
132
- }
133
- },
134
- "download_checksums": {
135
- "https://s3.amazonaws.com/research.metamind.io/wikitext/wikitext-103-raw-v1.zip": {
136
- "num_bytes": 191984949,
137
- "checksum": "91c00ae287f0d699e18605c84afc9e45c192bc6b7797ff8837e5474655a33794"
138
  }
139
  },
140
- "download_size": 191984949,
141
- "post_processing_size": null,
142
- "dataset_size": 548966057,
143
- "size_in_bytes": 740951006
144
  },
145
  "wikitext-2-raw-v1": {
146
  "description": " The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\n Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike\n License.\n",
 
95
  "features": {
96
  "text": {
97
  "dtype": "string",
 
98
  "_type": "Value"
99
  }
100
  },
 
 
 
101
  "builder_name": "wikitext",
102
+ "dataset_name": "wikitext",
103
  "config_name": "wikitext-103-raw-v1",
104
  "version": {
105
  "version_str": "1.0.0",
 
106
  "major": 1,
107
  "minor": 0,
108
  "patch": 0
 
110
  "splits": {
111
  "test": {
112
  "name": "test",
113
+ "num_bytes": 1305088,
114
  "num_examples": 4358,
115
+ "dataset_name": null
116
  },
117
  "train": {
118
  "name": "train",
119
+ "num_bytes": 546500949,
120
  "num_examples": 1801350,
121
+ "dataset_name": null
122
  },
123
  "validation": {
124
  "name": "validation",
125
+ "num_bytes": 1159288,
126
  "num_examples": 3760,
127
+ "dataset_name": null
 
 
 
 
 
 
128
  }
129
  },
130
+ "download_size": 315466397,
131
+ "dataset_size": 548965325,
132
+ "size_in_bytes": 864431722
 
133
  },
134
  "wikitext-2-raw-v1": {
135
  "description": " The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified\n Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike\n License.\n",
wikitext-103-raw-v1/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f1bea067869d04849c0f975a2b29c4ff47d867f484f5010ea5e861eab246d91
3
+ size 732610
wikitext-103-raw-v1/train-00000-of-00002.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74da360f23826045b3e6ac6375411fdb15f003030aa74f2596ed08b857cb9212
3
+ size 156987808
wikitext-103-raw-v1/train-00001-of-00002.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba090ac30dbf5461e8dcbdd1a1b8e6f3cf9c2c756d64f0c1220450acd514f720
3
+ size 157088770
wikitext-103-raw-v1/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:204929b7ff9d6184953f867dedb860e40aa69c078fc1e54b3baaa8fb28511c4c
3
+ size 657209