Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
Qingyun commited on
Commit
2fad0a7
1 Parent(s): 6f244cf

Upload dataset

Browse files
CC-MAIN-2014-23/train-00000-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:593815fce8c92dcab6aa47e664efdafb6acec3327ed47b60f80fb334aa1017cc
3
+ size 433554020
CC-MAIN-2014-23/train-00001-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c3b99ae4dabf963d343a019b95ad8669e17009cf9cb7d9db14473fcdde1fd4a
3
+ size 433759107
CC-MAIN-2014-23/train-00002-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a081184982d1e6ef35b37d13fbb51ca77d3cc6a127406b198024a093fb5628f4
3
+ size 434586617
CC-MAIN-2014-23/train-00003-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8663a0480effc800c5a9f6932df293a78ebf938a3b1138cde01921915b2695d0
3
+ size 432417522
CC-MAIN-2014-23/train-00004-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f451744378d8428f98987e7b5915512aa211bdcfd14fdc91816ae6bb17e1e371
3
+ size 433414565
CC-MAIN-2014-23/train-00005-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:041a97610572b7efea224d4ad89d83f840fb945e35cc24810d56223e652f7d24
3
+ size 433304384
CC-MAIN-2014-23/train-00006-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:005bafd2df445f6413b07d97b03162e128cbbff60ae7dd1385aae66202effbbf
3
+ size 432060462
CC-MAIN-2014-23/train-00007-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a12813e4fe630c56809a5f9093028458bd3abd9bed324dafb2b84f36017ec350
3
+ size 435756228
README.md CHANGED
@@ -216,6 +216,58 @@ dataset_info:
216
  num_examples: 1968361
217
  download_size: 4567738362
218
  dataset_size: 10106913108
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
219
  configs:
220
  - config_name: CC-MAIN-2013-20
221
  data_files:
@@ -233,6 +285,10 @@ configs:
233
  data_files:
234
  - split: train
235
  path: CC-MAIN-2014-15/train-*
 
 
 
 
236
  ---
237
 
238
  We are uploading the dataset files ~
 
216
  num_examples: 1968361
217
  download_size: 4567738362
218
  dataset_size: 10106913108
219
+ - config_name: CC-MAIN-2014-23
220
+ features:
221
+ - name: general_metadata
222
+ struct:
223
+ - name: domain
224
+ sequence: string
225
+ - name: fluency_prob
226
+ dtype: float64
227
+ - name: id
228
+ dtype: string
229
+ - name: non_advertisement_prob
230
+ dtype: float64
231
+ - name: politics_prob
232
+ dtype: float64
233
+ - name: porn_prob
234
+ dtype: float64
235
+ - name: toxic_prob
236
+ dtype: float64
237
+ - name: url
238
+ dtype: string
239
+ - name: images
240
+ sequence: string
241
+ - name: texts
242
+ sequence: string
243
+ - name: metadata
244
+ list:
245
+ - name: aesthetic_prob
246
+ dtype: float64
247
+ - name: bytes
248
+ dtype: int64
249
+ - name: d_hash
250
+ dtype: string
251
+ - name: d_hash_dup_count
252
+ dtype: int64
253
+ - name: height
254
+ dtype: int64
255
+ - name: img_url_sha
256
+ dtype: string
257
+ - name: p_hash
258
+ dtype: string
259
+ - name: p_hash_dup_count
260
+ dtype: int64
261
+ - name: unsafe_prob
262
+ dtype: float64
263
+ - name: width
264
+ dtype: int64
265
+ splits:
266
+ - name: train
267
+ num_bytes: 7997621043
268
+ num_examples: 1455331
269
+ download_size: 3468852905
270
+ dataset_size: 7997621043
271
  configs:
272
  - config_name: CC-MAIN-2013-20
273
  data_files:
 
285
  data_files:
286
  - split: train
287
  path: CC-MAIN-2014-15/train-*
288
+ - config_name: CC-MAIN-2014-23
289
+ data_files:
290
+ - split: train
291
+ path: CC-MAIN-2014-23/train-*
292
  ---
293
 
294
  We are uploading the dataset files ~