albertvillanova HF staff commited on
Commit
e0e0e06
1 Parent(s): 26a2ff7

Add dp data files

Browse files
README.md CHANGED
@@ -59,13 +59,13 @@ dataset_info:
59
  list: string
60
  splits:
61
  - name: train
62
- num_bytes: 7900009
63
  num_examples: 10000
64
  - name: validation
65
- num_bytes: 1557506
66
  num_examples: 2000
67
- download_size: 2033461
68
- dataset_size: 9457515
69
  - config_name: mrc
70
  features:
71
  - name: title
@@ -313,6 +313,12 @@ dataset_info:
313
  download_size: 5012303
314
  dataset_size: 12148765
315
  configs:
 
 
 
 
 
 
316
  - config_name: ner
317
  data_files:
318
  - split: train
 
59
  list: string
60
  splits:
61
  - name: train
62
+ num_bytes: 7899965
63
  num_examples: 10000
64
  - name: validation
65
+ num_bytes: 1557462
66
  num_examples: 2000
67
+ download_size: 3742577
68
+ dataset_size: 9457427
69
  - config_name: mrc
70
  features:
71
  - name: title
 
313
  download_size: 5012303
314
  dataset_size: 12148765
315
  configs:
316
+ - config_name: dp
317
+ data_files:
318
+ - split: train
319
+ path: dp/train-*
320
+ - split: validation
321
+ path: dp/validation-*
322
  - config_name: ner
323
  data_files:
324
  - split: train
dataset_infos.json CHANGED
@@ -382,56 +382,47 @@
382
  "features": {
383
  "sentence": {
384
  "dtype": "string",
385
- "id": null,
386
  "_type": "Value"
387
  },
388
  "index": [
389
  {
390
  "dtype": "int32",
391
- "id": null,
392
  "_type": "Value"
393
  }
394
  ],
395
  "word_form": [
396
  {
397
  "dtype": "string",
398
- "id": null,
399
  "_type": "Value"
400
  }
401
  ],
402
  "lemma": [
403
  {
404
  "dtype": "string",
405
- "id": null,
406
  "_type": "Value"
407
  }
408
  ],
409
  "pos": [
410
  {
411
  "dtype": "string",
412
- "id": null,
413
  "_type": "Value"
414
  }
415
  ],
416
  "head": [
417
  {
418
  "dtype": "int32",
419
- "id": null,
420
  "_type": "Value"
421
  }
422
  ],
423
  "deprel": [
424
  {
425
  "dtype": "string",
426
- "id": null,
427
  "_type": "Value"
428
  }
429
  ]
430
  },
431
- "post_processed": null,
432
- "supervised_keys": null,
433
- "task_templates": null,
434
  "builder_name": "klue",
 
435
  "config_name": "dp",
436
  "version": {
437
  "version_str": "1.0.0",
@@ -443,27 +434,20 @@
443
  "splits": {
444
  "train": {
445
  "name": "train",
446
- "num_bytes": 7900009,
447
  "num_examples": 10000,
448
- "dataset_name": "klue"
449
  },
450
  "validation": {
451
  "name": "validation",
452
- "num_bytes": 1557506,
453
  "num_examples": 2000,
454
- "dataset_name": "klue"
455
- }
456
- },
457
- "download_checksums": {
458
- "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000071/data/klue-dp-v1.tar.gz": {
459
- "num_bytes": 2033461,
460
- "checksum": "2c76a3543a50599ac6640ad360ba00eac36e0b5b2363f708a614d6e50844d17b"
461
  }
462
  },
463
- "download_size": 2033461,
464
- "post_processing_size": null,
465
- "dataset_size": 9457515,
466
- "size_in_bytes": 11490976
467
  },
468
  "mrc": {
469
  "description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n",
 
382
  "features": {
383
  "sentence": {
384
  "dtype": "string",
 
385
  "_type": "Value"
386
  },
387
  "index": [
388
  {
389
  "dtype": "int32",
 
390
  "_type": "Value"
391
  }
392
  ],
393
  "word_form": [
394
  {
395
  "dtype": "string",
 
396
  "_type": "Value"
397
  }
398
  ],
399
  "lemma": [
400
  {
401
  "dtype": "string",
 
402
  "_type": "Value"
403
  }
404
  ],
405
  "pos": [
406
  {
407
  "dtype": "string",
 
408
  "_type": "Value"
409
  }
410
  ],
411
  "head": [
412
  {
413
  "dtype": "int32",
 
414
  "_type": "Value"
415
  }
416
  ],
417
  "deprel": [
418
  {
419
  "dtype": "string",
 
420
  "_type": "Value"
421
  }
422
  ]
423
  },
 
 
 
424
  "builder_name": "klue",
425
+ "dataset_name": "klue",
426
  "config_name": "dp",
427
  "version": {
428
  "version_str": "1.0.0",
 
434
  "splits": {
435
  "train": {
436
  "name": "train",
437
+ "num_bytes": 7899965,
438
  "num_examples": 10000,
439
+ "dataset_name": null
440
  },
441
  "validation": {
442
  "name": "validation",
443
+ "num_bytes": 1557462,
444
  "num_examples": 2000,
445
+ "dataset_name": null
 
 
 
 
 
 
446
  }
447
  },
448
+ "download_size": 3742577,
449
+ "dataset_size": 9457427,
450
+ "size_in_bytes": 13200004
 
451
  },
452
  "mrc": {
453
  "description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n",
dp/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3f570c15f1da1b70d353da6a5896dc0a4f32d2b178e23a4def7b6e953dff30d
3
+ size 3124932
dp/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f43905816424723d15c63fe9ea724fb2c7dacfb88289b2b4e345f259d6119f1c
3
+ size 617645