Datasets:
Languages:
Korean
Multilinguality:
monolingual
Size Categories:
10K<n<100K
Language Creators:
expert-generated
Annotations Creators:
expert-generated
Source Datasets:
original
ArXiv:
License:
Commit
•
e0e0e06
1
Parent(s):
26a2ff7
Add dp data files
Browse files- README.md +10 -4
- dataset_infos.json +8 -24
- dp/train-00000-of-00001.parquet +3 -0
- dp/validation-00000-of-00001.parquet +3 -0
README.md
CHANGED
@@ -59,13 +59,13 @@ dataset_info:
|
|
59 |
list: string
|
60 |
splits:
|
61 |
- name: train
|
62 |
-
num_bytes:
|
63 |
num_examples: 10000
|
64 |
- name: validation
|
65 |
-
num_bytes:
|
66 |
num_examples: 2000
|
67 |
-
download_size:
|
68 |
-
dataset_size:
|
69 |
- config_name: mrc
|
70 |
features:
|
71 |
- name: title
|
@@ -313,6 +313,12 @@ dataset_info:
|
|
313 |
download_size: 5012303
|
314 |
dataset_size: 12148765
|
315 |
configs:
|
|
|
|
|
|
|
|
|
|
|
|
|
316 |
- config_name: ner
|
317 |
data_files:
|
318 |
- split: train
|
|
|
59 |
list: string
|
60 |
splits:
|
61 |
- name: train
|
62 |
+
num_bytes: 7899965
|
63 |
num_examples: 10000
|
64 |
- name: validation
|
65 |
+
num_bytes: 1557462
|
66 |
num_examples: 2000
|
67 |
+
download_size: 3742577
|
68 |
+
dataset_size: 9457427
|
69 |
- config_name: mrc
|
70 |
features:
|
71 |
- name: title
|
|
|
313 |
download_size: 5012303
|
314 |
dataset_size: 12148765
|
315 |
configs:
|
316 |
+
- config_name: dp
|
317 |
+
data_files:
|
318 |
+
- split: train
|
319 |
+
path: dp/train-*
|
320 |
+
- split: validation
|
321 |
+
path: dp/validation-*
|
322 |
- config_name: ner
|
323 |
data_files:
|
324 |
- split: train
|
dataset_infos.json
CHANGED
@@ -382,56 +382,47 @@
|
|
382 |
"features": {
|
383 |
"sentence": {
|
384 |
"dtype": "string",
|
385 |
-
"id": null,
|
386 |
"_type": "Value"
|
387 |
},
|
388 |
"index": [
|
389 |
{
|
390 |
"dtype": "int32",
|
391 |
-
"id": null,
|
392 |
"_type": "Value"
|
393 |
}
|
394 |
],
|
395 |
"word_form": [
|
396 |
{
|
397 |
"dtype": "string",
|
398 |
-
"id": null,
|
399 |
"_type": "Value"
|
400 |
}
|
401 |
],
|
402 |
"lemma": [
|
403 |
{
|
404 |
"dtype": "string",
|
405 |
-
"id": null,
|
406 |
"_type": "Value"
|
407 |
}
|
408 |
],
|
409 |
"pos": [
|
410 |
{
|
411 |
"dtype": "string",
|
412 |
-
"id": null,
|
413 |
"_type": "Value"
|
414 |
}
|
415 |
],
|
416 |
"head": [
|
417 |
{
|
418 |
"dtype": "int32",
|
419 |
-
"id": null,
|
420 |
"_type": "Value"
|
421 |
}
|
422 |
],
|
423 |
"deprel": [
|
424 |
{
|
425 |
"dtype": "string",
|
426 |
-
"id": null,
|
427 |
"_type": "Value"
|
428 |
}
|
429 |
]
|
430 |
},
|
431 |
-
"post_processed": null,
|
432 |
-
"supervised_keys": null,
|
433 |
-
"task_templates": null,
|
434 |
"builder_name": "klue",
|
|
|
435 |
"config_name": "dp",
|
436 |
"version": {
|
437 |
"version_str": "1.0.0",
|
@@ -443,27 +434,20 @@
|
|
443 |
"splits": {
|
444 |
"train": {
|
445 |
"name": "train",
|
446 |
-
"num_bytes":
|
447 |
"num_examples": 10000,
|
448 |
-
"dataset_name":
|
449 |
},
|
450 |
"validation": {
|
451 |
"name": "validation",
|
452 |
-
"num_bytes":
|
453 |
"num_examples": 2000,
|
454 |
-
"dataset_name":
|
455 |
-
}
|
456 |
-
},
|
457 |
-
"download_checksums": {
|
458 |
-
"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000071/data/klue-dp-v1.tar.gz": {
|
459 |
-
"num_bytes": 2033461,
|
460 |
-
"checksum": "2c76a3543a50599ac6640ad360ba00eac36e0b5b2363f708a614d6e50844d17b"
|
461 |
}
|
462 |
},
|
463 |
-
"download_size":
|
464 |
-
"
|
465 |
-
"
|
466 |
-
"size_in_bytes": 11490976
|
467 |
},
|
468 |
"mrc": {
|
469 |
"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n",
|
|
|
382 |
"features": {
|
383 |
"sentence": {
|
384 |
"dtype": "string",
|
|
|
385 |
"_type": "Value"
|
386 |
},
|
387 |
"index": [
|
388 |
{
|
389 |
"dtype": "int32",
|
|
|
390 |
"_type": "Value"
|
391 |
}
|
392 |
],
|
393 |
"word_form": [
|
394 |
{
|
395 |
"dtype": "string",
|
|
|
396 |
"_type": "Value"
|
397 |
}
|
398 |
],
|
399 |
"lemma": [
|
400 |
{
|
401 |
"dtype": "string",
|
|
|
402 |
"_type": "Value"
|
403 |
}
|
404 |
],
|
405 |
"pos": [
|
406 |
{
|
407 |
"dtype": "string",
|
|
|
408 |
"_type": "Value"
|
409 |
}
|
410 |
],
|
411 |
"head": [
|
412 |
{
|
413 |
"dtype": "int32",
|
|
|
414 |
"_type": "Value"
|
415 |
}
|
416 |
],
|
417 |
"deprel": [
|
418 |
{
|
419 |
"dtype": "string",
|
|
|
420 |
"_type": "Value"
|
421 |
}
|
422 |
]
|
423 |
},
|
|
|
|
|
|
|
424 |
"builder_name": "klue",
|
425 |
+
"dataset_name": "klue",
|
426 |
"config_name": "dp",
|
427 |
"version": {
|
428 |
"version_str": "1.0.0",
|
|
|
434 |
"splits": {
|
435 |
"train": {
|
436 |
"name": "train",
|
437 |
+
"num_bytes": 7899965,
|
438 |
"num_examples": 10000,
|
439 |
+
"dataset_name": null
|
440 |
},
|
441 |
"validation": {
|
442 |
"name": "validation",
|
443 |
+
"num_bytes": 1557462,
|
444 |
"num_examples": 2000,
|
445 |
+
"dataset_name": null
|
|
|
|
|
|
|
|
|
|
|
|
|
446 |
}
|
447 |
},
|
448 |
+
"download_size": 3742577,
|
449 |
+
"dataset_size": 9457427,
|
450 |
+
"size_in_bytes": 13200004
|
|
|
451 |
},
|
452 |
"mrc": {
|
453 |
"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n",
|
dp/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3f570c15f1da1b70d353da6a5896dc0a4f32d2b178e23a4def7b6e953dff30d
|
3 |
+
size 3124932
|
dp/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f43905816424723d15c63fe9ea724fb2c7dacfb88289b2b4e345f259d6119f1c
|
3 |
+
size 617645
|