albertvillanova HF staff commited on
Commit
349481e
1 Parent(s): 750599f

Convert dataset to Parquet (#2)

Browse files

- Convert dataset to Parquet (aa733cddfbffca00259b4f4a058b92d74b65380c)
- Add sts data files (b7f730b37c6eaa998178d7550ac35827fc0bf5ca)
- Add nli data files (0e05c2a3649a814bc206ae51d13d0a38ac991f64)
- Add ner data files (c302a004d918a49b00002e45475a0aeaa239aa1e)
- Add re data files (26a2ff79dc50e1b253243f1a511b031f56a38499)
- Add dp data files (e0e0e06ec993811871d9d81a2777a7fc8a61896c)
- Add mrc data files (9dc841ba15d7f362b0e55b4579307319e3cc167b)
- Add wos data files (5454eaaa94dafb629516e236faa87dbe0501cb09)
- Delete loading script (89d2a4e6e716de45c3e19ffe056665df11b3aac6)
- Delete legacy dataset_infos.json (2e32b5c21562b629dbd5140500af66e5db303bf9)

README.md CHANGED
@@ -29,96 +29,76 @@ task_ids:
29
  - topic-classification
30
  paperswithcode_id: klue
31
  pretty_name: KLUE
 
 
 
 
 
 
 
 
 
32
  tags:
33
  - relation-extraction
34
  dataset_info:
35
- - config_name: ynat
36
  features:
37
- - name: guid
38
- dtype: string
39
- - name: title
40
- dtype: string
41
- - name: label
42
- dtype:
43
- class_label:
44
- names:
45
- '0': IT과학
46
- '1': 경제
47
- '2': 사회
48
- '3': 생활문화
49
- '4': 세계
50
- '5': 스포츠
51
- '6': 정치
52
- - name: url
53
- dtype: string
54
- - name: date
55
  dtype: string
 
 
 
 
 
 
 
 
 
 
 
 
56
  splits:
57
  - name: train
58
- num_bytes: 10109664
59
- num_examples: 45678
60
  - name: validation
61
- num_bytes: 2039197
62
- num_examples: 9107
63
- download_size: 4932555
64
- dataset_size: 12148861
65
- - config_name: sts
66
  features:
67
- - name: guid
68
- dtype: string
69
- - name: source
70
- dtype: string
71
- - name: sentence1
72
  dtype: string
73
- - name: sentence2
74
  dtype: string
75
- - name: labels
76
- struct:
77
- - name: label
78
- dtype: float64
79
- - name: real-label
80
- dtype: float64
81
- - name: binary-label
82
- dtype:
83
- class_label:
84
- names:
85
- '0': negative
86
- '1': positive
87
- splits:
88
- - name: train
89
- num_bytes: 2832921
90
- num_examples: 11668
91
- - name: validation
92
- num_bytes: 122657
93
- num_examples: 519
94
- download_size: 1349875
95
- dataset_size: 2955578
96
- - config_name: nli
97
- features:
98
- - name: guid
99
  dtype: string
100
  - name: source
101
  dtype: string
102
- - name: premise
103
  dtype: string
104
- - name: hypothesis
 
 
 
 
105
  dtype: string
106
- - name: label
107
- dtype:
108
- class_label:
109
- names:
110
- '0': entailment
111
- '1': neutral
112
- '2': contradiction
113
  splits:
114
  - name: train
115
- num_bytes: 5719930
116
- num_examples: 24998
117
  - name: validation
118
- num_bytes: 673276
119
- num_examples: 3000
120
- download_size: 1257374
121
- dataset_size: 6393206
122
  - config_name: ner
123
  features:
124
  - name: sentence
@@ -144,13 +124,39 @@ dataset_info:
144
  '12': O
145
  splits:
146
  - name: train
147
- num_bytes: 19891953
148
  num_examples: 21008
149
  - name: validation
150
- num_bytes: 4937579
151
  num_examples: 5000
152
- download_size: 4308644
153
- dataset_size: 24829532
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
154
  - config_name: re
155
  features:
156
  - name: guid
@@ -215,71 +221,44 @@ dataset_info:
215
  dtype: string
216
  splits:
217
  - name: train
218
- num_bytes: 11145538
219
  num_examples: 32470
220
  - name: validation
221
- num_bytes: 2559300
222
  num_examples: 7765
223
- download_size: 5669259
224
- dataset_size: 13704838
225
- - config_name: dp
226
- features:
227
- - name: sentence
228
- dtype: string
229
- - name: index
230
- list: int32
231
- - name: word_form
232
- list: string
233
- - name: lemma
234
- list: string
235
- - name: pos
236
- list: string
237
- - name: head
238
- list: int32
239
- - name: deprel
240
- list: string
241
- splits:
242
- - name: train
243
- num_bytes: 7900009
244
- num_examples: 10000
245
- - name: validation
246
- num_bytes: 1557506
247
- num_examples: 2000
248
- download_size: 2033461
249
- dataset_size: 9457515
250
- - config_name: mrc
251
  features:
252
- - name: title
253
- dtype: string
254
- - name: context
255
- dtype: string
256
- - name: news_category
257
  dtype: string
258
  - name: source
259
  dtype: string
260
- - name: guid
261
  dtype: string
262
- - name: is_impossible
263
- dtype: bool
264
- - name: question_type
265
- dtype: int32
266
- - name: question
267
  dtype: string
268
- - name: answers
269
- sequence:
270
- - name: answer_start
271
- dtype: int32
272
- - name: text
273
- dtype: string
 
 
 
 
 
 
274
  splits:
275
  - name: train
276
- num_bytes: 46505665
277
- num_examples: 17554
278
  - name: validation
279
- num_bytes: 15583053
280
- num_examples: 5841
281
- download_size: 19218422
282
- dataset_size: 62088718
283
  - config_name: wos
284
  features:
285
  - name: guid
@@ -296,22 +275,92 @@ dataset_info:
296
  list: string
297
  splits:
298
  - name: train
299
- num_bytes: 26677002
300
  num_examples: 8000
301
  - name: validation
302
- num_bytes: 3488943
303
  num_examples: 1000
304
- download_size: 4785657
305
- dataset_size: 30165945
306
- config_names:
307
- - dp
308
- - mrc
309
- - ner
310
- - nli
311
- - re
312
- - sts
313
- - wos
314
- - ynat
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
315
  ---
316
 
317
  # Dataset Card for KLUE
29
  - topic-classification
30
  paperswithcode_id: klue
31
  pretty_name: KLUE
32
+ config_names:
33
+ - dp
34
+ - mrc
35
+ - ner
36
+ - nli
37
+ - re
38
+ - sts
39
+ - wos
40
+ - ynat
41
  tags:
42
  - relation-extraction
43
  dataset_info:
44
+ - config_name: dp
45
  features:
46
+ - name: sentence
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  dtype: string
48
+ - name: index
49
+ list: int32
50
+ - name: word_form
51
+ list: string
52
+ - name: lemma
53
+ list: string
54
+ - name: pos
55
+ list: string
56
+ - name: head
57
+ list: int32
58
+ - name: deprel
59
+ list: string
60
  splits:
61
  - name: train
62
+ num_bytes: 7899965
63
+ num_examples: 10000
64
  - name: validation
65
+ num_bytes: 1557462
66
+ num_examples: 2000
67
+ download_size: 3742577
68
+ dataset_size: 9457427
69
+ - config_name: mrc
70
  features:
71
+ - name: title
 
 
 
 
72
  dtype: string
73
+ - name: context
74
  dtype: string
75
+ - name: news_category
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76
  dtype: string
77
  - name: source
78
  dtype: string
79
+ - name: guid
80
  dtype: string
81
+ - name: is_impossible
82
+ dtype: bool
83
+ - name: question_type
84
+ dtype: int32
85
+ - name: question
86
  dtype: string
87
+ - name: answers
88
+ sequence:
89
+ - name: answer_start
90
+ dtype: int32
91
+ - name: text
92
+ dtype: string
 
93
  splits:
94
  - name: train
95
+ num_bytes: 46505593
96
+ num_examples: 17554
97
  - name: validation
98
+ num_bytes: 15583017
99
+ num_examples: 5841
100
+ download_size: 30098472
101
+ dataset_size: 62088610
102
  - config_name: ner
103
  features:
104
  - name: sentence
124
  '12': O
125
  splits:
126
  - name: train
127
+ num_bytes: 19891905
128
  num_examples: 21008
129
  - name: validation
130
+ num_bytes: 4937563
131
  num_examples: 5000
132
+ download_size: 5265887
133
+ dataset_size: 24829468
134
+ - config_name: nli
135
+ features:
136
+ - name: guid
137
+ dtype: string
138
+ - name: source
139
+ dtype: string
140
+ - name: premise
141
+ dtype: string
142
+ - name: hypothesis
143
+ dtype: string
144
+ - name: label
145
+ dtype:
146
+ class_label:
147
+ names:
148
+ '0': entailment
149
+ '1': neutral
150
+ '2': contradiction
151
+ splits:
152
+ - name: train
153
+ num_bytes: 5719882
154
+ num_examples: 24998
155
+ - name: validation
156
+ num_bytes: 673260
157
+ num_examples: 3000
158
+ download_size: 2056116
159
+ dataset_size: 6393142
160
  - config_name: re
161
  features:
162
  - name: guid
221
  dtype: string
222
  splits:
223
  - name: train
224
+ num_bytes: 11145426
225
  num_examples: 32470
226
  - name: validation
227
+ num_bytes: 2559272
228
  num_examples: 7765
229
+ download_size: 8190257
230
+ dataset_size: 13704698
231
+ - config_name: sts
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
232
  features:
233
+ - name: guid
 
 
 
 
234
  dtype: string
235
  - name: source
236
  dtype: string
237
+ - name: sentence1
238
  dtype: string
239
+ - name: sentence2
 
 
 
 
240
  dtype: string
241
+ - name: labels
242
+ struct:
243
+ - name: label
244
+ dtype: float64
245
+ - name: real-label
246
+ dtype: float64
247
+ - name: binary-label
248
+ dtype:
249
+ class_label:
250
+ names:
251
+ '0': negative
252
+ '1': positive
253
  splits:
254
  - name: train
255
+ num_bytes: 2832889
256
+ num_examples: 11668
257
  - name: validation
258
+ num_bytes: 122641
259
+ num_examples: 519
260
+ download_size: 1587855
261
+ dataset_size: 2955530
262
  - config_name: wos
263
  features:
264
  - name: guid
275
  list: string
276
  splits:
277
  - name: train
278
+ num_bytes: 26676970
279
  num_examples: 8000
280
  - name: validation
281
+ num_bytes: 3488911
282
  num_examples: 1000
283
+ download_size: 6358855
284
+ dataset_size: 30165881
285
+ - config_name: ynat
286
+ features:
287
+ - name: guid
288
+ dtype: string
289
+ - name: title
290
+ dtype: string
291
+ - name: label
292
+ dtype:
293
+ class_label:
294
+ names:
295
+ '0': IT과학
296
+ '1': 경제
297
+ '2': 사회
298
+ '3': 생활문화
299
+ '4': 세계
300
+ '5': 스포츠
301
+ '6': 정치
302
+ - name: url
303
+ dtype: string
304
+ - name: date
305
+ dtype: string
306
+ splits:
307
+ - name: train
308
+ num_bytes: 10109584
309
+ num_examples: 45678
310
+ - name: validation
311
+ num_bytes: 2039181
312
+ num_examples: 9107
313
+ download_size: 5012303
314
+ dataset_size: 12148765
315
+ configs:
316
+ - config_name: dp
317
+ data_files:
318
+ - split: train
319
+ path: dp/train-*
320
+ - split: validation
321
+ path: dp/validation-*
322
+ - config_name: mrc
323
+ data_files:
324
+ - split: train
325
+ path: mrc/train-*
326
+ - split: validation
327
+ path: mrc/validation-*
328
+ - config_name: ner
329
+ data_files:
330
+ - split: train
331
+ path: ner/train-*
332
+ - split: validation
333
+ path: ner/validation-*
334
+ - config_name: nli
335
+ data_files:
336
+ - split: train
337
+ path: nli/train-*
338
+ - split: validation
339
+ path: nli/validation-*
340
+ - config_name: re
341
+ data_files:
342
+ - split: train
343
+ path: re/train-*
344
+ - split: validation
345
+ path: re/validation-*
346
+ - config_name: sts
347
+ data_files:
348
+ - split: train
349
+ path: sts/train-*
350
+ - split: validation
351
+ path: sts/validation-*
352
+ - config_name: wos
353
+ data_files:
354
+ - split: train
355
+ path: wos/train-*
356
+ - split: validation
357
+ path: wos/validation-*
358
+ - config_name: ynat
359
+ data_files:
360
+ - split: train
361
+ path: ynat/train-*
362
+ - split: validation
363
+ path: ynat/validation-*
364
  ---
365
 
366
  # Dataset Card for KLUE
dataset_infos.json DELETED
@@ -1 +0,0 @@
1
- {"ynat": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/66/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "title": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 7, "names": ["IT\uacfc\ud559", "\uacbd\uc81c", "\uc0ac\ud68c", "\uc0dd\ud65c\ubb38\ud654", "\uc138\uacc4", "\uc2a4\ud3ec\uce20", "\uc815\uce58"], "names_file": null, "id": null, "_type": "ClassLabel"}, "url": {"dtype": "string", "id": null, "_type": "Value"}, "date": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "ynat", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 10109664, "num_examples": 45678, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 2039197, "num_examples": 9107, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000066/data/ynat-v1.tar.gz": {"num_bytes": 4932555, "checksum": "820a4d1d6d1fd83e2a421f856965d3cfc5c93627935ce8c5b27468c6113fc482"}}, "download_size": 4932555, "post_processing_size": null, "dataset_size": 12148861, "size_in_bytes": 17081416}, "sts": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/67/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "source": {"dtype": "string", "id": null, "_type": "Value"}, "sentence1": {"dtype": "string", "id": null, "_type": "Value"}, "sentence2": {"dtype": "string", "id": null, "_type": "Value"}, "labels": {"label": {"dtype": "float64", "id": null, "_type": "Value"}, "real-label": {"dtype": "float64", "id": null, "_type": "Value"}, "binary-label": {"num_classes": 2, "names": ["negative", "positive"], "names_file": null, "id": null, "_type": "ClassLabel"}}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "sts", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2832921, "num_examples": 11668, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 122657, "num_examples": 519, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000067/data/klue-sts-v1.tar.gz": {"num_bytes": 1349875, "checksum": "539341ba78a3b351c686cf70a448ac7a5886ed95f0719d5e3d2378ba703213bd"}}, "download_size": 1349875, "post_processing_size": null, "dataset_size": 2955578, "size_in_bytes": 4305453}, "nli": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/68/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "source": {"dtype": "string", "id": null, "_type": "Value"}, "premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "nli", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 5719930, "num_examples": 24998, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 673276, "num_examples": 3000, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000068/data/klue-nli-v1.tar.gz": {"num_bytes": 1257374, "checksum": "388be2033ef712072201903795a35b4f86826ee3ed3b62dc0c98e1721baa8850"}}, "download_size": 1257374, "post_processing_size": null, "dataset_size": 6393206, "size_in_bytes": 7650580}, "ner": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/69/overview/description", "license": "CC-BY-SA-4.0", "features": {"sentence": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 13, "names": ["B-DT", "I-DT", "B-LC", "I-LC", "B-OG", "I-OG", "B-PS", "I-PS", "B-QT", "I-QT", "B-TI", "I-TI", "O"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "ner", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 19891953, "num_examples": 21008, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 4937579, "num_examples": 5000, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000069/data/klue-ner-v1.tar.gz": {"num_bytes": 4308644, "checksum": "848a89759ac6b7c149c9a00d820726fe2a140c22782201f1a40d856672e7ea8e"}}, "download_size": 4308644, "post_processing_size": null, "dataset_size": 24829532, "size_in_bytes": 29138176}, "re": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/70/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "sentence": {"dtype": "string", "id": null, "_type": "Value"}, "subject_entity": {"word": {"dtype": "string", "id": null, "_type": "Value"}, "start_idx": {"dtype": "int32", "id": null, "_type": "Value"}, "end_idx": {"dtype": "int32", "id": null, "_type": "Value"}, "type": {"dtype": "string", "id": null, "_type": "Value"}}, "object_entity": {"word": {"dtype": "string", "id": null, "_type": "Value"}, "start_idx": {"dtype": "int32", "id": null, "_type": "Value"}, "end_idx": {"dtype": "int32", "id": null, "_type": "Value"}, "type": {"dtype": "string", "id": null, "_type": "Value"}}, "label": {"num_classes": 30, "names": ["no_relation", "org:dissolved", "org:founded", "org:place_of_headquarters", "org:alternate_names", "org:member_of", "org:members", "org:political/religious_affiliation", "org:product", "org:founded_by", "org:top_members/employees", "org:number_of_employees/members", "per:date_of_birth", "per:date_of_death", "per:place_of_birth", "per:place_of_death", "per:place_of_residence", "per:origin", "per:employee_of", "per:schools_attended", "per:alternate_names", "per:parents", "per:children", "per:siblings", "per:spouse", "per:other_family", "per:colleagues", "per:product", "per:religion", "per:title"], "names_file": null, "id": null, "_type": "ClassLabel"}, "source": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "re", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 11145538, "num_examples": 32470, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 2559300, "num_examples": 7765, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000070/data/klue-re-v1.tar.gz": {"num_bytes": 5669259, "checksum": "b09ceac0d986cc09e42fcda9c7f2873c0eea8ec0629baf91fead36580790f8f5"}}, "download_size": 5669259, "post_processing_size": null, "dataset_size": 13704838, "size_in_bytes": 19374097}, "dp": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/71/overview/description", "license": "CC-BY-SA-4.0", "features": {"sentence": {"dtype": "string", "id": null, "_type": "Value"}, "index": [{"dtype": "int32", "id": null, "_type": "Value"}], "word_form": [{"dtype": "string", "id": null, "_type": "Value"}], "lemma": [{"dtype": "string", "id": null, "_type": "Value"}], "pos": [{"dtype": "string", "id": null, "_type": "Value"}], "head": [{"dtype": "int32", "id": null, "_type": "Value"}], "deprel": [{"dtype": "string", "id": null, "_type": "Value"}]}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "dp", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 7900009, "num_examples": 10000, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 1557506, "num_examples": 2000, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000071/data/klue-dp-v1.tar.gz": {"num_bytes": 2033461, "checksum": "2c76a3543a50599ac6640ad360ba00eac36e0b5b2363f708a614d6e50844d17b"}}, "download_size": 2033461, "post_processing_size": null, "dataset_size": 9457515, "size_in_bytes": 11490976}, "mrc": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/72/overview/description", "license": "CC-BY-SA-4.0", "features": {"title": {"dtype": "string", "id": null, "_type": "Value"}, "context": {"dtype": "string", "id": null, "_type": "Value"}, "news_category": {"dtype": "string", "id": null, "_type": "Value"}, "source": {"dtype": "string", "id": null, "_type": "Value"}, "guid": {"dtype": "string", "id": null, "_type": "Value"}, "is_impossible": {"dtype": "bool", "id": null, "_type": "Value"}, "question_type": {"dtype": "int32", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "answers": {"feature": {"answer_start": {"dtype": "int32", "id": null, "_type": "Value"}, "text": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "mrc", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 46505665, "num_examples": 17554, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 15583053, "num_examples": 5841, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000072/data/klue-mrc-v1.tar.gz": {"num_bytes": 19218422, "checksum": "a444af252901452380d58a6320908ce4a86759bb6f38ad95d0ca98584ad33d14"}}, "download_size": 19218422, "post_processing_size": null, "dataset_size": 62088718, "size_in_bytes": 81307140}, "wos": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/73/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "domains": [{"dtype": "string", "id": null, "_type": "Value"}], "dialogue": [{"role": {"dtype": "string", "id": null, "_type": "Value"}, "text": {"dtype": "string", "id": null, "_type": "Value"}, "state": [{"dtype": "string", "id": null, "_type": "Value"}]}]}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "wos", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 26677002, "num_examples": 8000, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 3488943, "num_examples": 1000, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000073/data/wos-v1.tar.gz": {"num_bytes": 4785657, "checksum": "da17829300271560afc6e7fc330503c2ca6f7ae7721d9bb94308579542a5871f"}}, "download_size": 4785657, "post_processing_size": null, "dataset_size": 30165945, "size_in_bytes": 34951602}}
 
dp/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3f570c15f1da1b70d353da6a5896dc0a4f32d2b178e23a4def7b6e953dff30d
3
+ size 3124932
dp/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f43905816424723d15c63fe9ea724fb2c7dacfb88289b2b4e345f259d6119f1c
3
+ size 617645
klue.py DELETED
@@ -1,533 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- """KLUE (Korean Language Understanding Evaluation) benchmark."""
17
-
18
-
19
- import csv
20
- import json
21
- import textwrap
22
-
23
- import datasets
24
-
25
-
26
- _KLUE_CITATION = """\
27
- @misc{park2021klue,
28
- title={KLUE: Korean Language Understanding Evaluation},
29
- author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},
30
- year={2021},
31
- eprint={2105.09680},
32
- archivePrefix={arXiv},
33
- primaryClass={cs.CL}
34
- }
35
- """
36
-
37
- _KLUE_DESCRIPTION = """\
38
- KLUE (Korean Language Understanding Evaluation)
39
- Korean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language
40
- understanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible
41
- to anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain
42
- unambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics
43
- for every task, thus establishing fair comparison across Korean language models.
44
- """
45
-
46
- _DATA_URLs = {
47
- "ynat": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000066/data/ynat-v1.tar.gz",
48
- "sts": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000067/data/klue-sts-v1.tar.gz",
49
- "nli": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000068/data/klue-nli-v1.tar.gz",
50
- "ner": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000069/data/klue-ner-v1.tar.gz",
51
- "re": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000070/data/klue-re-v1.tar.gz",
52
- "dp": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000071/data/klue-dp-v1.tar.gz",
53
- "mrc": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000072/data/klue-mrc-v1.tar.gz",
54
- "wos": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000073/data/wos-v1.tar.gz",
55
- }
56
-
57
- _DESCRIPTION_URLs = {
58
- "ynat": "https://klue-benchmark.com/tasks/66/overview/description",
59
- "sts": "https://klue-benchmark.com/tasks/67/overview/description",
60
- "nli": "https://klue-benchmark.com/tasks/68/overview/description",
61
- "ner": "https://klue-benchmark.com/tasks/69/overview/description",
62
- "re": "https://klue-benchmark.com/tasks/70/overview/description",
63
- "dp": "https://klue-benchmark.com/tasks/71/overview/description",
64
- "mrc": "https://klue-benchmark.com/tasks/72/overview/description",
65
- "wos": "https://klue-benchmark.com/tasks/73/overview/description",
66
- }
67
-
68
- _LICENSE = "CC-BY-SA-4.0"
69
-
70
-
71
- class KlueConfig(datasets.BuilderConfig):
72
- """BuilderConfig for KLUE."""
73
-
74
- def __init__(
75
- self,
76
- features,
77
- data_url,
78
- url,
79
- file_map,
80
- **kwargs,
81
- ):
82
- """BuilderConfig for KLUE."""
83
-
84
- super(KlueConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs)
85
- self.features = features
86
- self.data_url = data_url
87
- self.url = url
88
- self.file_map = file_map
89
-
90
-
91
- class Klue(datasets.GeneratorBasedBuilder):
92
- """The General Language Understanding Evaluation (GLUE) benchmark."""
93
-
94
- BUILDER_CONFIGS = [
95
- KlueConfig(
96
- name="ynat",
97
- features={
98
- "guid": datasets.Value("string"),
99
- "title": datasets.Value("string"),
100
- "label": datasets.features.ClassLabel(names=["IT과학", "경제", "사회", "생활문화", "세계", "스포츠", "정치"]),
101
- "url": datasets.Value("string"),
102
- "date": datasets.Value("string"),
103
- },
104
- description=textwrap.dedent(
105
- """\
106
- In topic classification (TC), the goal is to predict the topic of a given text
107
- snippet. We include TC in our KLUE benchmark, as inferring the topic of a text is a key
108
- capability that should be possessed by a language understanding system.
109
- Following a typical single sentence classification task, we introduce YNAT, a Younhap
110
- News Agency news headlines for Topic Classification. For Korean, no dataset has been
111
- proposed for this task, which motivates us to construct the first Korean topic
112
- classification benchmark. In this task, given a news headline, a text classifier must
113
- predict a topic which is one of politics, economy, society, culture, world, IT/science,
114
- and sports. Macro-F1 score is used to evaluate a system."""
115
- ),
116
- data_url=_DATA_URLs["ynat"],
117
- url=_DESCRIPTION_URLs["ynat"],
118
- file_map={
119
- "train": "ynat-v1_train.json",
120
- "dev": "ynat-v1_dev.json",
121
- },
122
- ),
123
- KlueConfig(
124
- name="sts",
125
- features={
126
- "guid": datasets.Value("string"),
127
- "source": datasets.Value("string"),
128
- "sentence1": datasets.Value("string"),
129
- "sentence2": datasets.Value("string"),
130
- "labels": {
131
- "label": datasets.Value("float64"),
132
- "real-label": datasets.Value("float64"),
133
- "binary-label": datasets.ClassLabel(names=["negative", "positive"]),
134
- },
135
- },
136
- description=textwrap.dedent(
137
- """\
138
- STS is a task which aims to predict the semantic similarity of two input sentences as
139
- a real value between 0 and 5. Note that we furthure binarized the prediction scores
140
- into two classes with a threshold score 3.0 (paraphrased or not) and evaluated with
141
- a classification metric.
142
- """
143
- ),
144
- data_url=_DATA_URLs["sts"],
145
- url=_DESCRIPTION_URLs["sts"],
146
- file_map={
147
- "train": "klue-sts-v1_train.json",
148
- "dev": "klue-sts-v1_dev.json",
149
- },
150
- ),
151
- KlueConfig(
152
- name="nli",
153
- features={
154
- "guid": datasets.Value("string"),
155
- "source": datasets.Value("string"),
156
- "premise": datasets.Value("string"),
157
- "hypothesis": datasets.Value("string"),
158
- "label": datasets.ClassLabel(names=["entailment", "neutral", "contradiction"]),
159
- },
160
- description=textwrap.dedent(
161
- """\
162
- NLI is a task to infer the relationship between a hypothesis sentence and a premise
163
- sentence. Given the premise, the model determines if the hypothesis is true (entailment),
164
- false (contradiction), or undetermined (neutral).
165
- """
166
- ),
167
- data_url=_DATA_URLs["nli"],
168
- url=_DESCRIPTION_URLs["nli"],
169
- file_map={
170
- "train": "klue-nli-v1_train.json",
171
- "dev": "klue-nli-v1_dev.json",
172
- },
173
- ),
174
- KlueConfig(
175
- name="ner",
176
- features={
177
- "sentence": datasets.Value("string"),
178
- "tokens": datasets.Sequence(datasets.Value("string")),
179
- "ner_tags": datasets.Sequence(
180
- datasets.ClassLabel(
181
- names=[
182
- "B-DT",
183
- "I-DT",
184
- "B-LC",
185
- "I-LC",
186
- "B-OG",
187
- "I-OG",
188
- "B-PS",
189
- "I-PS",
190
- "B-QT",
191
- "I-QT",
192
- "B-TI",
193
- "I-TI",
194
- "O",
195
- ]
196
- )
197
- ),
198
- },
199
- description=textwrap.dedent(
200
- """\
201
- NER is a task to detect the boundaries of named entities in unstructured text and to
202
- classify the types. A named entity can be of one of predefined entity types such as
203
- person, location, organization, time expressions, quantities and monetary values.
204
- """
205
- ),
206
- data_url=_DATA_URLs["ner"],
207
- url=_DESCRIPTION_URLs["ner"],
208
- file_map={
209
- "train": "klue-ner-v1_train.tsv",
210
- "dev": "klue-ner-v1_dev.tsv",
211
- },
212
- ),
213
- KlueConfig(
214
- name="re",
215
- features={
216
- "guid": datasets.Value("string"),
217
- "sentence": datasets.Value("string"),
218
- "subject_entity": {
219
- "word": datasets.Value("string"),
220
- "start_idx": datasets.Value("int32"),
221
- "end_idx": datasets.Value("int32"),
222
- "type": datasets.Value("string"),
223
- },
224
- "object_entity": {
225
- "word": datasets.Value("string"),
226
- "start_idx": datasets.Value("int32"),
227
- "end_idx": datasets.Value("int32"),
228
- "type": datasets.Value("string"),
229
- },
230
- "label": datasets.ClassLabel(
231
- names=[
232
- "no_relation",
233
- "org:dissolved",
234
- "org:founded",
235
- "org:place_of_headquarters",
236
- "org:alternate_names",
237
- "org:member_of",
238
- "org:members",
239
- "org:political/religious_affiliation",
240
- "org:product",
241
- "org:founded_by",
242
- "org:top_members/employees",
243
- "org:number_of_employees/members",
244
- "per:date_of_birth",
245
- "per:date_of_death",
246
- "per:place_of_birth",
247
- "per:place_of_death",
248
- "per:place_of_residence",
249
- "per:origin",
250
- "per:employee_of",
251
- "per:schools_attended",
252
- "per:alternate_names",
253
- "per:parents",
254
- "per:children",
255
- "per:siblings",
256
- "per:spouse",
257
- "per:other_family",
258
- "per:colleagues",
259
- "per:product",
260
- "per:religion",
261
- "per:title",
262
- ]
263
- ),
264
- "source": datasets.Value("string"),
265
- },
266
- description=textwrap.dedent(
267
- """\
268
- RE is a task to identify semantic relations between entity pairs in a text. The relation
269
- is defined between an entity pair consisting of subject entity and object entity.
270
- The goal is then to pick an appropriate relationship between these two entities.
271
- """
272
- ),
273
- data_url=_DATA_URLs["re"],
274
- url=_DESCRIPTION_URLs["re"],
275
- file_map={
276
- "train": "klue-re-v1_train.json",
277
- "dev": "klue-re-v1_dev.json",
278
- },
279
- ),
280
- KlueConfig(
281
- name="dp",
282
- features={
283
- "sentence": datasets.Value("string"),
284
- "index": [datasets.Value("int32")],
285
- "word_form": [datasets.Value("string")],
286
- "lemma": [datasets.Value("string")],
287
- "pos": [datasets.Value("string")],
288
- "head": [datasets.Value("int32")],
289
- "deprel": [datasets.Value("string")],
290
- },
291
- description=textwrap.dedent(
292
- """\
293
- DP is a task that aims at finding relational information among words.
294
- The goal is to predict a graph structure and a dependency label of an input sentence
295
- based on the dependency grammar.
296
- """
297
- ),
298
- data_url=_DATA_URLs["dp"],
299
- url=_DESCRIPTION_URLs["dp"],
300
- file_map={
301
- "train": "klue-dp-v1_train.tsv",
302
- "dev": "klue-dp-v1_dev.tsv",
303
- },
304
- ),
305
- KlueConfig(
306
- name="mrc",
307
- features={
308
- "title": datasets.Value("string"),
309
- "context": datasets.Value("string"),
310
- "news_category": datasets.Value("string"),
311
- "source": datasets.Value("string"),
312
- "guid": datasets.Value("string"),
313
- "is_impossible": datasets.Value("bool"),
314
- "question_type": datasets.Value("int32"),
315
- "question": datasets.Value("string"),
316
- "answers": datasets.features.Sequence(
317
- {
318
- "answer_start": datasets.Value("int32"),
319
- "text": datasets.Value("string"),
320
- },
321
- ),
322
- },
323
- description=textwrap.dedent(
324
- """\
325
- MRC is a task of evaluating model that can answer a question about a given text
326
- passage. Specifically, we formulate the task as a span prediction task, where the
327
- answer is a text segment (coined as spans) in the passage.
328
- """
329
- ),
330
- data_url=_DATA_URLs["mrc"],
331
- url=_DESCRIPTION_URLs["mrc"],
332
- file_map={
333
- "train": "klue-mrc-v1_train.json",
334
- "dev": "klue-mrc-v1_dev.json",
335
- },
336
- ),
337
- KlueConfig(
338
- name="wos",
339
- features={
340
- "guid": datasets.Value("string"),
341
- "domains": [datasets.Value("string")],
342
- "dialogue": [
343
- {
344
- "role": datasets.Value("string"),
345
- "text": datasets.Value("string"),
346
- "state": [datasets.Value("string")],
347
- }
348
- ],
349
- },
350
- description=textwrap.dedent(
351
- """\
352
- DST is a task to predict slot and value pairs (dialogue states) from a task-oriented
353
- dialogue. The potential pairs are predefined by a given task schema and knowledge
354
- base (KB).
355
- """
356
- ),
357
- data_url=_DATA_URLs["wos"],
358
- url=_DESCRIPTION_URLs["wos"],
359
- file_map={
360
- "train": "wos-v1_train.json",
361
- "dev": "wos-v1_dev.json",
362
- },
363
- ),
364
- ]
365
-
366
- def _info(self):
367
- return datasets.DatasetInfo(
368
- description=_KLUE_DESCRIPTION,
369
- features=datasets.Features(self.config.features),
370
- homepage=self.config.url,
371
- citation=_KLUE_CITATION,
372
- license=_LICENSE,
373
- )
374
-
375
- def _split_generators(self, dl_manager):
376
- archive = dl_manager.download(self.config.data_url)
377
- dir_name = self.config.data_url.split("/")[-1].replace(".tar.gz", "")
378
- return [
379
- datasets.SplitGenerator(
380
- name=datasets.Split.TRAIN,
381
- gen_kwargs={
382
- "data_file": dir_name + "/" + self.config.file_map["train"],
383
- "files": dl_manager.iter_archive(archive),
384
- },
385
- ),
386
- datasets.SplitGenerator(
387
- name=datasets.Split.VALIDATION,
388
- gen_kwargs={
389
- "data_file": dir_name + "/" + self.config.file_map["dev"],
390
- "files": dl_manager.iter_archive(archive),
391
- },
392
- ),
393
- ]
394
-
395
- def _generate_examples(self, data_file, files):
396
- if self.config.name in ["ynat", "sts", "re"]:
397
- for path, f in files:
398
- if path == data_file:
399
- f = json.load(f)
400
- for id_, row in enumerate(f):
401
- features = {key: row[key] for key in row if key in self.config.features}
402
- yield id_, features
403
- break
404
-
405
- if self.config.name == "nli":
406
- for path, f in files:
407
- if path == data_file:
408
- f = json.load(f)
409
- for id_, row in enumerate(f):
410
- # In train file, "source" is written as "genre"
411
- features = {
412
- "guid": row["guid"],
413
- "source": row["source"] if "source" in row else row["genre"],
414
- "premise": row["premise"],
415
- "hypothesis": row["hypothesis"],
416
- "label": row["gold_label"],
417
- }
418
- yield id_, features
419
- break
420
-
421
- if self.config.name == "ner":
422
- for path, f in files:
423
- if path == data_file:
424
- f = (line.decode("utf-8") for line in f)
425
- reader = csv.reader(f, delimiter="\t", quoting=csv.QUOTE_NONE)
426
- for _ in range(5): # skip headers
427
- next(reader)
428
- id_ = -1
429
- for row in reader:
430
- if row:
431
- if row[0].startswith("##"):
432
- id_ += 1
433
- tokens, ner_tags = [], []
434
- sentence = row[1]
435
- else:
436
- tokens.append(row[0])
437
- ner_tags.append(row[1])
438
- else: # new line
439
- assert len(tokens) == len(ner_tags)
440
- yield id_, {"sentence": sentence, "tokens": tokens, "ner_tags": ner_tags}
441
- break
442
-
443
- if self.config.name == "dp":
444
- for path, f in files:
445
- if path == data_file:
446
- f = (line.decode("utf-8") for line in f)
447
- reader = csv.reader(f, delimiter="\t", quoting=csv.QUOTE_NONE)
448
- for _ in range(5): # skip headers
449
- next(reader)
450
- id_ = -1
451
- for row in reader:
452
- if row:
453
- if row[0].startswith("##"):
454
- id_ += 1
455
- index = []
456
- word_form = []
457
- lemma = []
458
- pos = []
459
- head = []
460
- deprel = []
461
- sentence = row[1]
462
- else:
463
- index.append(row[0])
464
- word_form.append(row[1])
465
- lemma.append(row[2])
466
- pos.append(row[3])
467
- head.append(row[4])
468
- deprel.append(row[5])
469
- else: # new line
470
- assert len(index) == len(word_form) == len(lemma) == len(pos) == len(head) == len(deprel)
471
- yield id_, {
472
- "sentence": sentence,
473
- "index": index,
474
- "word_form": word_form,
475
- "lemma": lemma,
476
- "pos": pos,
477
- "head": head,
478
- "deprel": deprel,
479
- }
480
- break
481
-
482
- if self.config.name == "mrc":
483
- for path, f in files:
484
- if path == data_file:
485
- f = json.load(f)
486
- id_ = -1
487
- for example in f["data"]:
488
- title = example.get("title", "")
489
- news_category = example.get("news_category", "")
490
- source = example["source"]
491
- for paragraph in example["paragraphs"]:
492
- context = paragraph["context"].strip()
493
- for qa in paragraph["qas"]:
494
- guid = qa["guid"]
495
- question_type = qa["question_type"]
496
- is_impossible = qa["is_impossible"]
497
- question = qa["question"].strip()
498
-
499
- if "plausible_answers" in qa:
500
- qa["answers"].extend(qa["plausible_answers"])
501
- answer_starts = [answer["answer_start"] for answer in qa["answers"]]
502
- answers = [answer["text"].strip() for answer in qa["answers"]]
503
- id_ += 1
504
-
505
- yield id_, {
506
- "guid": guid,
507
- "title": title,
508
- "context": context,
509
- "news_category": news_category,
510
- "source": source,
511
- "question_type": question_type,
512
- "is_impossible": is_impossible,
513
- "question": question,
514
- "answers": {
515
- "answer_start": answer_starts,
516
- "text": answers,
517
- },
518
- }
519
- break
520
-
521
- if self.config.name == "wos":
522
- for path, f in files:
523
- if path == data_file:
524
- f = json.load(f)
525
- for id_, row in enumerate(f):
526
- guid = row["guid"]
527
- domains = row["domains"]
528
- dialogue = row["dialogue"]
529
- for utterance in dialogue:
530
- if "state" not in utterance:
531
- utterance["state"] = []
532
- yield id_, {"guid": guid, "domains": domains, "dialogue": dialogue}
533
- break
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
mrc/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45fcfd50b99aed4cd55c8c5856e42aa4a5139808cc36b36994b3c1991aca93dd
3
+ size 21415334
mrc/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f576c924f471702a6fa3369d314bea2e47545c56736f371a2adb90ed15913738
3
+ size 8683138
ner/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:291ded874432fa413e5007ec3770b41bb2e76f44a4e42cc7cc22cb4ebd8c27b6
3
+ size 4209983
ner/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db2129c2ae38fc9a7de630d902cd41f80ac6174f620a8318b4d80569e31b9f84
3
+ size 1055904
nli/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35a8eb31818b94d9cfcb3477914360ce038b46b3bce4075eaef92309caa2c07c
3
+ size 1832150
nli/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b809f8f0aac2a2f41cd095bf15d8dbca9c931d7539390b6ef6241ca976195423
3
+ size 223966
re/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:329e15f482c794900d53627730448945607f2edb12397811eb6f671bf44d57da
3
+ size 6647272
re/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10aade1fdb0244eb867eb8a7e3459d391546016c792989eb1030e23603a28531
3
+ size 1542985
sts/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa624db134d4242f24e232432055f09f34652c3852eb77ff98ca4784edfed184
3
+ size 1519075
sts/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67611cf65f0d35feb244af81a1609186c55cc3a4aee293d1ceb02fb0d6b721ea
3
+ size 68780
wos/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36003fc8ef605306180591e708ace9043b30ec8fbc0c657ed1f01d53cd4a8a01
3
+ size 5643747
wos/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5342b51a1565aa9719adcb02125c8719c6ba8d3dda5be5c948e642db451fbdf
3
+ size 715108
ynat/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:062c3b51c1ca34ed23c8fd19ffa5ea0ddcd914a95aeb9077d89576d3ef71d123
3
+ size 4165783
ynat/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:521a6785a2492a9142474ad23ae038b1a6f333fa7e39aab8fd9dd19a20c9ce92
3
+ size 846520