Datasets:
Languages:
Korean
Multilinguality:
monolingual
Size Categories:
10K<n<100K
Language Creators:
expert-generated
Annotations Creators:
expert-generated
Source Datasets:
original
ArXiv:
License:
Commit
•
349481e
1
Parent(s):
750599f
Convert dataset to Parquet (#2)
Browse files- Convert dataset to Parquet (aa733cddfbffca00259b4f4a058b92d74b65380c)
- Add sts data files (b7f730b37c6eaa998178d7550ac35827fc0bf5ca)
- Add nli data files (0e05c2a3649a814bc206ae51d13d0a38ac991f64)
- Add ner data files (c302a004d918a49b00002e45475a0aeaa239aa1e)
- Add re data files (26a2ff79dc50e1b253243f1a511b031f56a38499)
- Add dp data files (e0e0e06ec993811871d9d81a2777a7fc8a61896c)
- Add mrc data files (9dc841ba15d7f362b0e55b4579307319e3cc167b)
- Add wos data files (5454eaaa94dafb629516e236faa87dbe0501cb09)
- Delete loading script (89d2a4e6e716de45c3e19ffe056665df11b3aac6)
- Delete legacy dataset_infos.json (2e32b5c21562b629dbd5140500af66e5db303bf9)
- README.md +190 -141
- dataset_infos.json +0 -1
- dp/train-00000-of-00001.parquet +3 -0
- dp/validation-00000-of-00001.parquet +3 -0
- klue.py +0 -533
- mrc/train-00000-of-00001.parquet +3 -0
- mrc/validation-00000-of-00001.parquet +3 -0
- ner/train-00000-of-00001.parquet +3 -0
- ner/validation-00000-of-00001.parquet +3 -0
- nli/train-00000-of-00001.parquet +3 -0
- nli/validation-00000-of-00001.parquet +3 -0
- re/train-00000-of-00001.parquet +3 -0
- re/validation-00000-of-00001.parquet +3 -0
- sts/train-00000-of-00001.parquet +3 -0
- sts/validation-00000-of-00001.parquet +3 -0
- wos/train-00000-of-00001.parquet +3 -0
- wos/validation-00000-of-00001.parquet +3 -0
- ynat/train-00000-of-00001.parquet +3 -0
- ynat/validation-00000-of-00001.parquet +3 -0
README.md
CHANGED
@@ -29,96 +29,76 @@ task_ids:
|
|
29 |
- topic-classification
|
30 |
paperswithcode_id: klue
|
31 |
pretty_name: KLUE
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
tags:
|
33 |
- relation-extraction
|
34 |
dataset_info:
|
35 |
-
- config_name:
|
36 |
features:
|
37 |
-
- name:
|
38 |
-
dtype: string
|
39 |
-
- name: title
|
40 |
-
dtype: string
|
41 |
-
- name: label
|
42 |
-
dtype:
|
43 |
-
class_label:
|
44 |
-
names:
|
45 |
-
'0': IT과학
|
46 |
-
'1': 경제
|
47 |
-
'2': 사회
|
48 |
-
'3': 생활문화
|
49 |
-
'4': 세계
|
50 |
-
'5': 스포츠
|
51 |
-
'6': 정치
|
52 |
-
- name: url
|
53 |
-
dtype: string
|
54 |
-
- name: date
|
55 |
dtype: string
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
56 |
splits:
|
57 |
- name: train
|
58 |
-
num_bytes:
|
59 |
-
num_examples:
|
60 |
- name: validation
|
61 |
-
num_bytes:
|
62 |
-
num_examples:
|
63 |
-
download_size:
|
64 |
-
dataset_size:
|
65 |
-
- config_name:
|
66 |
features:
|
67 |
-
- name:
|
68 |
-
dtype: string
|
69 |
-
- name: source
|
70 |
-
dtype: string
|
71 |
-
- name: sentence1
|
72 |
dtype: string
|
73 |
-
- name:
|
74 |
dtype: string
|
75 |
-
- name:
|
76 |
-
struct:
|
77 |
-
- name: label
|
78 |
-
dtype: float64
|
79 |
-
- name: real-label
|
80 |
-
dtype: float64
|
81 |
-
- name: binary-label
|
82 |
-
dtype:
|
83 |
-
class_label:
|
84 |
-
names:
|
85 |
-
'0': negative
|
86 |
-
'1': positive
|
87 |
-
splits:
|
88 |
-
- name: train
|
89 |
-
num_bytes: 2832921
|
90 |
-
num_examples: 11668
|
91 |
-
- name: validation
|
92 |
-
num_bytes: 122657
|
93 |
-
num_examples: 519
|
94 |
-
download_size: 1349875
|
95 |
-
dataset_size: 2955578
|
96 |
-
- config_name: nli
|
97 |
-
features:
|
98 |
-
- name: guid
|
99 |
dtype: string
|
100 |
- name: source
|
101 |
dtype: string
|
102 |
-
- name:
|
103 |
dtype: string
|
104 |
-
- name:
|
|
|
|
|
|
|
|
|
105 |
dtype: string
|
106 |
-
- name:
|
107 |
-
|
108 |
-
|
109 |
-
|
110 |
-
|
111 |
-
|
112 |
-
'2': contradiction
|
113 |
splits:
|
114 |
- name: train
|
115 |
-
num_bytes:
|
116 |
-
num_examples:
|
117 |
- name: validation
|
118 |
-
num_bytes:
|
119 |
-
num_examples:
|
120 |
-
download_size:
|
121 |
-
dataset_size:
|
122 |
- config_name: ner
|
123 |
features:
|
124 |
- name: sentence
|
@@ -144,13 +124,39 @@ dataset_info:
|
|
144 |
'12': O
|
145 |
splits:
|
146 |
- name: train
|
147 |
-
num_bytes:
|
148 |
num_examples: 21008
|
149 |
- name: validation
|
150 |
-
num_bytes:
|
151 |
num_examples: 5000
|
152 |
-
download_size:
|
153 |
-
dataset_size:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
154 |
- config_name: re
|
155 |
features:
|
156 |
- name: guid
|
@@ -215,71 +221,44 @@ dataset_info:
|
|
215 |
dtype: string
|
216 |
splits:
|
217 |
- name: train
|
218 |
-
num_bytes:
|
219 |
num_examples: 32470
|
220 |
- name: validation
|
221 |
-
num_bytes:
|
222 |
num_examples: 7765
|
223 |
-
download_size:
|
224 |
-
dataset_size:
|
225 |
-
- config_name:
|
226 |
-
features:
|
227 |
-
- name: sentence
|
228 |
-
dtype: string
|
229 |
-
- name: index
|
230 |
-
list: int32
|
231 |
-
- name: word_form
|
232 |
-
list: string
|
233 |
-
- name: lemma
|
234 |
-
list: string
|
235 |
-
- name: pos
|
236 |
-
list: string
|
237 |
-
- name: head
|
238 |
-
list: int32
|
239 |
-
- name: deprel
|
240 |
-
list: string
|
241 |
-
splits:
|
242 |
-
- name: train
|
243 |
-
num_bytes: 7900009
|
244 |
-
num_examples: 10000
|
245 |
-
- name: validation
|
246 |
-
num_bytes: 1557506
|
247 |
-
num_examples: 2000
|
248 |
-
download_size: 2033461
|
249 |
-
dataset_size: 9457515
|
250 |
-
- config_name: mrc
|
251 |
features:
|
252 |
-
- name:
|
253 |
-
dtype: string
|
254 |
-
- name: context
|
255 |
-
dtype: string
|
256 |
-
- name: news_category
|
257 |
dtype: string
|
258 |
- name: source
|
259 |
dtype: string
|
260 |
-
- name:
|
261 |
dtype: string
|
262 |
-
- name:
|
263 |
-
dtype: bool
|
264 |
-
- name: question_type
|
265 |
-
dtype: int32
|
266 |
-
- name: question
|
267 |
dtype: string
|
268 |
-
- name:
|
269 |
-
|
270 |
-
- name:
|
271 |
-
dtype:
|
272 |
-
- name:
|
273 |
-
dtype:
|
|
|
|
|
|
|
|
|
|
|
|
|
274 |
splits:
|
275 |
- name: train
|
276 |
-
num_bytes:
|
277 |
-
num_examples:
|
278 |
- name: validation
|
279 |
-
num_bytes:
|
280 |
-
num_examples:
|
281 |
-
download_size:
|
282 |
-
dataset_size:
|
283 |
- config_name: wos
|
284 |
features:
|
285 |
- name: guid
|
@@ -296,22 +275,92 @@ dataset_info:
|
|
296 |
list: string
|
297 |
splits:
|
298 |
- name: train
|
299 |
-
num_bytes:
|
300 |
num_examples: 8000
|
301 |
- name: validation
|
302 |
-
num_bytes:
|
303 |
num_examples: 1000
|
304 |
-
download_size:
|
305 |
-
dataset_size:
|
306 |
-
|
307 |
-
|
308 |
-
-
|
309 |
-
|
310 |
-
-
|
311 |
-
|
312 |
-
-
|
313 |
-
|
314 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
315 |
---
|
316 |
|
317 |
# Dataset Card for KLUE
|
29 |
- topic-classification
|
30 |
paperswithcode_id: klue
|
31 |
pretty_name: KLUE
|
32 |
+
config_names:
|
33 |
+
- dp
|
34 |
+
- mrc
|
35 |
+
- ner
|
36 |
+
- nli
|
37 |
+
- re
|
38 |
+
- sts
|
39 |
+
- wos
|
40 |
+
- ynat
|
41 |
tags:
|
42 |
- relation-extraction
|
43 |
dataset_info:
|
44 |
+
- config_name: dp
|
45 |
features:
|
46 |
+
- name: sentence
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
47 |
dtype: string
|
48 |
+
- name: index
|
49 |
+
list: int32
|
50 |
+
- name: word_form
|
51 |
+
list: string
|
52 |
+
- name: lemma
|
53 |
+
list: string
|
54 |
+
- name: pos
|
55 |
+
list: string
|
56 |
+
- name: head
|
57 |
+
list: int32
|
58 |
+
- name: deprel
|
59 |
+
list: string
|
60 |
splits:
|
61 |
- name: train
|
62 |
+
num_bytes: 7899965
|
63 |
+
num_examples: 10000
|
64 |
- name: validation
|
65 |
+
num_bytes: 1557462
|
66 |
+
num_examples: 2000
|
67 |
+
download_size: 3742577
|
68 |
+
dataset_size: 9457427
|
69 |
+
- config_name: mrc
|
70 |
features:
|
71 |
+
- name: title
|
|
|
|
|
|
|
|
|
72 |
dtype: string
|
73 |
+
- name: context
|
74 |
dtype: string
|
75 |
+
- name: news_category
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
76 |
dtype: string
|
77 |
- name: source
|
78 |
dtype: string
|
79 |
+
- name: guid
|
80 |
dtype: string
|
81 |
+
- name: is_impossible
|
82 |
+
dtype: bool
|
83 |
+
- name: question_type
|
84 |
+
dtype: int32
|
85 |
+
- name: question
|
86 |
dtype: string
|
87 |
+
- name: answers
|
88 |
+
sequence:
|
89 |
+
- name: answer_start
|
90 |
+
dtype: int32
|
91 |
+
- name: text
|
92 |
+
dtype: string
|
|
|
93 |
splits:
|
94 |
- name: train
|
95 |
+
num_bytes: 46505593
|
96 |
+
num_examples: 17554
|
97 |
- name: validation
|
98 |
+
num_bytes: 15583017
|
99 |
+
num_examples: 5841
|
100 |
+
download_size: 30098472
|
101 |
+
dataset_size: 62088610
|
102 |
- config_name: ner
|
103 |
features:
|
104 |
- name: sentence
|
124 |
'12': O
|
125 |
splits:
|
126 |
- name: train
|
127 |
+
num_bytes: 19891905
|
128 |
num_examples: 21008
|
129 |
- name: validation
|
130 |
+
num_bytes: 4937563
|
131 |
num_examples: 5000
|
132 |
+
download_size: 5265887
|
133 |
+
dataset_size: 24829468
|
134 |
+
- config_name: nli
|
135 |
+
features:
|
136 |
+
- name: guid
|
137 |
+
dtype: string
|
138 |
+
- name: source
|
139 |
+
dtype: string
|
140 |
+
- name: premise
|
141 |
+
dtype: string
|
142 |
+
- name: hypothesis
|
143 |
+
dtype: string
|
144 |
+
- name: label
|
145 |
+
dtype:
|
146 |
+
class_label:
|
147 |
+
names:
|
148 |
+
'0': entailment
|
149 |
+
'1': neutral
|
150 |
+
'2': contradiction
|
151 |
+
splits:
|
152 |
+
- name: train
|
153 |
+
num_bytes: 5719882
|
154 |
+
num_examples: 24998
|
155 |
+
- name: validation
|
156 |
+
num_bytes: 673260
|
157 |
+
num_examples: 3000
|
158 |
+
download_size: 2056116
|
159 |
+
dataset_size: 6393142
|
160 |
- config_name: re
|
161 |
features:
|
162 |
- name: guid
|
221 |
dtype: string
|
222 |
splits:
|
223 |
- name: train
|
224 |
+
num_bytes: 11145426
|
225 |
num_examples: 32470
|
226 |
- name: validation
|
227 |
+
num_bytes: 2559272
|
228 |
num_examples: 7765
|
229 |
+
download_size: 8190257
|
230 |
+
dataset_size: 13704698
|
231 |
+
- config_name: sts
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
232 |
features:
|
233 |
+
- name: guid
|
|
|
|
|
|
|
|
|
234 |
dtype: string
|
235 |
- name: source
|
236 |
dtype: string
|
237 |
+
- name: sentence1
|
238 |
dtype: string
|
239 |
+
- name: sentence2
|
|
|
|
|
|
|
|
|
240 |
dtype: string
|
241 |
+
- name: labels
|
242 |
+
struct:
|
243 |
+
- name: label
|
244 |
+
dtype: float64
|
245 |
+
- name: real-label
|
246 |
+
dtype: float64
|
247 |
+
- name: binary-label
|
248 |
+
dtype:
|
249 |
+
class_label:
|
250 |
+
names:
|
251 |
+
'0': negative
|
252 |
+
'1': positive
|
253 |
splits:
|
254 |
- name: train
|
255 |
+
num_bytes: 2832889
|
256 |
+
num_examples: 11668
|
257 |
- name: validation
|
258 |
+
num_bytes: 122641
|
259 |
+
num_examples: 519
|
260 |
+
download_size: 1587855
|
261 |
+
dataset_size: 2955530
|
262 |
- config_name: wos
|
263 |
features:
|
264 |
- name: guid
|
275 |
list: string
|
276 |
splits:
|
277 |
- name: train
|
278 |
+
num_bytes: 26676970
|
279 |
num_examples: 8000
|
280 |
- name: validation
|
281 |
+
num_bytes: 3488911
|
282 |
num_examples: 1000
|
283 |
+
download_size: 6358855
|
284 |
+
dataset_size: 30165881
|
285 |
+
- config_name: ynat
|
286 |
+
features:
|
287 |
+
- name: guid
|
288 |
+
dtype: string
|
289 |
+
- name: title
|
290 |
+
dtype: string
|
291 |
+
- name: label
|
292 |
+
dtype:
|
293 |
+
class_label:
|
294 |
+
names:
|
295 |
+
'0': IT과학
|
296 |
+
'1': 경제
|
297 |
+
'2': 사회
|
298 |
+
'3': 생활문화
|
299 |
+
'4': 세계
|
300 |
+
'5': 스포츠
|
301 |
+
'6': 정치
|
302 |
+
- name: url
|
303 |
+
dtype: string
|
304 |
+
- name: date
|
305 |
+
dtype: string
|
306 |
+
splits:
|
307 |
+
- name: train
|
308 |
+
num_bytes: 10109584
|
309 |
+
num_examples: 45678
|
310 |
+
- name: validation
|
311 |
+
num_bytes: 2039181
|
312 |
+
num_examples: 9107
|
313 |
+
download_size: 5012303
|
314 |
+
dataset_size: 12148765
|
315 |
+
configs:
|
316 |
+
- config_name: dp
|
317 |
+
data_files:
|
318 |
+
- split: train
|
319 |
+
path: dp/train-*
|
320 |
+
- split: validation
|
321 |
+
path: dp/validation-*
|
322 |
+
- config_name: mrc
|
323 |
+
data_files:
|
324 |
+
- split: train
|
325 |
+
path: mrc/train-*
|
326 |
+
- split: validation
|
327 |
+
path: mrc/validation-*
|
328 |
+
- config_name: ner
|
329 |
+
data_files:
|
330 |
+
- split: train
|
331 |
+
path: ner/train-*
|
332 |
+
- split: validation
|
333 |
+
path: ner/validation-*
|
334 |
+
- config_name: nli
|
335 |
+
data_files:
|
336 |
+
- split: train
|
337 |
+
path: nli/train-*
|
338 |
+
- split: validation
|
339 |
+
path: nli/validation-*
|
340 |
+
- config_name: re
|
341 |
+
data_files:
|
342 |
+
- split: train
|
343 |
+
path: re/train-*
|
344 |
+
- split: validation
|
345 |
+
path: re/validation-*
|
346 |
+
- config_name: sts
|
347 |
+
data_files:
|
348 |
+
- split: train
|
349 |
+
path: sts/train-*
|
350 |
+
- split: validation
|
351 |
+
path: sts/validation-*
|
352 |
+
- config_name: wos
|
353 |
+
data_files:
|
354 |
+
- split: train
|
355 |
+
path: wos/train-*
|
356 |
+
- split: validation
|
357 |
+
path: wos/validation-*
|
358 |
+
- config_name: ynat
|
359 |
+
data_files:
|
360 |
+
- split: train
|
361 |
+
path: ynat/train-*
|
362 |
+
- split: validation
|
363 |
+
path: ynat/validation-*
|
364 |
---
|
365 |
|
366 |
# Dataset Card for KLUE
|
dataset_infos.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"ynat": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/66/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "title": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 7, "names": ["IT\uacfc\ud559", "\uacbd\uc81c", "\uc0ac\ud68c", "\uc0dd\ud65c\ubb38\ud654", "\uc138\uacc4", "\uc2a4\ud3ec\uce20", "\uc815\uce58"], "names_file": null, "id": null, "_type": "ClassLabel"}, "url": {"dtype": "string", "id": null, "_type": "Value"}, "date": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "ynat", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 10109664, "num_examples": 45678, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 2039197, "num_examples": 9107, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000066/data/ynat-v1.tar.gz": {"num_bytes": 4932555, "checksum": "820a4d1d6d1fd83e2a421f856965d3cfc5c93627935ce8c5b27468c6113fc482"}}, "download_size": 4932555, "post_processing_size": null, "dataset_size": 12148861, "size_in_bytes": 17081416}, "sts": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/67/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "source": {"dtype": "string", "id": null, "_type": "Value"}, "sentence1": {"dtype": "string", "id": null, "_type": "Value"}, "sentence2": {"dtype": "string", "id": null, "_type": "Value"}, "labels": {"label": {"dtype": "float64", "id": null, "_type": "Value"}, "real-label": {"dtype": "float64", "id": null, "_type": "Value"}, "binary-label": {"num_classes": 2, "names": ["negative", "positive"], "names_file": null, "id": null, "_type": "ClassLabel"}}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "sts", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2832921, "num_examples": 11668, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 122657, "num_examples": 519, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000067/data/klue-sts-v1.tar.gz": {"num_bytes": 1349875, "checksum": "539341ba78a3b351c686cf70a448ac7a5886ed95f0719d5e3d2378ba703213bd"}}, "download_size": 1349875, "post_processing_size": null, "dataset_size": 2955578, "size_in_bytes": 4305453}, "nli": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/68/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "source": {"dtype": "string", "id": null, "_type": "Value"}, "premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "nli", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 5719930, "num_examples": 24998, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 673276, "num_examples": 3000, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000068/data/klue-nli-v1.tar.gz": {"num_bytes": 1257374, "checksum": "388be2033ef712072201903795a35b4f86826ee3ed3b62dc0c98e1721baa8850"}}, "download_size": 1257374, "post_processing_size": null, "dataset_size": 6393206, "size_in_bytes": 7650580}, "ner": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/69/overview/description", "license": "CC-BY-SA-4.0", "features": {"sentence": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 13, "names": ["B-DT", "I-DT", "B-LC", "I-LC", "B-OG", "I-OG", "B-PS", "I-PS", "B-QT", "I-QT", "B-TI", "I-TI", "O"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "ner", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 19891953, "num_examples": 21008, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 4937579, "num_examples": 5000, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000069/data/klue-ner-v1.tar.gz": {"num_bytes": 4308644, "checksum": "848a89759ac6b7c149c9a00d820726fe2a140c22782201f1a40d856672e7ea8e"}}, "download_size": 4308644, "post_processing_size": null, "dataset_size": 24829532, "size_in_bytes": 29138176}, "re": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/70/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "sentence": {"dtype": "string", "id": null, "_type": "Value"}, "subject_entity": {"word": {"dtype": "string", "id": null, "_type": "Value"}, "start_idx": {"dtype": "int32", "id": null, "_type": "Value"}, "end_idx": {"dtype": "int32", "id": null, "_type": "Value"}, "type": {"dtype": "string", "id": null, "_type": "Value"}}, "object_entity": {"word": {"dtype": "string", "id": null, "_type": "Value"}, "start_idx": {"dtype": "int32", "id": null, "_type": "Value"}, "end_idx": {"dtype": "int32", "id": null, "_type": "Value"}, "type": {"dtype": "string", "id": null, "_type": "Value"}}, "label": {"num_classes": 30, "names": ["no_relation", "org:dissolved", "org:founded", "org:place_of_headquarters", "org:alternate_names", "org:member_of", "org:members", "org:political/religious_affiliation", "org:product", "org:founded_by", "org:top_members/employees", "org:number_of_employees/members", "per:date_of_birth", "per:date_of_death", "per:place_of_birth", "per:place_of_death", "per:place_of_residence", "per:origin", "per:employee_of", "per:schools_attended", "per:alternate_names", "per:parents", "per:children", "per:siblings", "per:spouse", "per:other_family", "per:colleagues", "per:product", "per:religion", "per:title"], "names_file": null, "id": null, "_type": "ClassLabel"}, "source": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "re", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 11145538, "num_examples": 32470, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 2559300, "num_examples": 7765, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000070/data/klue-re-v1.tar.gz": {"num_bytes": 5669259, "checksum": "b09ceac0d986cc09e42fcda9c7f2873c0eea8ec0629baf91fead36580790f8f5"}}, "download_size": 5669259, "post_processing_size": null, "dataset_size": 13704838, "size_in_bytes": 19374097}, "dp": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/71/overview/description", "license": "CC-BY-SA-4.0", "features": {"sentence": {"dtype": "string", "id": null, "_type": "Value"}, "index": [{"dtype": "int32", "id": null, "_type": "Value"}], "word_form": [{"dtype": "string", "id": null, "_type": "Value"}], "lemma": [{"dtype": "string", "id": null, "_type": "Value"}], "pos": [{"dtype": "string", "id": null, "_type": "Value"}], "head": [{"dtype": "int32", "id": null, "_type": "Value"}], "deprel": [{"dtype": "string", "id": null, "_type": "Value"}]}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "dp", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 7900009, "num_examples": 10000, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 1557506, "num_examples": 2000, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000071/data/klue-dp-v1.tar.gz": {"num_bytes": 2033461, "checksum": "2c76a3543a50599ac6640ad360ba00eac36e0b5b2363f708a614d6e50844d17b"}}, "download_size": 2033461, "post_processing_size": null, "dataset_size": 9457515, "size_in_bytes": 11490976}, "mrc": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/72/overview/description", "license": "CC-BY-SA-4.0", "features": {"title": {"dtype": "string", "id": null, "_type": "Value"}, "context": {"dtype": "string", "id": null, "_type": "Value"}, "news_category": {"dtype": "string", "id": null, "_type": "Value"}, "source": {"dtype": "string", "id": null, "_type": "Value"}, "guid": {"dtype": "string", "id": null, "_type": "Value"}, "is_impossible": {"dtype": "bool", "id": null, "_type": "Value"}, "question_type": {"dtype": "int32", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "answers": {"feature": {"answer_start": {"dtype": "int32", "id": null, "_type": "Value"}, "text": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "mrc", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 46505665, "num_examples": 17554, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 15583053, "num_examples": 5841, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000072/data/klue-mrc-v1.tar.gz": {"num_bytes": 19218422, "checksum": "a444af252901452380d58a6320908ce4a86759bb6f38ad95d0ca98584ad33d14"}}, "download_size": 19218422, "post_processing_size": null, "dataset_size": 62088718, "size_in_bytes": 81307140}, "wos": {"description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n", "citation": "@misc{park2021klue,\n title={KLUE: Korean Language Understanding Evaluation},\n author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},\n year={2021},\n eprint={2105.09680},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "https://klue-benchmark.com/tasks/73/overview/description", "license": "CC-BY-SA-4.0", "features": {"guid": {"dtype": "string", "id": null, "_type": "Value"}, "domains": [{"dtype": "string", "id": null, "_type": "Value"}], "dialogue": [{"role": {"dtype": "string", "id": null, "_type": "Value"}, "text": {"dtype": "string", "id": null, "_type": "Value"}, "state": [{"dtype": "string", "id": null, "_type": "Value"}]}]}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "klue", "config_name": "wos", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 26677002, "num_examples": 8000, "dataset_name": "klue"}, "validation": {"name": "validation", "num_bytes": 3488943, "num_examples": 1000, "dataset_name": "klue"}}, "download_checksums": {"http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000073/data/wos-v1.tar.gz": {"num_bytes": 4785657, "checksum": "da17829300271560afc6e7fc330503c2ca6f7ae7721d9bb94308579542a5871f"}}, "download_size": 4785657, "post_processing_size": null, "dataset_size": 30165945, "size_in_bytes": 34951602}}
|
|
dp/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3f570c15f1da1b70d353da6a5896dc0a4f32d2b178e23a4def7b6e953dff30d
|
3 |
+
size 3124932
|
dp/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f43905816424723d15c63fe9ea724fb2c7dacfb88289b2b4e345f259d6119f1c
|
3 |
+
size 617645
|
klue.py
DELETED
@@ -1,533 +0,0 @@
|
|
1 |
-
# coding=utf-8
|
2 |
-
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
|
3 |
-
#
|
4 |
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
5 |
-
# you may not use this file except in compliance with the License.
|
6 |
-
# You may obtain a copy of the License at
|
7 |
-
#
|
8 |
-
# http://www.apache.org/licenses/LICENSE-2.0
|
9 |
-
#
|
10 |
-
# Unless required by applicable law or agreed to in writing, software
|
11 |
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
12 |
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
-
# See the License for the specific language governing permissions and
|
14 |
-
# limitations under the License.
|
15 |
-
|
16 |
-
"""KLUE (Korean Language Understanding Evaluation) benchmark."""
|
17 |
-
|
18 |
-
|
19 |
-
import csv
|
20 |
-
import json
|
21 |
-
import textwrap
|
22 |
-
|
23 |
-
import datasets
|
24 |
-
|
25 |
-
|
26 |
-
_KLUE_CITATION = """\
|
27 |
-
@misc{park2021klue,
|
28 |
-
title={KLUE: Korean Language Understanding Evaluation},
|
29 |
-
author={Sungjoon Park and Jihyung Moon and Sungdong Kim and Won Ik Cho and Jiyoon Han and Jangwon Park and Chisung Song and Junseong Kim and Yongsook Song and Taehwan Oh and Joohong Lee and Juhyun Oh and Sungwon Lyu and Younghoon Jeong and Inkwon Lee and Sangwoo Seo and Dongjun Lee and Hyunwoo Kim and Myeonghwa Lee and Seongbo Jang and Seungwon Do and Sunkyoung Kim and Kyungtae Lim and Jongwon Lee and Kyumin Park and Jamin Shin and Seonghyun Kim and Lucy Park and Alice Oh and Jungwoo Ha and Kyunghyun Cho},
|
30 |
-
year={2021},
|
31 |
-
eprint={2105.09680},
|
32 |
-
archivePrefix={arXiv},
|
33 |
-
primaryClass={cs.CL}
|
34 |
-
}
|
35 |
-
"""
|
36 |
-
|
37 |
-
_KLUE_DESCRIPTION = """\
|
38 |
-
KLUE (Korean Language Understanding Evaluation)
|
39 |
-
Korean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language
|
40 |
-
understanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible
|
41 |
-
to anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain
|
42 |
-
unambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics
|
43 |
-
for every task, thus establishing fair comparison across Korean language models.
|
44 |
-
"""
|
45 |
-
|
46 |
-
_DATA_URLs = {
|
47 |
-
"ynat": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000066/data/ynat-v1.tar.gz",
|
48 |
-
"sts": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000067/data/klue-sts-v1.tar.gz",
|
49 |
-
"nli": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000068/data/klue-nli-v1.tar.gz",
|
50 |
-
"ner": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000069/data/klue-ner-v1.tar.gz",
|
51 |
-
"re": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000070/data/klue-re-v1.tar.gz",
|
52 |
-
"dp": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000071/data/klue-dp-v1.tar.gz",
|
53 |
-
"mrc": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000072/data/klue-mrc-v1.tar.gz",
|
54 |
-
"wos": "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000073/data/wos-v1.tar.gz",
|
55 |
-
}
|
56 |
-
|
57 |
-
_DESCRIPTION_URLs = {
|
58 |
-
"ynat": "https://klue-benchmark.com/tasks/66/overview/description",
|
59 |
-
"sts": "https://klue-benchmark.com/tasks/67/overview/description",
|
60 |
-
"nli": "https://klue-benchmark.com/tasks/68/overview/description",
|
61 |
-
"ner": "https://klue-benchmark.com/tasks/69/overview/description",
|
62 |
-
"re": "https://klue-benchmark.com/tasks/70/overview/description",
|
63 |
-
"dp": "https://klue-benchmark.com/tasks/71/overview/description",
|
64 |
-
"mrc": "https://klue-benchmark.com/tasks/72/overview/description",
|
65 |
-
"wos": "https://klue-benchmark.com/tasks/73/overview/description",
|
66 |
-
}
|
67 |
-
|
68 |
-
_LICENSE = "CC-BY-SA-4.0"
|
69 |
-
|
70 |
-
|
71 |
-
class KlueConfig(datasets.BuilderConfig):
|
72 |
-
"""BuilderConfig for KLUE."""
|
73 |
-
|
74 |
-
def __init__(
|
75 |
-
self,
|
76 |
-
features,
|
77 |
-
data_url,
|
78 |
-
url,
|
79 |
-
file_map,
|
80 |
-
**kwargs,
|
81 |
-
):
|
82 |
-
"""BuilderConfig for KLUE."""
|
83 |
-
|
84 |
-
super(KlueConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs)
|
85 |
-
self.features = features
|
86 |
-
self.data_url = data_url
|
87 |
-
self.url = url
|
88 |
-
self.file_map = file_map
|
89 |
-
|
90 |
-
|
91 |
-
class Klue(datasets.GeneratorBasedBuilder):
|
92 |
-
"""The General Language Understanding Evaluation (GLUE) benchmark."""
|
93 |
-
|
94 |
-
BUILDER_CONFIGS = [
|
95 |
-
KlueConfig(
|
96 |
-
name="ynat",
|
97 |
-
features={
|
98 |
-
"guid": datasets.Value("string"),
|
99 |
-
"title": datasets.Value("string"),
|
100 |
-
"label": datasets.features.ClassLabel(names=["IT과학", "경제", "사회", "생활문화", "세계", "스포츠", "정치"]),
|
101 |
-
"url": datasets.Value("string"),
|
102 |
-
"date": datasets.Value("string"),
|
103 |
-
},
|
104 |
-
description=textwrap.dedent(
|
105 |
-
"""\
|
106 |
-
In topic classification (TC), the goal is to predict the topic of a given text
|
107 |
-
snippet. We include TC in our KLUE benchmark, as inferring the topic of a text is a key
|
108 |
-
capability that should be possessed by a language understanding system.
|
109 |
-
Following a typical single sentence classification task, we introduce YNAT, a Younhap
|
110 |
-
News Agency news headlines for Topic Classification. For Korean, no dataset has been
|
111 |
-
proposed for this task, which motivates us to construct the first Korean topic
|
112 |
-
classification benchmark. In this task, given a news headline, a text classifier must
|
113 |
-
predict a topic which is one of politics, economy, society, culture, world, IT/science,
|
114 |
-
and sports. Macro-F1 score is used to evaluate a system."""
|
115 |
-
),
|
116 |
-
data_url=_DATA_URLs["ynat"],
|
117 |
-
url=_DESCRIPTION_URLs["ynat"],
|
118 |
-
file_map={
|
119 |
-
"train": "ynat-v1_train.json",
|
120 |
-
"dev": "ynat-v1_dev.json",
|
121 |
-
},
|
122 |
-
),
|
123 |
-
KlueConfig(
|
124 |
-
name="sts",
|
125 |
-
features={
|
126 |
-
"guid": datasets.Value("string"),
|
127 |
-
"source": datasets.Value("string"),
|
128 |
-
"sentence1": datasets.Value("string"),
|
129 |
-
"sentence2": datasets.Value("string"),
|
130 |
-
"labels": {
|
131 |
-
"label": datasets.Value("float64"),
|
132 |
-
"real-label": datasets.Value("float64"),
|
133 |
-
"binary-label": datasets.ClassLabel(names=["negative", "positive"]),
|
134 |
-
},
|
135 |
-
},
|
136 |
-
description=textwrap.dedent(
|
137 |
-
"""\
|
138 |
-
STS is a task which aims to predict the semantic similarity of two input sentences as
|
139 |
-
a real value between 0 and 5. Note that we furthure binarized the prediction scores
|
140 |
-
into two classes with a threshold score 3.0 (paraphrased or not) and evaluated with
|
141 |
-
a classification metric.
|
142 |
-
"""
|
143 |
-
),
|
144 |
-
data_url=_DATA_URLs["sts"],
|
145 |
-
url=_DESCRIPTION_URLs["sts"],
|
146 |
-
file_map={
|
147 |
-
"train": "klue-sts-v1_train.json",
|
148 |
-
"dev": "klue-sts-v1_dev.json",
|
149 |
-
},
|
150 |
-
),
|
151 |
-
KlueConfig(
|
152 |
-
name="nli",
|
153 |
-
features={
|
154 |
-
"guid": datasets.Value("string"),
|
155 |
-
"source": datasets.Value("string"),
|
156 |
-
"premise": datasets.Value("string"),
|
157 |
-
"hypothesis": datasets.Value("string"),
|
158 |
-
"label": datasets.ClassLabel(names=["entailment", "neutral", "contradiction"]),
|
159 |
-
},
|
160 |
-
description=textwrap.dedent(
|
161 |
-
"""\
|
162 |
-
NLI is a task to infer the relationship between a hypothesis sentence and a premise
|
163 |
-
sentence. Given the premise, the model determines if the hypothesis is true (entailment),
|
164 |
-
false (contradiction), or undetermined (neutral).
|
165 |
-
"""
|
166 |
-
),
|
167 |
-
data_url=_DATA_URLs["nli"],
|
168 |
-
url=_DESCRIPTION_URLs["nli"],
|
169 |
-
file_map={
|
170 |
-
"train": "klue-nli-v1_train.json",
|
171 |
-
"dev": "klue-nli-v1_dev.json",
|
172 |
-
},
|
173 |
-
),
|
174 |
-
KlueConfig(
|
175 |
-
name="ner",
|
176 |
-
features={
|
177 |
-
"sentence": datasets.Value("string"),
|
178 |
-
"tokens": datasets.Sequence(datasets.Value("string")),
|
179 |
-
"ner_tags": datasets.Sequence(
|
180 |
-
datasets.ClassLabel(
|
181 |
-
names=[
|
182 |
-
"B-DT",
|
183 |
-
"I-DT",
|
184 |
-
"B-LC",
|
185 |
-
"I-LC",
|
186 |
-
"B-OG",
|
187 |
-
"I-OG",
|
188 |
-
"B-PS",
|
189 |
-
"I-PS",
|
190 |
-
"B-QT",
|
191 |
-
"I-QT",
|
192 |
-
"B-TI",
|
193 |
-
"I-TI",
|
194 |
-
"O",
|
195 |
-
]
|
196 |
-
)
|
197 |
-
),
|
198 |
-
},
|
199 |
-
description=textwrap.dedent(
|
200 |
-
"""\
|
201 |
-
NER is a task to detect the boundaries of named entities in unstructured text and to
|
202 |
-
classify the types. A named entity can be of one of predefined entity types such as
|
203 |
-
person, location, organization, time expressions, quantities and monetary values.
|
204 |
-
"""
|
205 |
-
),
|
206 |
-
data_url=_DATA_URLs["ner"],
|
207 |
-
url=_DESCRIPTION_URLs["ner"],
|
208 |
-
file_map={
|
209 |
-
"train": "klue-ner-v1_train.tsv",
|
210 |
-
"dev": "klue-ner-v1_dev.tsv",
|
211 |
-
},
|
212 |
-
),
|
213 |
-
KlueConfig(
|
214 |
-
name="re",
|
215 |
-
features={
|
216 |
-
"guid": datasets.Value("string"),
|
217 |
-
"sentence": datasets.Value("string"),
|
218 |
-
"subject_entity": {
|
219 |
-
"word": datasets.Value("string"),
|
220 |
-
"start_idx": datasets.Value("int32"),
|
221 |
-
"end_idx": datasets.Value("int32"),
|
222 |
-
"type": datasets.Value("string"),
|
223 |
-
},
|
224 |
-
"object_entity": {
|
225 |
-
"word": datasets.Value("string"),
|
226 |
-
"start_idx": datasets.Value("int32"),
|
227 |
-
"end_idx": datasets.Value("int32"),
|
228 |
-
"type": datasets.Value("string"),
|
229 |
-
},
|
230 |
-
"label": datasets.ClassLabel(
|
231 |
-
names=[
|
232 |
-
"no_relation",
|
233 |
-
"org:dissolved",
|
234 |
-
"org:founded",
|
235 |
-
"org:place_of_headquarters",
|
236 |
-
"org:alternate_names",
|
237 |
-
"org:member_of",
|
238 |
-
"org:members",
|
239 |
-
"org:political/religious_affiliation",
|
240 |
-
"org:product",
|
241 |
-
"org:founded_by",
|
242 |
-
"org:top_members/employees",
|
243 |
-
"org:number_of_employees/members",
|
244 |
-
"per:date_of_birth",
|
245 |
-
"per:date_of_death",
|
246 |
-
"per:place_of_birth",
|
247 |
-
"per:place_of_death",
|
248 |
-
"per:place_of_residence",
|
249 |
-
"per:origin",
|
250 |
-
"per:employee_of",
|
251 |
-
"per:schools_attended",
|
252 |
-
"per:alternate_names",
|
253 |
-
"per:parents",
|
254 |
-
"per:children",
|
255 |
-
"per:siblings",
|
256 |
-
"per:spouse",
|
257 |
-
"per:other_family",
|
258 |
-
"per:colleagues",
|
259 |
-
"per:product",
|
260 |
-
"per:religion",
|
261 |
-
"per:title",
|
262 |
-
]
|
263 |
-
),
|
264 |
-
"source": datasets.Value("string"),
|
265 |
-
},
|
266 |
-
description=textwrap.dedent(
|
267 |
-
"""\
|
268 |
-
RE is a task to identify semantic relations between entity pairs in a text. The relation
|
269 |
-
is defined between an entity pair consisting of subject entity and object entity.
|
270 |
-
The goal is then to pick an appropriate relationship between these two entities.
|
271 |
-
"""
|
272 |
-
),
|
273 |
-
data_url=_DATA_URLs["re"],
|
274 |
-
url=_DESCRIPTION_URLs["re"],
|
275 |
-
file_map={
|
276 |
-
"train": "klue-re-v1_train.json",
|
277 |
-
"dev": "klue-re-v1_dev.json",
|
278 |
-
},
|
279 |
-
),
|
280 |
-
KlueConfig(
|
281 |
-
name="dp",
|
282 |
-
features={
|
283 |
-
"sentence": datasets.Value("string"),
|
284 |
-
"index": [datasets.Value("int32")],
|
285 |
-
"word_form": [datasets.Value("string")],
|
286 |
-
"lemma": [datasets.Value("string")],
|
287 |
-
"pos": [datasets.Value("string")],
|
288 |
-
"head": [datasets.Value("int32")],
|
289 |
-
"deprel": [datasets.Value("string")],
|
290 |
-
},
|
291 |
-
description=textwrap.dedent(
|
292 |
-
"""\
|
293 |
-
DP is a task that aims at finding relational information among words.
|
294 |
-
The goal is to predict a graph structure and a dependency label of an input sentence
|
295 |
-
based on the dependency grammar.
|
296 |
-
"""
|
297 |
-
),
|
298 |
-
data_url=_DATA_URLs["dp"],
|
299 |
-
url=_DESCRIPTION_URLs["dp"],
|
300 |
-
file_map={
|
301 |
-
"train": "klue-dp-v1_train.tsv",
|
302 |
-
"dev": "klue-dp-v1_dev.tsv",
|
303 |
-
},
|
304 |
-
),
|
305 |
-
KlueConfig(
|
306 |
-
name="mrc",
|
307 |
-
features={
|
308 |
-
"title": datasets.Value("string"),
|
309 |
-
"context": datasets.Value("string"),
|
310 |
-
"news_category": datasets.Value("string"),
|
311 |
-
"source": datasets.Value("string"),
|
312 |
-
"guid": datasets.Value("string"),
|
313 |
-
"is_impossible": datasets.Value("bool"),
|
314 |
-
"question_type": datasets.Value("int32"),
|
315 |
-
"question": datasets.Value("string"),
|
316 |
-
"answers": datasets.features.Sequence(
|
317 |
-
{
|
318 |
-
"answer_start": datasets.Value("int32"),
|
319 |
-
"text": datasets.Value("string"),
|
320 |
-
},
|
321 |
-
),
|
322 |
-
},
|
323 |
-
description=textwrap.dedent(
|
324 |
-
"""\
|
325 |
-
MRC is a task of evaluating model that can answer a question about a given text
|
326 |
-
passage. Specifically, we formulate the task as a span prediction task, where the
|
327 |
-
answer is a text segment (coined as spans) in the passage.
|
328 |
-
"""
|
329 |
-
),
|
330 |
-
data_url=_DATA_URLs["mrc"],
|
331 |
-
url=_DESCRIPTION_URLs["mrc"],
|
332 |
-
file_map={
|
333 |
-
"train": "klue-mrc-v1_train.json",
|
334 |
-
"dev": "klue-mrc-v1_dev.json",
|
335 |
-
},
|
336 |
-
),
|
337 |
-
KlueConfig(
|
338 |
-
name="wos",
|
339 |
-
features={
|
340 |
-
"guid": datasets.Value("string"),
|
341 |
-
"domains": [datasets.Value("string")],
|
342 |
-
"dialogue": [
|
343 |
-
{
|
344 |
-
"role": datasets.Value("string"),
|
345 |
-
"text": datasets.Value("string"),
|
346 |
-
"state": [datasets.Value("string")],
|
347 |
-
}
|
348 |
-
],
|
349 |
-
},
|
350 |
-
description=textwrap.dedent(
|
351 |
-
"""\
|
352 |
-
DST is a task to predict slot and value pairs (dialogue states) from a task-oriented
|
353 |
-
dialogue. The potential pairs are predefined by a given task schema and knowledge
|
354 |
-
base (KB).
|
355 |
-
"""
|
356 |
-
),
|
357 |
-
data_url=_DATA_URLs["wos"],
|
358 |
-
url=_DESCRIPTION_URLs["wos"],
|
359 |
-
file_map={
|
360 |
-
"train": "wos-v1_train.json",
|
361 |
-
"dev": "wos-v1_dev.json",
|
362 |
-
},
|
363 |
-
),
|
364 |
-
]
|
365 |
-
|
366 |
-
def _info(self):
|
367 |
-
return datasets.DatasetInfo(
|
368 |
-
description=_KLUE_DESCRIPTION,
|
369 |
-
features=datasets.Features(self.config.features),
|
370 |
-
homepage=self.config.url,
|
371 |
-
citation=_KLUE_CITATION,
|
372 |
-
license=_LICENSE,
|
373 |
-
)
|
374 |
-
|
375 |
-
def _split_generators(self, dl_manager):
|
376 |
-
archive = dl_manager.download(self.config.data_url)
|
377 |
-
dir_name = self.config.data_url.split("/")[-1].replace(".tar.gz", "")
|
378 |
-
return [
|
379 |
-
datasets.SplitGenerator(
|
380 |
-
name=datasets.Split.TRAIN,
|
381 |
-
gen_kwargs={
|
382 |
-
"data_file": dir_name + "/" + self.config.file_map["train"],
|
383 |
-
"files": dl_manager.iter_archive(archive),
|
384 |
-
},
|
385 |
-
),
|
386 |
-
datasets.SplitGenerator(
|
387 |
-
name=datasets.Split.VALIDATION,
|
388 |
-
gen_kwargs={
|
389 |
-
"data_file": dir_name + "/" + self.config.file_map["dev"],
|
390 |
-
"files": dl_manager.iter_archive(archive),
|
391 |
-
},
|
392 |
-
),
|
393 |
-
]
|
394 |
-
|
395 |
-
def _generate_examples(self, data_file, files):
|
396 |
-
if self.config.name in ["ynat", "sts", "re"]:
|
397 |
-
for path, f in files:
|
398 |
-
if path == data_file:
|
399 |
-
f = json.load(f)
|
400 |
-
for id_, row in enumerate(f):
|
401 |
-
features = {key: row[key] for key in row if key in self.config.features}
|
402 |
-
yield id_, features
|
403 |
-
break
|
404 |
-
|
405 |
-
if self.config.name == "nli":
|
406 |
-
for path, f in files:
|
407 |
-
if path == data_file:
|
408 |
-
f = json.load(f)
|
409 |
-
for id_, row in enumerate(f):
|
410 |
-
# In train file, "source" is written as "genre"
|
411 |
-
features = {
|
412 |
-
"guid": row["guid"],
|
413 |
-
"source": row["source"] if "source" in row else row["genre"],
|
414 |
-
"premise": row["premise"],
|
415 |
-
"hypothesis": row["hypothesis"],
|
416 |
-
"label": row["gold_label"],
|
417 |
-
}
|
418 |
-
yield id_, features
|
419 |
-
break
|
420 |
-
|
421 |
-
if self.config.name == "ner":
|
422 |
-
for path, f in files:
|
423 |
-
if path == data_file:
|
424 |
-
f = (line.decode("utf-8") for line in f)
|
425 |
-
reader = csv.reader(f, delimiter="\t", quoting=csv.QUOTE_NONE)
|
426 |
-
for _ in range(5): # skip headers
|
427 |
-
next(reader)
|
428 |
-
id_ = -1
|
429 |
-
for row in reader:
|
430 |
-
if row:
|
431 |
-
if row[0].startswith("##"):
|
432 |
-
id_ += 1
|
433 |
-
tokens, ner_tags = [], []
|
434 |
-
sentence = row[1]
|
435 |
-
else:
|
436 |
-
tokens.append(row[0])
|
437 |
-
ner_tags.append(row[1])
|
438 |
-
else: # new line
|
439 |
-
assert len(tokens) == len(ner_tags)
|
440 |
-
yield id_, {"sentence": sentence, "tokens": tokens, "ner_tags": ner_tags}
|
441 |
-
break
|
442 |
-
|
443 |
-
if self.config.name == "dp":
|
444 |
-
for path, f in files:
|
445 |
-
if path == data_file:
|
446 |
-
f = (line.decode("utf-8") for line in f)
|
447 |
-
reader = csv.reader(f, delimiter="\t", quoting=csv.QUOTE_NONE)
|
448 |
-
for _ in range(5): # skip headers
|
449 |
-
next(reader)
|
450 |
-
id_ = -1
|
451 |
-
for row in reader:
|
452 |
-
if row:
|
453 |
-
if row[0].startswith("##"):
|
454 |
-
id_ += 1
|
455 |
-
index = []
|
456 |
-
word_form = []
|
457 |
-
lemma = []
|
458 |
-
pos = []
|
459 |
-
head = []
|
460 |
-
deprel = []
|
461 |
-
sentence = row[1]
|
462 |
-
else:
|
463 |
-
index.append(row[0])
|
464 |
-
word_form.append(row[1])
|
465 |
-
lemma.append(row[2])
|
466 |
-
pos.append(row[3])
|
467 |
-
head.append(row[4])
|
468 |
-
deprel.append(row[5])
|
469 |
-
else: # new line
|
470 |
-
assert len(index) == len(word_form) == len(lemma) == len(pos) == len(head) == len(deprel)
|
471 |
-
yield id_, {
|
472 |
-
"sentence": sentence,
|
473 |
-
"index": index,
|
474 |
-
"word_form": word_form,
|
475 |
-
"lemma": lemma,
|
476 |
-
"pos": pos,
|
477 |
-
"head": head,
|
478 |
-
"deprel": deprel,
|
479 |
-
}
|
480 |
-
break
|
481 |
-
|
482 |
-
if self.config.name == "mrc":
|
483 |
-
for path, f in files:
|
484 |
-
if path == data_file:
|
485 |
-
f = json.load(f)
|
486 |
-
id_ = -1
|
487 |
-
for example in f["data"]:
|
488 |
-
title = example.get("title", "")
|
489 |
-
news_category = example.get("news_category", "")
|
490 |
-
source = example["source"]
|
491 |
-
for paragraph in example["paragraphs"]:
|
492 |
-
context = paragraph["context"].strip()
|
493 |
-
for qa in paragraph["qas"]:
|
494 |
-
guid = qa["guid"]
|
495 |
-
question_type = qa["question_type"]
|
496 |
-
is_impossible = qa["is_impossible"]
|
497 |
-
question = qa["question"].strip()
|
498 |
-
|
499 |
-
if "plausible_answers" in qa:
|
500 |
-
qa["answers"].extend(qa["plausible_answers"])
|
501 |
-
answer_starts = [answer["answer_start"] for answer in qa["answers"]]
|
502 |
-
answers = [answer["text"].strip() for answer in qa["answers"]]
|
503 |
-
id_ += 1
|
504 |
-
|
505 |
-
yield id_, {
|
506 |
-
"guid": guid,
|
507 |
-
"title": title,
|
508 |
-
"context": context,
|
509 |
-
"news_category": news_category,
|
510 |
-
"source": source,
|
511 |
-
"question_type": question_type,
|
512 |
-
"is_impossible": is_impossible,
|
513 |
-
"question": question,
|
514 |
-
"answers": {
|
515 |
-
"answer_start": answer_starts,
|
516 |
-
"text": answers,
|
517 |
-
},
|
518 |
-
}
|
519 |
-
break
|
520 |
-
|
521 |
-
if self.config.name == "wos":
|
522 |
-
for path, f in files:
|
523 |
-
if path == data_file:
|
524 |
-
f = json.load(f)
|
525 |
-
for id_, row in enumerate(f):
|
526 |
-
guid = row["guid"]
|
527 |
-
domains = row["domains"]
|
528 |
-
dialogue = row["dialogue"]
|
529 |
-
for utterance in dialogue:
|
530 |
-
if "state" not in utterance:
|
531 |
-
utterance["state"] = []
|
532 |
-
yield id_, {"guid": guid, "domains": domains, "dialogue": dialogue}
|
533 |
-
break
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
mrc/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:45fcfd50b99aed4cd55c8c5856e42aa4a5139808cc36b36994b3c1991aca93dd
|
3 |
+
size 21415334
|
mrc/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f576c924f471702a6fa3369d314bea2e47545c56736f371a2adb90ed15913738
|
3 |
+
size 8683138
|
ner/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:291ded874432fa413e5007ec3770b41bb2e76f44a4e42cc7cc22cb4ebd8c27b6
|
3 |
+
size 4209983
|
ner/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:db2129c2ae38fc9a7de630d902cd41f80ac6174f620a8318b4d80569e31b9f84
|
3 |
+
size 1055904
|
nli/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35a8eb31818b94d9cfcb3477914360ce038b46b3bce4075eaef92309caa2c07c
|
3 |
+
size 1832150
|
nli/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b809f8f0aac2a2f41cd095bf15d8dbca9c931d7539390b6ef6241ca976195423
|
3 |
+
size 223966
|
re/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:329e15f482c794900d53627730448945607f2edb12397811eb6f671bf44d57da
|
3 |
+
size 6647272
|
re/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10aade1fdb0244eb867eb8a7e3459d391546016c792989eb1030e23603a28531
|
3 |
+
size 1542985
|
sts/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fa624db134d4242f24e232432055f09f34652c3852eb77ff98ca4784edfed184
|
3 |
+
size 1519075
|
sts/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:67611cf65f0d35feb244af81a1609186c55cc3a4aee293d1ceb02fb0d6b721ea
|
3 |
+
size 68780
|
wos/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:36003fc8ef605306180591e708ace9043b30ec8fbc0c657ed1f01d53cd4a8a01
|
3 |
+
size 5643747
|
wos/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f5342b51a1565aa9719adcb02125c8719c6ba8d3dda5be5c948e642db451fbdf
|
3 |
+
size 715108
|
ynat/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:062c3b51c1ca34ed23c8fd19ffa5ea0ddcd914a95aeb9077d89576d3ef71d123
|
3 |
+
size 4165783
|
ynat/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:521a6785a2492a9142474ad23ae038b1a6f333fa7e39aab8fd9dd19a20c9ce92
|
3 |
+
size 846520
|