albertvillanova HF staff commited on
Commit
b7f730b
1 Parent(s): aa733cd

Add sts data files

Browse files
README.md CHANGED
@@ -252,13 +252,13 @@ dataset_info:
252
  '1': positive
253
  splits:
254
  - name: train
255
- num_bytes: 2832921
256
  num_examples: 11668
257
  - name: validation
258
- num_bytes: 122657
259
  num_examples: 519
260
- download_size: 1349875
261
- dataset_size: 2955578
262
  - config_name: wos
263
  features:
264
  - name: guid
@@ -313,6 +313,12 @@ dataset_info:
313
  download_size: 5012303
314
  dataset_size: 12148765
315
  configs:
 
 
 
 
 
 
316
  - config_name: ynat
317
  data_files:
318
  - split: train
 
252
  '1': positive
253
  splits:
254
  - name: train
255
+ num_bytes: 2832889
256
  num_examples: 11668
257
  - name: validation
258
+ num_bytes: 122641
259
  num_examples: 519
260
+ download_size: 1587855
261
+ dataset_size: 2955530
262
  - config_name: wos
263
  features:
264
  - name: guid
 
313
  download_size: 5012303
314
  dataset_size: 12148765
315
  configs:
316
+ - config_name: sts
317
+ data_files:
318
+ - split: train
319
+ path: sts/train-*
320
+ - split: validation
321
+ path: sts/validation-*
322
  - config_name: ynat
323
  data_files:
324
  - split: train
dataset_infos.json CHANGED
@@ -70,51 +70,40 @@
70
  "features": {
71
  "guid": {
72
  "dtype": "string",
73
- "id": null,
74
  "_type": "Value"
75
  },
76
  "source": {
77
  "dtype": "string",
78
- "id": null,
79
  "_type": "Value"
80
  },
81
  "sentence1": {
82
  "dtype": "string",
83
- "id": null,
84
  "_type": "Value"
85
  },
86
  "sentence2": {
87
  "dtype": "string",
88
- "id": null,
89
  "_type": "Value"
90
  },
91
  "labels": {
92
  "label": {
93
  "dtype": "float64",
94
- "id": null,
95
  "_type": "Value"
96
  },
97
  "real-label": {
98
  "dtype": "float64",
99
- "id": null,
100
  "_type": "Value"
101
  },
102
  "binary-label": {
103
- "num_classes": 2,
104
  "names": [
105
  "negative",
106
  "positive"
107
  ],
108
- "names_file": null,
109
- "id": null,
110
  "_type": "ClassLabel"
111
  }
112
  }
113
  },
114
- "post_processed": null,
115
- "supervised_keys": null,
116
- "task_templates": null,
117
  "builder_name": "klue",
 
118
  "config_name": "sts",
119
  "version": {
120
  "version_str": "1.0.0",
@@ -126,27 +115,20 @@
126
  "splits": {
127
  "train": {
128
  "name": "train",
129
- "num_bytes": 2832921,
130
  "num_examples": 11668,
131
- "dataset_name": "klue"
132
  },
133
  "validation": {
134
  "name": "validation",
135
- "num_bytes": 122657,
136
  "num_examples": 519,
137
- "dataset_name": "klue"
138
- }
139
- },
140
- "download_checksums": {
141
- "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000067/data/klue-sts-v1.tar.gz": {
142
- "num_bytes": 1349875,
143
- "checksum": "539341ba78a3b351c686cf70a448ac7a5886ed95f0719d5e3d2378ba703213bd"
144
  }
145
  },
146
- "download_size": 1349875,
147
- "post_processing_size": null,
148
- "dataset_size": 2955578,
149
- "size_in_bytes": 4305453
150
  },
151
  "nli": {
152
  "description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n",
 
70
  "features": {
71
  "guid": {
72
  "dtype": "string",
 
73
  "_type": "Value"
74
  },
75
  "source": {
76
  "dtype": "string",
 
77
  "_type": "Value"
78
  },
79
  "sentence1": {
80
  "dtype": "string",
 
81
  "_type": "Value"
82
  },
83
  "sentence2": {
84
  "dtype": "string",
 
85
  "_type": "Value"
86
  },
87
  "labels": {
88
  "label": {
89
  "dtype": "float64",
 
90
  "_type": "Value"
91
  },
92
  "real-label": {
93
  "dtype": "float64",
 
94
  "_type": "Value"
95
  },
96
  "binary-label": {
 
97
  "names": [
98
  "negative",
99
  "positive"
100
  ],
 
 
101
  "_type": "ClassLabel"
102
  }
103
  }
104
  },
 
 
 
105
  "builder_name": "klue",
106
+ "dataset_name": "klue",
107
  "config_name": "sts",
108
  "version": {
109
  "version_str": "1.0.0",
 
115
  "splits": {
116
  "train": {
117
  "name": "train",
118
+ "num_bytes": 2832889,
119
  "num_examples": 11668,
120
+ "dataset_name": null
121
  },
122
  "validation": {
123
  "name": "validation",
124
+ "num_bytes": 122641,
125
  "num_examples": 519,
126
+ "dataset_name": null
 
 
 
 
 
 
127
  }
128
  },
129
+ "download_size": 1587855,
130
+ "dataset_size": 2955530,
131
+ "size_in_bytes": 4543385
 
132
  },
133
  "nli": {
134
  "description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n",
sts/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa624db134d4242f24e232432055f09f34652c3852eb77ff98ca4784edfed184
3
+ size 1519075
sts/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67611cf65f0d35feb244af81a1609186c55cc3a4aee293d1ceb02fb0d6b721ea
3
+ size 68780