open-ko-llm-bot commited on
Commit
594279f
1 Parent(s): bbd618a

Add results for 2023-10-19 08:44:42

Browse files
choco9966/Llama-2-7b-instruct-tuning/result_2023-10-19 08:44:42.json CHANGED
@@ -1,22 +1,22 @@
1
  {
2
  "results": {
3
  "harness|ko_arc_challenge|25": {
4
- "acc": 0.29692832764505117,
5
- "acc_stderr": 0.013352025976725222,
6
- "acc_norm": 0.33447098976109213,
7
- "acc_norm_stderr": 0.01378746032244138
8
  },
9
  "harness|ko_hellaswag|10": {
10
- "acc": 0.3454491137223661,
11
- "acc_stderr": 0.004745426656377564,
12
- "acc_norm": 0.4245170284803824,
13
- "acc_norm_stderr": 0.00493259334881363
14
  },
15
  "harness|ko_mmlu_world_religions|5": {
16
- "acc": 0.4269005847953216,
17
- "acc_stderr": 0.03793620616529916,
18
- "acc_norm": 0.4269005847953216,
19
- "acc_norm_stderr": 0.03793620616529916
20
  },
21
  "harness|ko_mmlu_management|5": {
22
  "acc": 0.34951456310679613,
@@ -49,10 +49,10 @@
49
  "acc_norm_stderr": 0.03148955829745529
50
  },
51
  "harness|ko_mmlu_virology|5": {
52
- "acc": 0.29518072289156627,
53
- "acc_stderr": 0.03550920185689629,
54
- "acc_norm": 0.29518072289156627,
55
- "acc_norm_stderr": 0.03550920185689629
56
  },
57
  "harness|ko_mmlu_philosophy|5": {
58
  "acc": 0.3858520900321543,
@@ -73,10 +73,10 @@
73
  "acc_norm_stderr": 0.04363643698524779
74
  },
75
  "harness|ko_mmlu_medical_genetics|5": {
76
- "acc": 0.35,
77
- "acc_stderr": 0.04793724854411021,
78
- "acc_norm": 0.35,
79
- "acc_norm_stderr": 0.04793724854411021
80
  },
81
  "harness|ko_mmlu_high_school_geography|5": {
82
  "acc": 0.35353535353535354,
@@ -97,10 +97,10 @@
97
  "acc_norm_stderr": 0.04092563958237654
98
  },
99
  "harness|ko_mmlu_high_school_microeconomics|5": {
100
- "acc": 0.33613445378151263,
101
- "acc_stderr": 0.030684737115135377,
102
- "acc_norm": 0.33613445378151263,
103
- "acc_norm_stderr": 0.030684737115135377
104
  },
105
  "harness|ko_mmlu_high_school_macroeconomics|5": {
106
  "acc": 0.31025641025641026,
@@ -127,16 +127,16 @@
127
  "acc_norm_stderr": 0.047128212574267705
128
  },
129
  "harness|ko_mmlu_high_school_chemistry|5": {
130
- "acc": 0.27586206896551724,
131
- "acc_stderr": 0.0314471258167824,
132
- "acc_norm": 0.27586206896551724,
133
- "acc_norm_stderr": 0.0314471258167824
134
  },
135
  "harness|ko_mmlu_high_school_biology|5": {
136
- "acc": 0.3580645161290323,
137
- "acc_stderr": 0.027273890594300642,
138
- "acc_norm": 0.3580645161290323,
139
- "acc_norm_stderr": 0.027273890594300642
140
  },
141
  "harness|ko_mmlu_marketing|5": {
142
  "acc": 0.5128205128205128,
@@ -145,16 +145,16 @@
145
  "acc_norm_stderr": 0.032745319388423504
146
  },
147
  "harness|ko_mmlu_clinical_knowledge|5": {
148
- "acc": 0.32452830188679244,
149
- "acc_stderr": 0.028815615713432115,
150
- "acc_norm": 0.32452830188679244,
151
- "acc_norm_stderr": 0.028815615713432115
152
  },
153
  "harness|ko_mmlu_public_relations|5": {
154
- "acc": 0.39090909090909093,
155
- "acc_stderr": 0.04673752333670237,
156
- "acc_norm": 0.39090909090909093,
157
- "acc_norm_stderr": 0.04673752333670237
158
  },
159
  "harness|ko_mmlu_high_school_mathematics|5": {
160
  "acc": 0.29259259259259257,
@@ -163,10 +163,10 @@
163
  "acc_norm_stderr": 0.027738969632176095
164
  },
165
  "harness|ko_mmlu_high_school_physics|5": {
166
- "acc": 0.26490066225165565,
167
- "acc_stderr": 0.03603038545360384,
168
- "acc_norm": 0.26490066225165565,
169
- "acc_norm_stderr": 0.03603038545360384
170
  },
171
  "harness|ko_mmlu_sociology|5": {
172
  "acc": 0.48258706467661694,
@@ -175,10 +175,10 @@
175
  "acc_norm_stderr": 0.03533389234739245
176
  },
177
  "harness|ko_mmlu_college_medicine|5": {
178
- "acc": 0.31213872832369943,
179
- "acc_stderr": 0.035331333893236574,
180
- "acc_norm": 0.31213872832369943,
181
- "acc_norm_stderr": 0.035331333893236574
182
  },
183
  "harness|ko_mmlu_elementary_mathematics|5": {
184
  "acc": 0.30158730158730157,
@@ -217,22 +217,22 @@
217
  "acc_norm_stderr": 0.03731133519673893
218
  },
219
  "harness|ko_mmlu_prehistory|5": {
220
- "acc": 0.38271604938271603,
221
- "acc_stderr": 0.027044538138402616,
222
- "acc_norm": 0.38271604938271603,
223
- "acc_norm_stderr": 0.027044538138402616
224
  },
225
  "harness|ko_mmlu_college_mathematics|5": {
226
  "acc": 0.28,
227
- "acc_stderr": 0.04512608598542129,
228
  "acc_norm": 0.28,
229
- "acc_norm_stderr": 0.04512608598542129
230
  },
231
  "harness|ko_mmlu_high_school_government_and_politics|5": {
232
- "acc": 0.43005181347150256,
233
- "acc_stderr": 0.03572954333144808,
234
- "acc_norm": 0.43005181347150256,
235
- "acc_norm_stderr": 0.03572954333144808
236
  },
237
  "harness|ko_mmlu_econometrics|5": {
238
  "acc": 0.2719298245614035,
@@ -241,10 +241,10 @@
241
  "acc_norm_stderr": 0.041857744240220575
242
  },
243
  "harness|ko_mmlu_high_school_psychology|5": {
244
- "acc": 0.3412844036697248,
245
- "acc_stderr": 0.020328612816592432,
246
- "acc_norm": 0.3412844036697248,
247
- "acc_norm_stderr": 0.020328612816592432
248
  },
249
  "harness|ko_mmlu_formal_logic|5": {
250
  "acc": 0.30158730158730157,
@@ -253,16 +253,16 @@
253
  "acc_norm_stderr": 0.04104947269903394
254
  },
255
  "harness|ko_mmlu_nutrition|5": {
256
- "acc": 0.39215686274509803,
257
- "acc_stderr": 0.027956046165424516,
258
- "acc_norm": 0.39215686274509803,
259
- "acc_norm_stderr": 0.027956046165424516
260
  },
261
  "harness|ko_mmlu_business_ethics|5": {
262
- "acc": 0.46,
263
- "acc_stderr": 0.05009082659620332,
264
- "acc_norm": 0.46,
265
- "acc_norm_stderr": 0.05009082659620332
266
  },
267
  "harness|ko_mmlu_international_law|5": {
268
  "acc": 0.49586776859504134,
@@ -313,16 +313,16 @@
313
  "acc_norm_stderr": 0.04760952285695235
314
  },
315
  "harness|ko_mmlu_high_school_computer_science|5": {
316
- "acc": 0.34,
317
- "acc_stderr": 0.04760952285695236,
318
- "acc_norm": 0.34,
319
- "acc_norm_stderr": 0.04760952285695236
320
  },
321
  "harness|ko_mmlu_professional_medicine|5": {
322
- "acc": 0.3088235294117647,
323
- "acc_stderr": 0.028064998167040094,
324
- "acc_norm": 0.3088235294117647,
325
- "acc_norm_stderr": 0.028064998167040094
326
  },
327
  "harness|ko_mmlu_security_studies|5": {
328
  "acc": 0.47346938775510206,
@@ -331,16 +331,16 @@
331
  "acc_norm_stderr": 0.03196412734523272
332
  },
333
  "harness|ko_mmlu_high_school_world_history|5": {
334
- "acc": 0.35443037974683544,
335
- "acc_stderr": 0.031137304297185798,
336
- "acc_norm": 0.35443037974683544,
337
- "acc_norm_stderr": 0.031137304297185798
338
  },
339
  "harness|ko_mmlu_professional_law|5": {
340
- "acc": 0.2627118644067797,
341
- "acc_stderr": 0.011240545514995669,
342
- "acc_norm": 0.2627118644067797,
343
- "acc_norm_stderr": 0.011240545514995669
344
  },
345
  "harness|ko_mmlu_high_school_us_history|5": {
346
  "acc": 0.31862745098039214,
@@ -357,12 +357,12 @@
357
  "harness|ko_truthfulqa_mc|0": {
358
  "mc1": 0.3317013463892289,
359
  "mc1_stderr": 0.01648214881024146,
360
- "mc2": 0.5139810284158008,
361
- "mc2_stderr": 0.01608284635540924
362
  },
363
  "harness|ko_commongen_v2|2": {
364
- "acc": 0.2585596221959858,
365
- "acc_stderr": 0.015053354438964,
366
  "acc_norm": 0.29161747343565525,
367
  "acc_norm_stderr": 0.01562627669007024
368
  }
 
1
  {
2
  "results": {
3
  "harness|ko_arc_challenge|25": {
4
+ "acc": 0.2960750853242321,
5
+ "acc_stderr": 0.013340916085246268,
6
+ "acc_norm": 0.33276450511945393,
7
+ "acc_norm_stderr": 0.013769863046192304
8
  },
9
  "harness|ko_hellaswag|10": {
10
+ "acc": 0.3458474407488548,
11
+ "acc_stderr": 0.004746716805735752,
12
+ "acc_norm": 0.42471619199362676,
13
+ "acc_norm_stderr": 0.00493289647246057
14
  },
15
  "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4327485380116959,
17
+ "acc_stderr": 0.03799978644370608,
18
+ "acc_norm": 0.4327485380116959,
19
+ "acc_norm_stderr": 0.03799978644370608
20
  },
21
  "harness|ko_mmlu_management|5": {
22
  "acc": 0.34951456310679613,
 
49
  "acc_norm_stderr": 0.03148955829745529
50
  },
51
  "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.2891566265060241,
53
+ "acc_stderr": 0.03529486801511114,
54
+ "acc_norm": 0.2891566265060241,
55
+ "acc_norm_stderr": 0.03529486801511114
56
  },
57
  "harness|ko_mmlu_philosophy|5": {
58
  "acc": 0.3858520900321543,
 
73
  "acc_norm_stderr": 0.04363643698524779
74
  },
75
  "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.36,
77
+ "acc_stderr": 0.04824181513244218,
78
+ "acc_norm": 0.36,
79
+ "acc_norm_stderr": 0.04824181513244218
80
  },
81
  "harness|ko_mmlu_high_school_geography|5": {
82
  "acc": 0.35353535353535354,
 
97
  "acc_norm_stderr": 0.04092563958237654
98
  },
99
  "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.3403361344537815,
101
+ "acc_stderr": 0.030778057422931673,
102
+ "acc_norm": 0.3403361344537815,
103
+ "acc_norm_stderr": 0.030778057422931673
104
  },
105
  "harness|ko_mmlu_high_school_macroeconomics|5": {
106
  "acc": 0.31025641025641026,
 
127
  "acc_norm_stderr": 0.047128212574267705
128
  },
129
  "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.28078817733990147,
131
+ "acc_stderr": 0.03161856335358611,
132
+ "acc_norm": 0.28078817733990147,
133
+ "acc_norm_stderr": 0.03161856335358611
134
  },
135
  "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.3548387096774194,
137
+ "acc_stderr": 0.02721888977330876,
138
+ "acc_norm": 0.3548387096774194,
139
+ "acc_norm_stderr": 0.02721888977330876
140
  },
141
  "harness|ko_mmlu_marketing|5": {
142
  "acc": 0.5128205128205128,
 
145
  "acc_norm_stderr": 0.032745319388423504
146
  },
147
  "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.32075471698113206,
149
+ "acc_stderr": 0.028727502957880274,
150
+ "acc_norm": 0.32075471698113206,
151
+ "acc_norm_stderr": 0.028727502957880274
152
  },
153
  "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.38181818181818183,
155
+ "acc_stderr": 0.04653429807913508,
156
+ "acc_norm": 0.38181818181818183,
157
+ "acc_norm_stderr": 0.04653429807913508
158
  },
159
  "harness|ko_mmlu_high_school_mathematics|5": {
160
  "acc": 0.29259259259259257,
 
163
  "acc_norm_stderr": 0.027738969632176095
164
  },
165
  "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2781456953642384,
167
+ "acc_stderr": 0.03658603262763743,
168
+ "acc_norm": 0.2781456953642384,
169
+ "acc_norm_stderr": 0.03658603262763743
170
  },
171
  "harness|ko_mmlu_sociology|5": {
172
  "acc": 0.48258706467661694,
 
175
  "acc_norm_stderr": 0.03533389234739245
176
  },
177
  "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3063583815028902,
179
+ "acc_stderr": 0.035149425512674394,
180
+ "acc_norm": 0.3063583815028902,
181
+ "acc_norm_stderr": 0.035149425512674394
182
  },
183
  "harness|ko_mmlu_elementary_mathematics|5": {
184
  "acc": 0.30158730158730157,
 
217
  "acc_norm_stderr": 0.03731133519673893
218
  },
219
  "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.38580246913580246,
221
+ "acc_stderr": 0.027085401226132143,
222
+ "acc_norm": 0.38580246913580246,
223
+ "acc_norm_stderr": 0.027085401226132143
224
  },
225
  "harness|ko_mmlu_college_mathematics|5": {
226
  "acc": 0.28,
227
+ "acc_stderr": 0.04512608598542128,
228
  "acc_norm": 0.28,
229
+ "acc_norm_stderr": 0.04512608598542128
230
  },
231
  "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.43523316062176165,
233
+ "acc_stderr": 0.03578038165008586,
234
+ "acc_norm": 0.43523316062176165,
235
+ "acc_norm_stderr": 0.03578038165008586
236
  },
237
  "harness|ko_mmlu_econometrics|5": {
238
  "acc": 0.2719298245614035,
 
241
  "acc_norm_stderr": 0.041857744240220575
242
  },
243
  "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.3376146788990826,
245
+ "acc_stderr": 0.020275265986638903,
246
+ "acc_norm": 0.3376146788990826,
247
+ "acc_norm_stderr": 0.020275265986638903
248
  },
249
  "harness|ko_mmlu_formal_logic|5": {
250
  "acc": 0.30158730158730157,
 
253
  "acc_norm_stderr": 0.04104947269903394
254
  },
255
  "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.3888888888888889,
257
+ "acc_stderr": 0.027914055510468008,
258
+ "acc_norm": 0.3888888888888889,
259
+ "acc_norm_stderr": 0.027914055510468008
260
  },
261
  "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.47,
263
+ "acc_stderr": 0.05016135580465919,
264
+ "acc_norm": 0.47,
265
+ "acc_norm_stderr": 0.05016135580465919
266
  },
267
  "harness|ko_mmlu_international_law|5": {
268
  "acc": 0.49586776859504134,
 
313
  "acc_norm_stderr": 0.04760952285695235
314
  },
315
  "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.35,
317
+ "acc_stderr": 0.047937248544110196,
318
+ "acc_norm": 0.35,
319
+ "acc_norm_stderr": 0.047937248544110196
320
  },
321
  "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.30514705882352944,
323
+ "acc_stderr": 0.0279715413701706,
324
+ "acc_norm": 0.30514705882352944,
325
+ "acc_norm_stderr": 0.0279715413701706
326
  },
327
  "harness|ko_mmlu_security_studies|5": {
328
  "acc": 0.47346938775510206,
 
331
  "acc_norm_stderr": 0.03196412734523272
332
  },
333
  "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.350210970464135,
335
+ "acc_stderr": 0.031052391937584356,
336
+ "acc_norm": 0.350210970464135,
337
+ "acc_norm_stderr": 0.031052391937584356
338
  },
339
  "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.2633637548891786,
341
+ "acc_stderr": 0.011249506403605284,
342
+ "acc_norm": 0.2633637548891786,
343
+ "acc_norm_stderr": 0.011249506403605284
344
  },
345
  "harness|ko_mmlu_high_school_us_history|5": {
346
  "acc": 0.31862745098039214,
 
357
  "harness|ko_truthfulqa_mc|0": {
358
  "mc1": 0.3317013463892289,
359
  "mc1_stderr": 0.01648214881024146,
360
+ "mc2": 0.5140886375597211,
361
+ "mc2_stderr": 0.016082111072689104
362
  },
363
  "harness|ko_commongen_v2|2": {
364
+ "acc": 0.25737898465171194,
365
+ "acc_stderr": 0.015030899730346749,
366
  "acc_norm": 0.29161747343565525,
367
  "acc_norm_stderr": 0.01562627669007024
368
  }