Commit
•
594279f
1
Parent(s):
bbd618a
Add results for 2023-10-19 08:44:42
Browse files
choco9966/Llama-2-7b-instruct-tuning/result_2023-10-19 08:44:42.json
CHANGED
@@ -1,22 +1,22 @@
|
|
1 |
{
|
2 |
"results": {
|
3 |
"harness|ko_arc_challenge|25": {
|
4 |
-
"acc": 0.
|
5 |
-
"acc_stderr": 0.
|
6 |
-
"acc_norm": 0.
|
7 |
-
"acc_norm_stderr": 0.
|
8 |
},
|
9 |
"harness|ko_hellaswag|10": {
|
10 |
-
"acc": 0.
|
11 |
-
"acc_stderr": 0.
|
12 |
-
"acc_norm": 0.
|
13 |
-
"acc_norm_stderr": 0.
|
14 |
},
|
15 |
"harness|ko_mmlu_world_religions|5": {
|
16 |
-
"acc": 0.
|
17 |
-
"acc_stderr": 0.
|
18 |
-
"acc_norm": 0.
|
19 |
-
"acc_norm_stderr": 0.
|
20 |
},
|
21 |
"harness|ko_mmlu_management|5": {
|
22 |
"acc": 0.34951456310679613,
|
@@ -49,10 +49,10 @@
|
|
49 |
"acc_norm_stderr": 0.03148955829745529
|
50 |
},
|
51 |
"harness|ko_mmlu_virology|5": {
|
52 |
-
"acc": 0.
|
53 |
-
"acc_stderr": 0.
|
54 |
-
"acc_norm": 0.
|
55 |
-
"acc_norm_stderr": 0.
|
56 |
},
|
57 |
"harness|ko_mmlu_philosophy|5": {
|
58 |
"acc": 0.3858520900321543,
|
@@ -73,10 +73,10 @@
|
|
73 |
"acc_norm_stderr": 0.04363643698524779
|
74 |
},
|
75 |
"harness|ko_mmlu_medical_genetics|5": {
|
76 |
-
"acc": 0.
|
77 |
-
"acc_stderr": 0.
|
78 |
-
"acc_norm": 0.
|
79 |
-
"acc_norm_stderr": 0.
|
80 |
},
|
81 |
"harness|ko_mmlu_high_school_geography|5": {
|
82 |
"acc": 0.35353535353535354,
|
@@ -97,10 +97,10 @@
|
|
97 |
"acc_norm_stderr": 0.04092563958237654
|
98 |
},
|
99 |
"harness|ko_mmlu_high_school_microeconomics|5": {
|
100 |
-
"acc": 0.
|
101 |
-
"acc_stderr": 0.
|
102 |
-
"acc_norm": 0.
|
103 |
-
"acc_norm_stderr": 0.
|
104 |
},
|
105 |
"harness|ko_mmlu_high_school_macroeconomics|5": {
|
106 |
"acc": 0.31025641025641026,
|
@@ -127,16 +127,16 @@
|
|
127 |
"acc_norm_stderr": 0.047128212574267705
|
128 |
},
|
129 |
"harness|ko_mmlu_high_school_chemistry|5": {
|
130 |
-
"acc": 0.
|
131 |
-
"acc_stderr": 0.
|
132 |
-
"acc_norm": 0.
|
133 |
-
"acc_norm_stderr": 0.
|
134 |
},
|
135 |
"harness|ko_mmlu_high_school_biology|5": {
|
136 |
-
"acc": 0.
|
137 |
-
"acc_stderr": 0.
|
138 |
-
"acc_norm": 0.
|
139 |
-
"acc_norm_stderr": 0.
|
140 |
},
|
141 |
"harness|ko_mmlu_marketing|5": {
|
142 |
"acc": 0.5128205128205128,
|
@@ -145,16 +145,16 @@
|
|
145 |
"acc_norm_stderr": 0.032745319388423504
|
146 |
},
|
147 |
"harness|ko_mmlu_clinical_knowledge|5": {
|
148 |
-
"acc": 0.
|
149 |
-
"acc_stderr": 0.
|
150 |
-
"acc_norm": 0.
|
151 |
-
"acc_norm_stderr": 0.
|
152 |
},
|
153 |
"harness|ko_mmlu_public_relations|5": {
|
154 |
-
"acc": 0.
|
155 |
-
"acc_stderr": 0.
|
156 |
-
"acc_norm": 0.
|
157 |
-
"acc_norm_stderr": 0.
|
158 |
},
|
159 |
"harness|ko_mmlu_high_school_mathematics|5": {
|
160 |
"acc": 0.29259259259259257,
|
@@ -163,10 +163,10 @@
|
|
163 |
"acc_norm_stderr": 0.027738969632176095
|
164 |
},
|
165 |
"harness|ko_mmlu_high_school_physics|5": {
|
166 |
-
"acc": 0.
|
167 |
-
"acc_stderr": 0.
|
168 |
-
"acc_norm": 0.
|
169 |
-
"acc_norm_stderr": 0.
|
170 |
},
|
171 |
"harness|ko_mmlu_sociology|5": {
|
172 |
"acc": 0.48258706467661694,
|
@@ -175,10 +175,10 @@
|
|
175 |
"acc_norm_stderr": 0.03533389234739245
|
176 |
},
|
177 |
"harness|ko_mmlu_college_medicine|5": {
|
178 |
-
"acc": 0.
|
179 |
-
"acc_stderr": 0.
|
180 |
-
"acc_norm": 0.
|
181 |
-
"acc_norm_stderr": 0.
|
182 |
},
|
183 |
"harness|ko_mmlu_elementary_mathematics|5": {
|
184 |
"acc": 0.30158730158730157,
|
@@ -217,22 +217,22 @@
|
|
217 |
"acc_norm_stderr": 0.03731133519673893
|
218 |
},
|
219 |
"harness|ko_mmlu_prehistory|5": {
|
220 |
-
"acc": 0.
|
221 |
-
"acc_stderr": 0.
|
222 |
-
"acc_norm": 0.
|
223 |
-
"acc_norm_stderr": 0.
|
224 |
},
|
225 |
"harness|ko_mmlu_college_mathematics|5": {
|
226 |
"acc": 0.28,
|
227 |
-
"acc_stderr": 0.
|
228 |
"acc_norm": 0.28,
|
229 |
-
"acc_norm_stderr": 0.
|
230 |
},
|
231 |
"harness|ko_mmlu_high_school_government_and_politics|5": {
|
232 |
-
"acc": 0.
|
233 |
-
"acc_stderr": 0.
|
234 |
-
"acc_norm": 0.
|
235 |
-
"acc_norm_stderr": 0.
|
236 |
},
|
237 |
"harness|ko_mmlu_econometrics|5": {
|
238 |
"acc": 0.2719298245614035,
|
@@ -241,10 +241,10 @@
|
|
241 |
"acc_norm_stderr": 0.041857744240220575
|
242 |
},
|
243 |
"harness|ko_mmlu_high_school_psychology|5": {
|
244 |
-
"acc": 0.
|
245 |
-
"acc_stderr": 0.
|
246 |
-
"acc_norm": 0.
|
247 |
-
"acc_norm_stderr": 0.
|
248 |
},
|
249 |
"harness|ko_mmlu_formal_logic|5": {
|
250 |
"acc": 0.30158730158730157,
|
@@ -253,16 +253,16 @@
|
|
253 |
"acc_norm_stderr": 0.04104947269903394
|
254 |
},
|
255 |
"harness|ko_mmlu_nutrition|5": {
|
256 |
-
"acc": 0.
|
257 |
-
"acc_stderr": 0.
|
258 |
-
"acc_norm": 0.
|
259 |
-
"acc_norm_stderr": 0.
|
260 |
},
|
261 |
"harness|ko_mmlu_business_ethics|5": {
|
262 |
-
"acc": 0.
|
263 |
-
"acc_stderr": 0.
|
264 |
-
"acc_norm": 0.
|
265 |
-
"acc_norm_stderr": 0.
|
266 |
},
|
267 |
"harness|ko_mmlu_international_law|5": {
|
268 |
"acc": 0.49586776859504134,
|
@@ -313,16 +313,16 @@
|
|
313 |
"acc_norm_stderr": 0.04760952285695235
|
314 |
},
|
315 |
"harness|ko_mmlu_high_school_computer_science|5": {
|
316 |
-
"acc": 0.
|
317 |
-
"acc_stderr": 0.
|
318 |
-
"acc_norm": 0.
|
319 |
-
"acc_norm_stderr": 0.
|
320 |
},
|
321 |
"harness|ko_mmlu_professional_medicine|5": {
|
322 |
-
"acc": 0.
|
323 |
-
"acc_stderr": 0.
|
324 |
-
"acc_norm": 0.
|
325 |
-
"acc_norm_stderr": 0.
|
326 |
},
|
327 |
"harness|ko_mmlu_security_studies|5": {
|
328 |
"acc": 0.47346938775510206,
|
@@ -331,16 +331,16 @@
|
|
331 |
"acc_norm_stderr": 0.03196412734523272
|
332 |
},
|
333 |
"harness|ko_mmlu_high_school_world_history|5": {
|
334 |
-
"acc": 0.
|
335 |
-
"acc_stderr": 0.
|
336 |
-
"acc_norm": 0.
|
337 |
-
"acc_norm_stderr": 0.
|
338 |
},
|
339 |
"harness|ko_mmlu_professional_law|5": {
|
340 |
-
"acc": 0.
|
341 |
-
"acc_stderr": 0.
|
342 |
-
"acc_norm": 0.
|
343 |
-
"acc_norm_stderr": 0.
|
344 |
},
|
345 |
"harness|ko_mmlu_high_school_us_history|5": {
|
346 |
"acc": 0.31862745098039214,
|
@@ -357,12 +357,12 @@
|
|
357 |
"harness|ko_truthfulqa_mc|0": {
|
358 |
"mc1": 0.3317013463892289,
|
359 |
"mc1_stderr": 0.01648214881024146,
|
360 |
-
"mc2": 0.
|
361 |
-
"mc2_stderr": 0.
|
362 |
},
|
363 |
"harness|ko_commongen_v2|2": {
|
364 |
-
"acc": 0.
|
365 |
-
"acc_stderr": 0.
|
366 |
"acc_norm": 0.29161747343565525,
|
367 |
"acc_norm_stderr": 0.01562627669007024
|
368 |
}
|
|
|
1 |
{
|
2 |
"results": {
|
3 |
"harness|ko_arc_challenge|25": {
|
4 |
+
"acc": 0.2960750853242321,
|
5 |
+
"acc_stderr": 0.013340916085246268,
|
6 |
+
"acc_norm": 0.33276450511945393,
|
7 |
+
"acc_norm_stderr": 0.013769863046192304
|
8 |
},
|
9 |
"harness|ko_hellaswag|10": {
|
10 |
+
"acc": 0.3458474407488548,
|
11 |
+
"acc_stderr": 0.004746716805735752,
|
12 |
+
"acc_norm": 0.42471619199362676,
|
13 |
+
"acc_norm_stderr": 0.00493289647246057
|
14 |
},
|
15 |
"harness|ko_mmlu_world_religions|5": {
|
16 |
+
"acc": 0.4327485380116959,
|
17 |
+
"acc_stderr": 0.03799978644370608,
|
18 |
+
"acc_norm": 0.4327485380116959,
|
19 |
+
"acc_norm_stderr": 0.03799978644370608
|
20 |
},
|
21 |
"harness|ko_mmlu_management|5": {
|
22 |
"acc": 0.34951456310679613,
|
|
|
49 |
"acc_norm_stderr": 0.03148955829745529
|
50 |
},
|
51 |
"harness|ko_mmlu_virology|5": {
|
52 |
+
"acc": 0.2891566265060241,
|
53 |
+
"acc_stderr": 0.03529486801511114,
|
54 |
+
"acc_norm": 0.2891566265060241,
|
55 |
+
"acc_norm_stderr": 0.03529486801511114
|
56 |
},
|
57 |
"harness|ko_mmlu_philosophy|5": {
|
58 |
"acc": 0.3858520900321543,
|
|
|
73 |
"acc_norm_stderr": 0.04363643698524779
|
74 |
},
|
75 |
"harness|ko_mmlu_medical_genetics|5": {
|
76 |
+
"acc": 0.36,
|
77 |
+
"acc_stderr": 0.04824181513244218,
|
78 |
+
"acc_norm": 0.36,
|
79 |
+
"acc_norm_stderr": 0.04824181513244218
|
80 |
},
|
81 |
"harness|ko_mmlu_high_school_geography|5": {
|
82 |
"acc": 0.35353535353535354,
|
|
|
97 |
"acc_norm_stderr": 0.04092563958237654
|
98 |
},
|
99 |
"harness|ko_mmlu_high_school_microeconomics|5": {
|
100 |
+
"acc": 0.3403361344537815,
|
101 |
+
"acc_stderr": 0.030778057422931673,
|
102 |
+
"acc_norm": 0.3403361344537815,
|
103 |
+
"acc_norm_stderr": 0.030778057422931673
|
104 |
},
|
105 |
"harness|ko_mmlu_high_school_macroeconomics|5": {
|
106 |
"acc": 0.31025641025641026,
|
|
|
127 |
"acc_norm_stderr": 0.047128212574267705
|
128 |
},
|
129 |
"harness|ko_mmlu_high_school_chemistry|5": {
|
130 |
+
"acc": 0.28078817733990147,
|
131 |
+
"acc_stderr": 0.03161856335358611,
|
132 |
+
"acc_norm": 0.28078817733990147,
|
133 |
+
"acc_norm_stderr": 0.03161856335358611
|
134 |
},
|
135 |
"harness|ko_mmlu_high_school_biology|5": {
|
136 |
+
"acc": 0.3548387096774194,
|
137 |
+
"acc_stderr": 0.02721888977330876,
|
138 |
+
"acc_norm": 0.3548387096774194,
|
139 |
+
"acc_norm_stderr": 0.02721888977330876
|
140 |
},
|
141 |
"harness|ko_mmlu_marketing|5": {
|
142 |
"acc": 0.5128205128205128,
|
|
|
145 |
"acc_norm_stderr": 0.032745319388423504
|
146 |
},
|
147 |
"harness|ko_mmlu_clinical_knowledge|5": {
|
148 |
+
"acc": 0.32075471698113206,
|
149 |
+
"acc_stderr": 0.028727502957880274,
|
150 |
+
"acc_norm": 0.32075471698113206,
|
151 |
+
"acc_norm_stderr": 0.028727502957880274
|
152 |
},
|
153 |
"harness|ko_mmlu_public_relations|5": {
|
154 |
+
"acc": 0.38181818181818183,
|
155 |
+
"acc_stderr": 0.04653429807913508,
|
156 |
+
"acc_norm": 0.38181818181818183,
|
157 |
+
"acc_norm_stderr": 0.04653429807913508
|
158 |
},
|
159 |
"harness|ko_mmlu_high_school_mathematics|5": {
|
160 |
"acc": 0.29259259259259257,
|
|
|
163 |
"acc_norm_stderr": 0.027738969632176095
|
164 |
},
|
165 |
"harness|ko_mmlu_high_school_physics|5": {
|
166 |
+
"acc": 0.2781456953642384,
|
167 |
+
"acc_stderr": 0.03658603262763743,
|
168 |
+
"acc_norm": 0.2781456953642384,
|
169 |
+
"acc_norm_stderr": 0.03658603262763743
|
170 |
},
|
171 |
"harness|ko_mmlu_sociology|5": {
|
172 |
"acc": 0.48258706467661694,
|
|
|
175 |
"acc_norm_stderr": 0.03533389234739245
|
176 |
},
|
177 |
"harness|ko_mmlu_college_medicine|5": {
|
178 |
+
"acc": 0.3063583815028902,
|
179 |
+
"acc_stderr": 0.035149425512674394,
|
180 |
+
"acc_norm": 0.3063583815028902,
|
181 |
+
"acc_norm_stderr": 0.035149425512674394
|
182 |
},
|
183 |
"harness|ko_mmlu_elementary_mathematics|5": {
|
184 |
"acc": 0.30158730158730157,
|
|
|
217 |
"acc_norm_stderr": 0.03731133519673893
|
218 |
},
|
219 |
"harness|ko_mmlu_prehistory|5": {
|
220 |
+
"acc": 0.38580246913580246,
|
221 |
+
"acc_stderr": 0.027085401226132143,
|
222 |
+
"acc_norm": 0.38580246913580246,
|
223 |
+
"acc_norm_stderr": 0.027085401226132143
|
224 |
},
|
225 |
"harness|ko_mmlu_college_mathematics|5": {
|
226 |
"acc": 0.28,
|
227 |
+
"acc_stderr": 0.04512608598542128,
|
228 |
"acc_norm": 0.28,
|
229 |
+
"acc_norm_stderr": 0.04512608598542128
|
230 |
},
|
231 |
"harness|ko_mmlu_high_school_government_and_politics|5": {
|
232 |
+
"acc": 0.43523316062176165,
|
233 |
+
"acc_stderr": 0.03578038165008586,
|
234 |
+
"acc_norm": 0.43523316062176165,
|
235 |
+
"acc_norm_stderr": 0.03578038165008586
|
236 |
},
|
237 |
"harness|ko_mmlu_econometrics|5": {
|
238 |
"acc": 0.2719298245614035,
|
|
|
241 |
"acc_norm_stderr": 0.041857744240220575
|
242 |
},
|
243 |
"harness|ko_mmlu_high_school_psychology|5": {
|
244 |
+
"acc": 0.3376146788990826,
|
245 |
+
"acc_stderr": 0.020275265986638903,
|
246 |
+
"acc_norm": 0.3376146788990826,
|
247 |
+
"acc_norm_stderr": 0.020275265986638903
|
248 |
},
|
249 |
"harness|ko_mmlu_formal_logic|5": {
|
250 |
"acc": 0.30158730158730157,
|
|
|
253 |
"acc_norm_stderr": 0.04104947269903394
|
254 |
},
|
255 |
"harness|ko_mmlu_nutrition|5": {
|
256 |
+
"acc": 0.3888888888888889,
|
257 |
+
"acc_stderr": 0.027914055510468008,
|
258 |
+
"acc_norm": 0.3888888888888889,
|
259 |
+
"acc_norm_stderr": 0.027914055510468008
|
260 |
},
|
261 |
"harness|ko_mmlu_business_ethics|5": {
|
262 |
+
"acc": 0.47,
|
263 |
+
"acc_stderr": 0.05016135580465919,
|
264 |
+
"acc_norm": 0.47,
|
265 |
+
"acc_norm_stderr": 0.05016135580465919
|
266 |
},
|
267 |
"harness|ko_mmlu_international_law|5": {
|
268 |
"acc": 0.49586776859504134,
|
|
|
313 |
"acc_norm_stderr": 0.04760952285695235
|
314 |
},
|
315 |
"harness|ko_mmlu_high_school_computer_science|5": {
|
316 |
+
"acc": 0.35,
|
317 |
+
"acc_stderr": 0.047937248544110196,
|
318 |
+
"acc_norm": 0.35,
|
319 |
+
"acc_norm_stderr": 0.047937248544110196
|
320 |
},
|
321 |
"harness|ko_mmlu_professional_medicine|5": {
|
322 |
+
"acc": 0.30514705882352944,
|
323 |
+
"acc_stderr": 0.0279715413701706,
|
324 |
+
"acc_norm": 0.30514705882352944,
|
325 |
+
"acc_norm_stderr": 0.0279715413701706
|
326 |
},
|
327 |
"harness|ko_mmlu_security_studies|5": {
|
328 |
"acc": 0.47346938775510206,
|
|
|
331 |
"acc_norm_stderr": 0.03196412734523272
|
332 |
},
|
333 |
"harness|ko_mmlu_high_school_world_history|5": {
|
334 |
+
"acc": 0.350210970464135,
|
335 |
+
"acc_stderr": 0.031052391937584356,
|
336 |
+
"acc_norm": 0.350210970464135,
|
337 |
+
"acc_norm_stderr": 0.031052391937584356
|
338 |
},
|
339 |
"harness|ko_mmlu_professional_law|5": {
|
340 |
+
"acc": 0.2633637548891786,
|
341 |
+
"acc_stderr": 0.011249506403605284,
|
342 |
+
"acc_norm": 0.2633637548891786,
|
343 |
+
"acc_norm_stderr": 0.011249506403605284
|
344 |
},
|
345 |
"harness|ko_mmlu_high_school_us_history|5": {
|
346 |
"acc": 0.31862745098039214,
|
|
|
357 |
"harness|ko_truthfulqa_mc|0": {
|
358 |
"mc1": 0.3317013463892289,
|
359 |
"mc1_stderr": 0.01648214881024146,
|
360 |
+
"mc2": 0.5140886375597211,
|
361 |
+
"mc2_stderr": 0.016082111072689104
|
362 |
},
|
363 |
"harness|ko_commongen_v2|2": {
|
364 |
+
"acc": 0.25737898465171194,
|
365 |
+
"acc_stderr": 0.015030899730346749,
|
366 |
"acc_norm": 0.29161747343565525,
|
367 |
"acc_norm_stderr": 0.01562627669007024
|
368 |
}
|