Sean Cho commited on
Commit
ec0015a
1 Parent(s): 81e9f42
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. 01-ai/Yi-6B-Chat/result_2023-12-27 01:03:40.json +444 -0
  2. 01-ai/Yi-6B/result_2023-12-27 01:03:17.json +444 -0
  3. 42MARU/GenAI-llama-2-ko-en-instruct-v1/result_2023-10-12 11:14:34.json +444 -0
  4. 42MARU/GenAI-llama2-ko-en-dpo-13b-test3/result_2023-11-30 08:08:14.json +444 -0
  5. 42MARU/GenAI-llama2-ko-en-dpo-13b-v1/result_2023-11-18 17:03:07.json +444 -0
  6. 42MARU/GenAI-llama2-ko-en-dpo-13b-v2/result_2023-11-19 10:36:38.json +444 -0
  7. 42MARU/GenAI-llama2-ko-en-instruct-20B-v1/result_2023-11-05 16:07:53.json +444 -0
  8. 42MARU/GenAI-llama2-ko-en-instruct-v2-13b/result_2023-10-18 01:40:38.json +444 -0
  9. 42MARU/GenAI-llama2-ko-en-instruct-v3-13B/result_2023-11-01 18:54:40.json +444 -0
  10. 42MARU/GenAI-llama2-ko-en-instruct-v3-13B/result_2023-11-02 01:15:35.json +444 -0
  11. 42MARU/GenAI-llama2-ko-en-instruct-v4-13B/result_2023-11-09 09:34:14.json +444 -0
  12. 42MARU/GenAI-llama2-ko-en-instruct-v5-13B/result_2023-11-16 17:37:54.json +444 -0
  13. 42MARU/GenAI-llama2-ko-en-instruct-v6-13B/result_2023-11-27 02:12:26.json +444 -0
  14. 42MARU/GenAI-llama2-ko-en-instruct-v7-13B/result_2023-12-03 17:19:11.json +444 -0
  15. 42MARU/GenAI-llama2-ko-en-platypus-13B-v2/result_2023-10-30 01:04:14.json +444 -0
  16. 42MARU/GenAI-llama2-ko-en-platypus-13B/result_2023-10-29 00:31:00.json +444 -0
  17. 42MARU/llama-2-ko-7b-instruct/result_2023-09-29 09:41:36.json +444 -0
  18. 42MARU/llama-2-ko-7b-instruction-v3/result_2023-10-01 18:41:33.json +444 -0
  19. 42MARU/polyglot-ko-12.8b-instruct/result_2023-09-27 21:10:18.json +444 -0
  20. 42MARU/sitebunny-13b/result_2023-09-27 08:17:31.json +444 -0
  21. 42dot/42dot_LLM-PLM-1.3B/result_2023-10-18 01:46:47.json +444 -0
  22. 42dot/42dot_LLM-SFT-1.3B/result_2023-10-18 01:47:03.json +444 -0
  23. AIFT/PACK-13b-v1.0/result_2023-12-07 02:16:32.json +444 -0
  24. AIFT/PACK-13b-v1.1/result_2023-12-11 08:43:39.json +444 -0
  25. AIFT/aift-llama2-koen-instruct-dpo-v1.01/result_2023-12-14 06:44:33.json +444 -0
  26. AIFT/aift-llama2-koen-instruct-dpo-v1.02/result_2023-12-15 03:06:44.json +444 -0
  27. AIFT/aift-llama2-koen-instruct-v1.0/result_2023-12-14 00:45:21.json +444 -0
  28. AIFT/aift-llama2-koen-instruct-v1.1-dpo-test1/result_2023-12-18 03:42:48.json +444 -0
  29. AIFT/aift-llama2-koen-instruct-v1.1/result_2023-12-15 09:57:09.json +444 -0
  30. AIFT/aift-llama2-koen-instruct-v1.2/result_2023-12-16 03:29:50.json +444 -0
  31. AIdenU/LLAMA-2-13b-ko-Y24-DPO_v0.1/result_2023-12-18 01:34:17.json +444 -0
  32. AIdenU/Mistral-7b-ko-Y24-DPO_v0.1/result_2023-12-21 04:18:43.json +444 -0
  33. AIdenU/Mistral-7b-ko-Y24_v0.1/result_2023-12-21 04:19:12.json +444 -0
  34. AtAndDev/ShortKingv0.1/result_2023-09-29 19:59:47.json +444 -0
  35. BM-K/llama-2-ko-7b-it-v1.0.0/result_2023-11-15 11:34:52.json +444 -0
  36. BM-K/mistral-7b-it-v1.0.1/result_2023-11-07 22:19:25.json +444 -0
  37. BM-K/mistral-7b-it-v1.0.2/result_2023-11-08 13:07:51.json +444 -0
  38. BM-K/mistral-7b-it-v1.0.3/result_2023-11-09 00:04:55.json +444 -0
  39. BM-K/mistral-7b-it-v1.0/result_2023-11-07 08:10:27.json +444 -0
  40. BM-K/mistral-7b-it-v1.1.0/result_2023-11-13 23:21:20.json +444 -0
  41. BM-K/mistral-7b-it-v1.2.0/result_2023-11-14 04:32:36.json +444 -0
  42. BM-K/mistral-7b-it-v1.3.0/result_2023-11-15 15:01:55.json +444 -0
  43. BM-K/mistral-7b-it-v1.5.0/result_2023-11-16 23:47:33.json +444 -0
  44. BM-K/mistral-7b-it-v1.6.0/result_2023-11-19 13:15:31.json +444 -0
  45. BM-K/mistral-7b-it-v1.7.0/result_2023-11-20 09:22:14.json +444 -0
  46. BM-K/mistral-7b-it-v1.7.1/result_2023-11-21 00:39:48.json +444 -0
  47. BM-K/mistral-7b-it-v1.7.2/result_2023-11-30 06:16:02.json +444 -0
  48. BM-K/mistral-7b-it-v1.7.3/result_2023-12-04 23:53:23.json +444 -0
  49. BM-K/mistral-ko-7b-it-v2.0.0/result_2023-12-23 00:56:47.json +444 -0
  50. BM-K/mistral-ko-7b-it-v2.0.1/result_2023-12-26 12:34:11.json +444 -0
01-ai/Yi-6B-Chat/result_2023-12-27 01:03:40.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.2235494880546075,
5
+ "acc_stderr": 0.012174896631202614,
6
+ "acc_norm": 0.26621160409556316,
7
+ "acc_norm_stderr": 0.012915774781523216
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.30611431985660226,
11
+ "acc_stderr": 0.004599358920909541,
12
+ "acc_norm": 0.35222067317267475,
13
+ "acc_norm_stderr": 0.004766860907171539
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4619883040935672,
17
+ "acc_stderr": 0.03823727092882307,
18
+ "acc_norm": 0.4619883040935672,
19
+ "acc_norm_stderr": 0.03823727092882307
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.4854368932038835,
23
+ "acc_stderr": 0.04948637324026637,
24
+ "acc_norm": 0.4854368932038835,
25
+ "acc_norm_stderr": 0.04948637324026637
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.3537675606641124,
29
+ "acc_stderr": 0.017098184708161906,
30
+ "acc_norm": 0.3537675606641124,
31
+ "acc_norm_stderr": 0.017098184708161906
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.34074074074074073,
35
+ "acc_stderr": 0.04094376269996794,
36
+ "acc_norm": 0.34074074074074073,
37
+ "acc_norm_stderr": 0.04094376269996794
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.24,
41
+ "acc_stderr": 0.04292346959909283,
42
+ "acc_norm": 0.24,
43
+ "acc_norm_stderr": 0.04292346959909283
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.3276595744680851,
47
+ "acc_stderr": 0.030683020843231004,
48
+ "acc_norm": 0.3276595744680851,
49
+ "acc_norm_stderr": 0.030683020843231004
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3614457831325301,
53
+ "acc_stderr": 0.03740059382029319,
54
+ "acc_norm": 0.3614457831325301,
55
+ "acc_norm_stderr": 0.03740059382029319
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.43086816720257237,
59
+ "acc_stderr": 0.028125340983972718,
60
+ "acc_norm": 0.43086816720257237,
61
+ "acc_norm_stderr": 0.028125340983972718
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.34977578475336324,
65
+ "acc_stderr": 0.03200736719484504,
66
+ "acc_norm": 0.34977578475336324,
67
+ "acc_norm_stderr": 0.03200736719484504
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.366412213740458,
71
+ "acc_stderr": 0.04225875451969638,
72
+ "acc_norm": 0.366412213740458,
73
+ "acc_norm_stderr": 0.04225875451969638
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.36,
77
+ "acc_stderr": 0.04824181513244218,
78
+ "acc_norm": 0.36,
79
+ "acc_norm_stderr": 0.04824181513244218
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.42424242424242425,
83
+ "acc_stderr": 0.035212249088415824,
84
+ "acc_norm": 0.42424242424242425,
85
+ "acc_norm_stderr": 0.035212249088415824
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.5172413793103449,
89
+ "acc_stderr": 0.04164188720169375,
90
+ "acc_norm": 0.5172413793103449,
91
+ "acc_norm_stderr": 0.04164188720169375
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.3137254901960784,
95
+ "acc_stderr": 0.04617034827006718,
96
+ "acc_norm": 0.3137254901960784,
97
+ "acc_norm_stderr": 0.04617034827006718
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.46638655462184875,
101
+ "acc_stderr": 0.03240501447690071,
102
+ "acc_norm": 0.46638655462184875,
103
+ "acc_norm_stderr": 0.03240501447690071
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.3641025641025641,
107
+ "acc_stderr": 0.024396672985094764,
108
+ "acc_norm": 0.3641025641025641,
109
+ "acc_norm_stderr": 0.024396672985094764
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.52,
113
+ "acc_stderr": 0.05021167315686779,
114
+ "acc_norm": 0.52,
115
+ "acc_norm_stderr": 0.05021167315686779
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.27,
119
+ "acc_stderr": 0.04461960433384739,
120
+ "acc_norm": 0.27,
121
+ "acc_norm_stderr": 0.04461960433384739
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.4722222222222222,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.4722222222222222,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3891625615763547,
131
+ "acc_stderr": 0.03430462416103872,
132
+ "acc_norm": 0.3891625615763547,
133
+ "acc_norm_stderr": 0.03430462416103872
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.38064516129032255,
137
+ "acc_stderr": 0.027621717832907046,
138
+ "acc_norm": 0.38064516129032255,
139
+ "acc_norm_stderr": 0.027621717832907046
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.594017094017094,
143
+ "acc_stderr": 0.03217180182641086,
144
+ "acc_norm": 0.594017094017094,
145
+ "acc_norm_stderr": 0.03217180182641086
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4037735849056604,
149
+ "acc_stderr": 0.03019761160019795,
150
+ "acc_norm": 0.4037735849056604,
151
+ "acc_norm_stderr": 0.03019761160019795
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.41818181818181815,
155
+ "acc_stderr": 0.0472457740573157,
156
+ "acc_norm": 0.41818181818181815,
157
+ "acc_norm_stderr": 0.0472457740573157
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.3074074074074074,
161
+ "acc_stderr": 0.02813325257881563,
162
+ "acc_norm": 0.3074074074074074,
163
+ "acc_norm_stderr": 0.02813325257881563
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.271523178807947,
167
+ "acc_stderr": 0.036313298039696545,
168
+ "acc_norm": 0.271523178807947,
169
+ "acc_norm_stderr": 0.036313298039696545
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.4577114427860697,
173
+ "acc_stderr": 0.035228658640995975,
174
+ "acc_norm": 0.4577114427860697,
175
+ "acc_norm_stderr": 0.035228658640995975
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.36416184971098264,
179
+ "acc_stderr": 0.03669072477416907,
180
+ "acc_norm": 0.36416184971098264,
181
+ "acc_norm_stderr": 0.03669072477416907
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.3941798941798942,
185
+ "acc_stderr": 0.02516798233389414,
186
+ "acc_norm": 0.3941798941798942,
187
+ "acc_norm_stderr": 0.02516798233389414
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3125,
191
+ "acc_stderr": 0.038760854559127644,
192
+ "acc_norm": 0.3125,
193
+ "acc_norm_stderr": 0.038760854559127644
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.35,
197
+ "acc_stderr": 0.047937248544110196,
198
+ "acc_norm": 0.35,
199
+ "acc_norm_stderr": 0.047937248544110196
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.5,
203
+ "acc_stderr": 0.050251890762960605,
204
+ "acc_norm": 0.5,
205
+ "acc_norm_stderr": 0.050251890762960605
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.4479768786127168,
209
+ "acc_stderr": 0.026772990653361813,
210
+ "acc_norm": 0.4479768786127168,
211
+ "acc_norm_stderr": 0.026772990653361813
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.3496932515337423,
215
+ "acc_stderr": 0.03746668325470022,
216
+ "acc_norm": 0.3496932515337423,
217
+ "acc_norm_stderr": 0.03746668325470022
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.37962962962962965,
221
+ "acc_stderr": 0.027002521034516475,
222
+ "acc_norm": 0.37962962962962965,
223
+ "acc_norm_stderr": 0.027002521034516475
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.27,
227
+ "acc_stderr": 0.04461960433384739,
228
+ "acc_norm": 0.27,
229
+ "acc_norm_stderr": 0.04461960433384739
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.43005181347150256,
233
+ "acc_stderr": 0.035729543331448094,
234
+ "acc_norm": 0.43005181347150256,
235
+ "acc_norm_stderr": 0.035729543331448094
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2719298245614035,
239
+ "acc_stderr": 0.041857744240220575,
240
+ "acc_norm": 0.2719298245614035,
241
+ "acc_norm_stderr": 0.041857744240220575
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.3779816513761468,
245
+ "acc_stderr": 0.02078918706672812,
246
+ "acc_norm": 0.3779816513761468,
247
+ "acc_norm_stderr": 0.02078918706672812
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.48412698412698413,
251
+ "acc_stderr": 0.04469881854072606,
252
+ "acc_norm": 0.48412698412698413,
253
+ "acc_norm_stderr": 0.04469881854072606
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4411764705882353,
257
+ "acc_stderr": 0.028431095444176643,
258
+ "acc_norm": 0.4411764705882353,
259
+ "acc_norm_stderr": 0.028431095444176643
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.46,
263
+ "acc_stderr": 0.05009082659620332,
264
+ "acc_norm": 0.46,
265
+ "acc_norm_stderr": 0.05009082659620332
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.5867768595041323,
269
+ "acc_stderr": 0.04495087843548408,
270
+ "acc_norm": 0.5867768595041323,
271
+ "acc_norm_stderr": 0.04495087843548408
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4144736842105263,
275
+ "acc_stderr": 0.04008973785779206,
276
+ "acc_norm": 0.4144736842105263,
277
+ "acc_norm_stderr": 0.04008973785779206
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3300653594771242,
281
+ "acc_stderr": 0.019023726160724556,
282
+ "acc_norm": 0.3300653594771242,
283
+ "acc_norm_stderr": 0.019023726160724556
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.36879432624113473,
287
+ "acc_stderr": 0.028782227561347247,
288
+ "acc_norm": 0.36879432624113473,
289
+ "acc_norm_stderr": 0.028782227561347247
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.30357142857142855,
293
+ "acc_stderr": 0.04364226155841044,
294
+ "acc_norm": 0.30357142857142855,
295
+ "acc_norm_stderr": 0.04364226155841044
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3611111111111111,
299
+ "acc_stderr": 0.032757734861009996,
300
+ "acc_norm": 0.3611111111111111,
301
+ "acc_norm_stderr": 0.032757734861009996
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.23910614525139665,
305
+ "acc_stderr": 0.014265554192331161,
306
+ "acc_norm": 0.23910614525139665,
307
+ "acc_norm_stderr": 0.014265554192331161
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.42,
311
+ "acc_stderr": 0.049604496374885836,
312
+ "acc_norm": 0.42,
313
+ "acc_norm_stderr": 0.049604496374885836
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.48,
317
+ "acc_stderr": 0.050211673156867795,
318
+ "acc_norm": 0.48,
319
+ "acc_norm_stderr": 0.050211673156867795
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3161764705882353,
323
+ "acc_stderr": 0.02824568739146291,
324
+ "acc_norm": 0.3161764705882353,
325
+ "acc_norm_stderr": 0.02824568739146291
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5755102040816327,
329
+ "acc_stderr": 0.031642094879429414,
330
+ "acc_norm": 0.5755102040816327,
331
+ "acc_norm_stderr": 0.031642094879429414
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.350210970464135,
335
+ "acc_stderr": 0.03105239193758435,
336
+ "acc_norm": 0.350210970464135,
337
+ "acc_norm_stderr": 0.03105239193758435
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3200782268578879,
341
+ "acc_stderr": 0.011914791947638519,
342
+ "acc_norm": 0.3200782268578879,
343
+ "acc_norm_stderr": 0.011914791947638519
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.39215686274509803,
347
+ "acc_stderr": 0.03426712349247271,
348
+ "acc_norm": 0.39215686274509803,
349
+ "acc_norm_stderr": 0.03426712349247271
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.3696969696969697,
353
+ "acc_stderr": 0.03769430314512567,
354
+ "acc_norm": 0.3696969696969697,
355
+ "acc_norm_stderr": 0.03769430314512567
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.3072215422276622,
359
+ "mc1_stderr": 0.016150201321323013,
360
+ "mc2": 0.48699251655132686,
361
+ "mc2_stderr": 0.016174272005682996
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.33530106257378983,
365
+ "acc_stderr": 0.016230981232989827,
366
+ "acc_norm": 0.3742621015348288,
367
+ "acc_norm_stderr": 0.016637917789798732
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "01-ai/Yi-6B-Chat",
436
+ "model_sha": "36326f9bc1c8020e0cf29ea830ee5e6679a66a23",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
01-ai/Yi-6B/result_2023-12-27 01:03:17.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.2175767918088737,
5
+ "acc_stderr": 0.012057262020972502,
6
+ "acc_norm": 0.26109215017064846,
7
+ "acc_norm_stderr": 0.012835523909473855
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3021310495917148,
11
+ "acc_stderr": 0.00458243310963648,
12
+ "acc_norm": 0.35012945628360886,
13
+ "acc_norm_stderr": 0.004760354191370852
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4678362573099415,
17
+ "acc_stderr": 0.03826882417660368,
18
+ "acc_norm": 0.4678362573099415,
19
+ "acc_norm_stderr": 0.03826882417660368
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.4368932038834951,
23
+ "acc_stderr": 0.04911147107365777,
24
+ "acc_norm": 0.4368932038834951,
25
+ "acc_norm_stderr": 0.04911147107365777
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.39080459770114945,
29
+ "acc_stderr": 0.01744836606706253,
30
+ "acc_norm": 0.39080459770114945,
31
+ "acc_norm_stderr": 0.01744836606706253
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.35555555555555557,
35
+ "acc_stderr": 0.04135176749720386,
36
+ "acc_norm": 0.35555555555555557,
37
+ "acc_norm_stderr": 0.04135176749720386
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.29,
41
+ "acc_stderr": 0.045604802157206845,
42
+ "acc_norm": 0.29,
43
+ "acc_norm_stderr": 0.045604802157206845
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.37446808510638296,
47
+ "acc_stderr": 0.031639106653672915,
48
+ "acc_norm": 0.37446808510638296,
49
+ "acc_norm_stderr": 0.031639106653672915
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3614457831325301,
53
+ "acc_stderr": 0.03740059382029319,
54
+ "acc_norm": 0.3614457831325301,
55
+ "acc_norm_stderr": 0.03740059382029319
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.43086816720257237,
59
+ "acc_stderr": 0.028125340983972714,
60
+ "acc_norm": 0.43086816720257237,
61
+ "acc_norm_stderr": 0.028125340983972714
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.3183856502242152,
65
+ "acc_stderr": 0.03126580522513713,
66
+ "acc_norm": 0.3183856502242152,
67
+ "acc_norm_stderr": 0.03126580522513713
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.3893129770992366,
71
+ "acc_stderr": 0.04276486542814591,
72
+ "acc_norm": 0.3893129770992366,
73
+ "acc_norm_stderr": 0.04276486542814591
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.4,
77
+ "acc_stderr": 0.049236596391733084,
78
+ "acc_norm": 0.4,
79
+ "acc_norm_stderr": 0.049236596391733084
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.42424242424242425,
83
+ "acc_stderr": 0.035212249088415824,
84
+ "acc_norm": 0.42424242424242425,
85
+ "acc_norm_stderr": 0.035212249088415824
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4896551724137931,
89
+ "acc_stderr": 0.041657747757287644,
90
+ "acc_norm": 0.4896551724137931,
91
+ "acc_norm_stderr": 0.041657747757287644
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.29411764705882354,
95
+ "acc_stderr": 0.04533838195929774,
96
+ "acc_norm": 0.29411764705882354,
97
+ "acc_norm_stderr": 0.04533838195929774
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.47058823529411764,
101
+ "acc_stderr": 0.03242225027115007,
102
+ "acc_norm": 0.47058823529411764,
103
+ "acc_norm_stderr": 0.03242225027115007
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.34102564102564104,
107
+ "acc_stderr": 0.02403548967633507,
108
+ "acc_norm": 0.34102564102564104,
109
+ "acc_norm_stderr": 0.02403548967633507
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.54,
113
+ "acc_stderr": 0.05009082659620332,
114
+ "acc_norm": 0.54,
115
+ "acc_norm_stderr": 0.05009082659620332
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.3,
119
+ "acc_stderr": 0.046056618647183814,
120
+ "acc_norm": 0.3,
121
+ "acc_norm_stderr": 0.046056618647183814
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.49074074074074076,
125
+ "acc_stderr": 0.04832853553437055,
126
+ "acc_norm": 0.49074074074074076,
127
+ "acc_norm_stderr": 0.04832853553437055
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3842364532019704,
131
+ "acc_stderr": 0.03422398565657551,
132
+ "acc_norm": 0.3842364532019704,
133
+ "acc_norm_stderr": 0.03422398565657551
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.36774193548387096,
137
+ "acc_stderr": 0.027430866579973474,
138
+ "acc_norm": 0.36774193548387096,
139
+ "acc_norm_stderr": 0.027430866579973474
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6410256410256411,
143
+ "acc_stderr": 0.03142616993791923,
144
+ "acc_norm": 0.6410256410256411,
145
+ "acc_norm_stderr": 0.03142616993791923
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.3660377358490566,
149
+ "acc_stderr": 0.029647813539365263,
150
+ "acc_norm": 0.3660377358490566,
151
+ "acc_norm_stderr": 0.029647813539365263
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.39090909090909093,
155
+ "acc_stderr": 0.04673752333670237,
156
+ "acc_norm": 0.39090909090909093,
157
+ "acc_norm_stderr": 0.04673752333670237
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2851851851851852,
161
+ "acc_stderr": 0.027528599210340492,
162
+ "acc_norm": 0.2851851851851852,
163
+ "acc_norm_stderr": 0.027528599210340492
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2913907284768212,
167
+ "acc_stderr": 0.03710185726119995,
168
+ "acc_norm": 0.2913907284768212,
169
+ "acc_norm_stderr": 0.03710185726119995
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.48756218905472637,
173
+ "acc_stderr": 0.03534439848539579,
174
+ "acc_norm": 0.48756218905472637,
175
+ "acc_norm_stderr": 0.03534439848539579
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3583815028901734,
179
+ "acc_stderr": 0.03656343653353159,
180
+ "acc_norm": 0.3583815028901734,
181
+ "acc_norm_stderr": 0.03656343653353159
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.36772486772486773,
185
+ "acc_stderr": 0.024833839825562424,
186
+ "acc_norm": 0.36772486772486773,
187
+ "acc_norm_stderr": 0.024833839825562424
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3055555555555556,
191
+ "acc_stderr": 0.03852084696008534,
192
+ "acc_norm": 0.3055555555555556,
193
+ "acc_norm_stderr": 0.03852084696008534
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.29,
197
+ "acc_stderr": 0.045604802157206845,
198
+ "acc_norm": 0.29,
199
+ "acc_norm_stderr": 0.045604802157206845
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.54,
203
+ "acc_stderr": 0.05009082659620333,
204
+ "acc_norm": 0.54,
205
+ "acc_norm_stderr": 0.05009082659620333
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.476878612716763,
209
+ "acc_stderr": 0.026890297881303128,
210
+ "acc_norm": 0.476878612716763,
211
+ "acc_norm_stderr": 0.026890297881303128
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.36809815950920244,
215
+ "acc_stderr": 0.03789213935838396,
216
+ "acc_norm": 0.36809815950920244,
217
+ "acc_norm_stderr": 0.03789213935838396
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4012345679012346,
221
+ "acc_stderr": 0.027272582849839792,
222
+ "acc_norm": 0.4012345679012346,
223
+ "acc_norm_stderr": 0.027272582849839792
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.32,
227
+ "acc_stderr": 0.04688261722621504,
228
+ "acc_norm": 0.32,
229
+ "acc_norm_stderr": 0.04688261722621504
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.37823834196891193,
233
+ "acc_stderr": 0.03499807276193338,
234
+ "acc_norm": 0.37823834196891193,
235
+ "acc_norm_stderr": 0.03499807276193338
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2543859649122807,
239
+ "acc_stderr": 0.040969851398436695,
240
+ "acc_norm": 0.2543859649122807,
241
+ "acc_norm_stderr": 0.040969851398436695
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.3669724770642202,
245
+ "acc_stderr": 0.020664675659520536,
246
+ "acc_norm": 0.3669724770642202,
247
+ "acc_norm_stderr": 0.020664675659520536
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.42857142857142855,
251
+ "acc_stderr": 0.0442626668137991,
252
+ "acc_norm": 0.42857142857142855,
253
+ "acc_norm_stderr": 0.0442626668137991
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4477124183006536,
257
+ "acc_stderr": 0.02847293847803353,
258
+ "acc_norm": 0.4477124183006536,
259
+ "acc_norm_stderr": 0.02847293847803353
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.48,
263
+ "acc_stderr": 0.050211673156867795,
264
+ "acc_norm": 0.48,
265
+ "acc_norm_stderr": 0.050211673156867795
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.5950413223140496,
269
+ "acc_stderr": 0.04481137755942469,
270
+ "acc_norm": 0.5950413223140496,
271
+ "acc_norm_stderr": 0.04481137755942469
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40131578947368424,
275
+ "acc_stderr": 0.039889037033362836,
276
+ "acc_norm": 0.40131578947368424,
277
+ "acc_norm_stderr": 0.039889037033362836
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.32189542483660133,
281
+ "acc_stderr": 0.018901015322093092,
282
+ "acc_norm": 0.32189542483660133,
283
+ "acc_norm_stderr": 0.018901015322093092
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3900709219858156,
287
+ "acc_stderr": 0.02909767559946393,
288
+ "acc_norm": 0.3900709219858156,
289
+ "acc_norm_stderr": 0.02909767559946393
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.3392857142857143,
293
+ "acc_stderr": 0.04493949068613539,
294
+ "acc_norm": 0.3392857142857143,
295
+ "acc_norm_stderr": 0.04493949068613539
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.39351851851851855,
299
+ "acc_stderr": 0.03331747876370312,
300
+ "acc_norm": 0.39351851851851855,
301
+ "acc_norm_stderr": 0.03331747876370312
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.24581005586592178,
305
+ "acc_stderr": 0.014400296429225612,
306
+ "acc_norm": 0.24581005586592178,
307
+ "acc_norm_stderr": 0.014400296429225612
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.41,
311
+ "acc_stderr": 0.049431107042371025,
312
+ "acc_norm": 0.41,
313
+ "acc_norm_stderr": 0.049431107042371025
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.5,
317
+ "acc_stderr": 0.050251890762960605,
318
+ "acc_norm": 0.5,
319
+ "acc_norm_stderr": 0.050251890762960605
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.35661764705882354,
323
+ "acc_stderr": 0.02909720956841196,
324
+ "acc_norm": 0.35661764705882354,
325
+ "acc_norm_stderr": 0.02909720956841196
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.563265306122449,
329
+ "acc_stderr": 0.031751952375833226,
330
+ "acc_norm": 0.563265306122449,
331
+ "acc_norm_stderr": 0.031751952375833226
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.3459915611814346,
335
+ "acc_stderr": 0.030964810588786713,
336
+ "acc_norm": 0.3459915611814346,
337
+ "acc_norm_stderr": 0.030964810588786713
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3318122555410691,
341
+ "acc_stderr": 0.012026088259897634,
342
+ "acc_norm": 0.3318122555410691,
343
+ "acc_norm_stderr": 0.012026088259897634
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4068627450980392,
347
+ "acc_stderr": 0.03447891136353382,
348
+ "acc_norm": 0.4068627450980392,
349
+ "acc_norm_stderr": 0.03447891136353382
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.4303030303030303,
353
+ "acc_stderr": 0.03866225962879077,
354
+ "acc_norm": 0.4303030303030303,
355
+ "acc_norm_stderr": 0.03866225962879077
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2998776009791922,
359
+ "mc1_stderr": 0.016040352966713616,
360
+ "mc2": 0.47599173122840593,
361
+ "mc2_stderr": 0.015773874222919516
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3305785123966942,
365
+ "acc_stderr": 0.0161734232988457,
366
+ "acc_norm": 0.40968122786304606,
367
+ "acc_norm_stderr": 0.01690756819221947
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "01-ai/Yi-6B",
436
+ "model_sha": "b881162e08d0fa65011cb53f2c51544e1b623112",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama-2-ko-en-instruct-v1/result_2023-10-12 11:14:34.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.39505119453924914,
5
+ "acc_stderr": 0.014285898292938165,
6
+ "acc_norm": 0.4445392491467577,
7
+ "acc_norm_stderr": 0.014521226405627077
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.41545508862776337,
11
+ "acc_stderr": 0.004917931778593191,
12
+ "acc_norm": 0.5571599283011353,
13
+ "acc_norm_stderr": 0.004957068377516512
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.47953216374269003,
17
+ "acc_stderr": 0.0383161053282193,
18
+ "acc_norm": 0.47953216374269003,
19
+ "acc_norm_stderr": 0.0383161053282193
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.4174757281553398,
23
+ "acc_stderr": 0.048828405482122375,
24
+ "acc_norm": 0.4174757281553398,
25
+ "acc_norm_stderr": 0.048828405482122375
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5108556832694764,
29
+ "acc_stderr": 0.017875748840242407,
30
+ "acc_norm": 0.5108556832694764,
31
+ "acc_norm_stderr": 0.017875748840242407
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4222222222222222,
35
+ "acc_stderr": 0.04266763404099582,
36
+ "acc_norm": 0.4222222222222222,
37
+ "acc_norm_stderr": 0.04266763404099582
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.25,
41
+ "acc_stderr": 0.04351941398892446,
42
+ "acc_norm": 0.25,
43
+ "acc_norm_stderr": 0.04351941398892446
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.32340425531914896,
47
+ "acc_stderr": 0.03057944277361033,
48
+ "acc_norm": 0.32340425531914896,
49
+ "acc_norm_stderr": 0.03057944277361033
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.41566265060240964,
53
+ "acc_stderr": 0.03836722176598053,
54
+ "acc_norm": 0.41566265060240964,
55
+ "acc_norm_stderr": 0.03836722176598053
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4758842443729904,
59
+ "acc_stderr": 0.02836504154256457,
60
+ "acc_norm": 0.4758842443729904,
61
+ "acc_norm_stderr": 0.02836504154256457
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.3811659192825112,
65
+ "acc_stderr": 0.03259625118416828,
66
+ "acc_norm": 0.3811659192825112,
67
+ "acc_norm_stderr": 0.03259625118416828
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.46564885496183206,
71
+ "acc_stderr": 0.043749285605997376,
72
+ "acc_norm": 0.46564885496183206,
73
+ "acc_norm_stderr": 0.043749285605997376
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.3,
77
+ "acc_stderr": 0.046056618647183814,
78
+ "acc_norm": 0.3,
79
+ "acc_norm_stderr": 0.046056618647183814
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.4696969696969697,
83
+ "acc_stderr": 0.03555804051763929,
84
+ "acc_norm": 0.4696969696969697,
85
+ "acc_norm_stderr": 0.03555804051763929
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4068965517241379,
89
+ "acc_stderr": 0.04093793981266237,
90
+ "acc_norm": 0.4068965517241379,
91
+ "acc_norm_stderr": 0.04093793981266237
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.2549019607843137,
95
+ "acc_stderr": 0.043364327079931785,
96
+ "acc_norm": 0.2549019607843137,
97
+ "acc_norm_stderr": 0.043364327079931785
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.36134453781512604,
101
+ "acc_stderr": 0.031204691225150013,
102
+ "acc_norm": 0.36134453781512604,
103
+ "acc_norm_stderr": 0.031204691225150013
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.3564102564102564,
107
+ "acc_stderr": 0.02428314052946728,
108
+ "acc_norm": 0.3564102564102564,
109
+ "acc_norm_stderr": 0.02428314052946728
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.47,
113
+ "acc_stderr": 0.050161355804659205,
114
+ "acc_norm": 0.47,
115
+ "acc_norm_stderr": 0.050161355804659205
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.35,
119
+ "acc_stderr": 0.047937248544110196,
120
+ "acc_norm": 0.35,
121
+ "acc_norm_stderr": 0.047937248544110196
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.4166666666666667,
125
+ "acc_stderr": 0.04766075165356461,
126
+ "acc_norm": 0.4166666666666667,
127
+ "acc_norm_stderr": 0.04766075165356461
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3448275862068966,
131
+ "acc_stderr": 0.03344283744280459,
132
+ "acc_norm": 0.3448275862068966,
133
+ "acc_norm_stderr": 0.03344283744280459
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4483870967741935,
137
+ "acc_stderr": 0.028292056830112735,
138
+ "acc_norm": 0.4483870967741935,
139
+ "acc_norm_stderr": 0.028292056830112735
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6068376068376068,
143
+ "acc_stderr": 0.03199957924651047,
144
+ "acc_norm": 0.6068376068376068,
145
+ "acc_norm_stderr": 0.03199957924651047
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.43018867924528303,
149
+ "acc_stderr": 0.030471445867183238,
150
+ "acc_norm": 0.43018867924528303,
151
+ "acc_norm_stderr": 0.030471445867183238
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4909090909090909,
155
+ "acc_stderr": 0.04788339768702861,
156
+ "acc_norm": 0.4909090909090909,
157
+ "acc_norm_stderr": 0.04788339768702861
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2518518518518518,
161
+ "acc_stderr": 0.026466117538959916,
162
+ "acc_norm": 0.2518518518518518,
163
+ "acc_norm_stderr": 0.026466117538959916
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.26490066225165565,
167
+ "acc_stderr": 0.03603038545360384,
168
+ "acc_norm": 0.26490066225165565,
169
+ "acc_norm_stderr": 0.03603038545360384
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5124378109452736,
173
+ "acc_stderr": 0.0353443984853958,
174
+ "acc_norm": 0.5124378109452736,
175
+ "acc_norm_stderr": 0.0353443984853958
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3583815028901734,
179
+ "acc_stderr": 0.03656343653353159,
180
+ "acc_norm": 0.3583815028901734,
181
+ "acc_norm_stderr": 0.03656343653353159
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.30687830687830686,
185
+ "acc_stderr": 0.023752928712112126,
186
+ "acc_norm": 0.30687830687830686,
187
+ "acc_norm_stderr": 0.023752928712112126
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.2986111111111111,
191
+ "acc_stderr": 0.03827052357950756,
192
+ "acc_norm": 0.2986111111111111,
193
+ "acc_norm_stderr": 0.03827052357950756
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.22,
197
+ "acc_stderr": 0.04163331998932269,
198
+ "acc_norm": 0.22,
199
+ "acc_norm_stderr": 0.04163331998932269
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.53,
203
+ "acc_stderr": 0.050161355804659205,
204
+ "acc_norm": 0.53,
205
+ "acc_norm_stderr": 0.050161355804659205
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.44508670520231214,
209
+ "acc_stderr": 0.02675625512966377,
210
+ "acc_norm": 0.44508670520231214,
211
+ "acc_norm_stderr": 0.02675625512966377
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.3558282208588957,
215
+ "acc_stderr": 0.03761521380046734,
216
+ "acc_norm": 0.3558282208588957,
217
+ "acc_norm_stderr": 0.03761521380046734
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.44135802469135804,
221
+ "acc_stderr": 0.027628737155668777,
222
+ "acc_norm": 0.44135802469135804,
223
+ "acc_norm_stderr": 0.027628737155668777
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.24,
227
+ "acc_stderr": 0.04292346959909283,
228
+ "acc_norm": 0.24,
229
+ "acc_norm_stderr": 0.04292346959909283
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.40414507772020725,
233
+ "acc_stderr": 0.0354150857888402,
234
+ "acc_norm": 0.40414507772020725,
235
+ "acc_norm_stderr": 0.0354150857888402
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2631578947368421,
239
+ "acc_stderr": 0.041424397194893624,
240
+ "acc_norm": 0.2631578947368421,
241
+ "acc_norm_stderr": 0.041424397194893624
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.4091743119266055,
245
+ "acc_stderr": 0.02108067026443373,
246
+ "acc_norm": 0.4091743119266055,
247
+ "acc_norm_stderr": 0.02108067026443373
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.1984126984126984,
251
+ "acc_stderr": 0.03567016675276863,
252
+ "acc_norm": 0.1984126984126984,
253
+ "acc_norm_stderr": 0.03567016675276863
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.39869281045751637,
257
+ "acc_stderr": 0.02803609227389177,
258
+ "acc_norm": 0.39869281045751637,
259
+ "acc_norm_stderr": 0.02803609227389177
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.42,
263
+ "acc_stderr": 0.049604496374885836,
264
+ "acc_norm": 0.42,
265
+ "acc_norm_stderr": 0.049604496374885836
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6033057851239669,
269
+ "acc_stderr": 0.04465869780531009,
270
+ "acc_norm": 0.6033057851239669,
271
+ "acc_norm_stderr": 0.04465869780531009
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4342105263157895,
275
+ "acc_stderr": 0.040335656678483184,
276
+ "acc_norm": 0.4342105263157895,
277
+ "acc_norm_stderr": 0.040335656678483184
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3333333333333333,
281
+ "acc_stderr": 0.0190709855896875,
282
+ "acc_norm": 0.3333333333333333,
283
+ "acc_norm_stderr": 0.0190709855896875
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3191489361702128,
287
+ "acc_stderr": 0.027807990141320207,
288
+ "acc_norm": 0.3191489361702128,
289
+ "acc_norm_stderr": 0.027807990141320207
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.19642857142857142,
293
+ "acc_stderr": 0.03770970049347019,
294
+ "acc_norm": 0.19642857142857142,
295
+ "acc_norm_stderr": 0.03770970049347019
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.25462962962962965,
299
+ "acc_stderr": 0.02971127586000534,
300
+ "acc_norm": 0.25462962962962965,
301
+ "acc_norm_stderr": 0.02971127586000534
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2424581005586592,
305
+ "acc_stderr": 0.01433352205921789,
306
+ "acc_norm": 0.2424581005586592,
307
+ "acc_norm_stderr": 0.01433352205921789
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.31,
311
+ "acc_stderr": 0.04648231987117316,
312
+ "acc_norm": 0.31,
313
+ "acc_norm_stderr": 0.04648231987117316
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.44,
317
+ "acc_stderr": 0.04988876515698589,
318
+ "acc_norm": 0.44,
319
+ "acc_norm_stderr": 0.04988876515698589
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.22426470588235295,
323
+ "acc_stderr": 0.02533684856333237,
324
+ "acc_norm": 0.22426470588235295,
325
+ "acc_norm_stderr": 0.02533684856333237
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4163265306122449,
329
+ "acc_stderr": 0.03155782816556164,
330
+ "acc_norm": 0.4163265306122449,
331
+ "acc_norm_stderr": 0.03155782816556164
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.41350210970464135,
335
+ "acc_stderr": 0.03205649904851859,
336
+ "acc_norm": 0.41350210970464135,
337
+ "acc_norm_stderr": 0.03205649904851859
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.29335071707953064,
341
+ "acc_stderr": 0.011628520449582076,
342
+ "acc_norm": 0.29335071707953064,
343
+ "acc_norm_stderr": 0.011628520449582076
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.3627450980392157,
347
+ "acc_stderr": 0.033744993563193555,
348
+ "acc_norm": 0.3627450980392157,
349
+ "acc_norm_stderr": 0.033744993563193555
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.43636363636363634,
353
+ "acc_stderr": 0.03872592983524754,
354
+ "acc_norm": 0.43636363636363634,
355
+ "acc_norm_stderr": 0.03872592983524754
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2962056303549572,
359
+ "mc1_stderr": 0.01598359510181139,
360
+ "mc2": 0.4602391231259313,
361
+ "mc2_stderr": 0.015191570633369808
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.4757969303423849,
365
+ "acc_stderr": 0.017170202466520748,
366
+ "acc_norm": 0.5454545454545454,
367
+ "acc_norm_stderr": 0.017119172208061504
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama-2-ko-en-instruct-v1",
436
+ "model_sha": "aee07500d61a1d5d214cf0bc0040650957cf3da0",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-dpo-13b-test3/result_2023-11-30 08:08:14.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.4112627986348123,
5
+ "acc_stderr": 0.014379441068522077,
6
+ "acc_norm": 0.45733788395904434,
7
+ "acc_norm_stderr": 0.014558106543924067
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.43288189603664606,
11
+ "acc_stderr": 0.004944620712318274,
12
+ "acc_norm": 0.5816570404301932,
13
+ "acc_norm_stderr": 0.004922789247319874
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5321637426900585,
17
+ "acc_stderr": 0.038268824176603704,
18
+ "acc_norm": 0.5321637426900585,
19
+ "acc_norm_stderr": 0.038268824176603704
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5242718446601942,
23
+ "acc_stderr": 0.049449010929737795,
24
+ "acc_norm": 0.5242718446601942,
25
+ "acc_norm_stderr": 0.049449010929737795
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5517241379310345,
29
+ "acc_stderr": 0.01778403453499242,
30
+ "acc_norm": 0.5517241379310345,
31
+ "acc_norm_stderr": 0.01778403453499242
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4074074074074074,
35
+ "acc_stderr": 0.0424463323835323,
36
+ "acc_norm": 0.4074074074074074,
37
+ "acc_norm_stderr": 0.0424463323835323
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.27,
41
+ "acc_stderr": 0.04461960433384741,
42
+ "acc_norm": 0.27,
43
+ "acc_norm_stderr": 0.04461960433384741
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.40425531914893614,
47
+ "acc_stderr": 0.03208115750788684,
48
+ "acc_norm": 0.40425531914893614,
49
+ "acc_norm_stderr": 0.03208115750788684
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.43373493975903615,
53
+ "acc_stderr": 0.03858158940685515,
54
+ "acc_norm": 0.43373493975903615,
55
+ "acc_norm_stderr": 0.03858158940685515
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4919614147909968,
59
+ "acc_stderr": 0.028394421370984545,
60
+ "acc_norm": 0.4919614147909968,
61
+ "acc_norm_stderr": 0.028394421370984545
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.5022421524663677,
65
+ "acc_stderr": 0.033557465352232634,
66
+ "acc_norm": 0.5022421524663677,
67
+ "acc_norm_stderr": 0.033557465352232634
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.48091603053435117,
71
+ "acc_stderr": 0.04382094705550989,
72
+ "acc_norm": 0.48091603053435117,
73
+ "acc_norm_stderr": 0.04382094705550989
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.39,
77
+ "acc_stderr": 0.04902071300001975,
78
+ "acc_norm": 0.39,
79
+ "acc_norm_stderr": 0.04902071300001975
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5858585858585859,
83
+ "acc_stderr": 0.035094383488796295,
84
+ "acc_norm": 0.5858585858585859,
85
+ "acc_norm_stderr": 0.035094383488796295
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.36551724137931035,
89
+ "acc_stderr": 0.040131241954243856,
90
+ "acc_norm": 0.36551724137931035,
91
+ "acc_norm_stderr": 0.040131241954243856
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.22549019607843138,
95
+ "acc_stderr": 0.04158307533083286,
96
+ "acc_norm": 0.22549019607843138,
97
+ "acc_norm_stderr": 0.04158307533083286
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.46638655462184875,
101
+ "acc_stderr": 0.03240501447690071,
102
+ "acc_norm": 0.46638655462184875,
103
+ "acc_norm_stderr": 0.03240501447690071
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.45384615384615384,
107
+ "acc_stderr": 0.02524277098712617,
108
+ "acc_norm": 0.45384615384615384,
109
+ "acc_norm_stderr": 0.02524277098712617
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.55,
113
+ "acc_stderr": 0.04999999999999999,
114
+ "acc_norm": 0.55,
115
+ "acc_norm_stderr": 0.04999999999999999
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.37,
119
+ "acc_stderr": 0.04852365870939099,
120
+ "acc_norm": 0.37,
121
+ "acc_norm_stderr": 0.04852365870939099
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5370370370370371,
125
+ "acc_stderr": 0.04820403072760627,
126
+ "acc_norm": 0.5370370370370371,
127
+ "acc_norm_stderr": 0.04820403072760627
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.35960591133004927,
131
+ "acc_stderr": 0.03376458246509567,
132
+ "acc_norm": 0.35960591133004927,
133
+ "acc_norm_stderr": 0.03376458246509567
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4612903225806452,
137
+ "acc_stderr": 0.02835863485983692,
138
+ "acc_norm": 0.4612903225806452,
139
+ "acc_norm_stderr": 0.02835863485983692
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6538461538461539,
143
+ "acc_stderr": 0.031166957367235903,
144
+ "acc_norm": 0.6538461538461539,
145
+ "acc_norm_stderr": 0.031166957367235903
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4490566037735849,
149
+ "acc_stderr": 0.030612730713641092,
150
+ "acc_norm": 0.4490566037735849,
151
+ "acc_norm_stderr": 0.030612730713641092
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.6,
155
+ "acc_stderr": 0.0469237132203465,
156
+ "acc_norm": 0.6,
157
+ "acc_norm_stderr": 0.0469237132203465
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2518518518518518,
161
+ "acc_stderr": 0.026466117538959916,
162
+ "acc_norm": 0.2518518518518518,
163
+ "acc_norm_stderr": 0.026466117538959916
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.31125827814569534,
167
+ "acc_stderr": 0.03780445850526733,
168
+ "acc_norm": 0.31125827814569534,
169
+ "acc_norm_stderr": 0.03780445850526733
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5572139303482587,
173
+ "acc_stderr": 0.03512310964123937,
174
+ "acc_norm": 0.5572139303482587,
175
+ "acc_norm_stderr": 0.03512310964123937
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.41040462427745666,
179
+ "acc_stderr": 0.03750757044895538,
180
+ "acc_norm": 0.41040462427745666,
181
+ "acc_norm_stderr": 0.03750757044895538
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.29894179894179895,
185
+ "acc_stderr": 0.023577604791655812,
186
+ "acc_norm": 0.29894179894179895,
187
+ "acc_norm_stderr": 0.023577604791655812
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3958333333333333,
191
+ "acc_stderr": 0.04089465449325583,
192
+ "acc_norm": 0.3958333333333333,
193
+ "acc_norm_stderr": 0.04089465449325583
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.3,
197
+ "acc_stderr": 0.046056618647183814,
198
+ "acc_norm": 0.3,
199
+ "acc_norm_stderr": 0.046056618647183814
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.65,
203
+ "acc_stderr": 0.04793724854411018,
204
+ "acc_norm": 0.65,
205
+ "acc_norm_stderr": 0.04793724854411018
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5,
209
+ "acc_stderr": 0.026919095102908273,
210
+ "acc_norm": 0.5,
211
+ "acc_norm_stderr": 0.026919095102908273
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.5828220858895705,
215
+ "acc_stderr": 0.03874102859818082,
216
+ "acc_norm": 0.5828220858895705,
217
+ "acc_norm_stderr": 0.03874102859818082
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.49691358024691357,
221
+ "acc_stderr": 0.027820214158594377,
222
+ "acc_norm": 0.49691358024691357,
223
+ "acc_norm_stderr": 0.027820214158594377
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.36,
227
+ "acc_stderr": 0.04824181513244218,
228
+ "acc_norm": 0.36,
229
+ "acc_norm_stderr": 0.04824181513244218
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.538860103626943,
233
+ "acc_stderr": 0.03597524411734578,
234
+ "acc_norm": 0.538860103626943,
235
+ "acc_norm_stderr": 0.03597524411734578
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.21052631578947367,
239
+ "acc_stderr": 0.03835153954399419,
240
+ "acc_norm": 0.21052631578947367,
241
+ "acc_norm_stderr": 0.03835153954399419
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5926605504587156,
245
+ "acc_stderr": 0.02106598624441288,
246
+ "acc_norm": 0.5926605504587156,
247
+ "acc_norm_stderr": 0.02106598624441288
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3333333333333333,
251
+ "acc_stderr": 0.04216370213557835,
252
+ "acc_norm": 0.3333333333333333,
253
+ "acc_norm_stderr": 0.04216370213557835
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.42483660130718953,
257
+ "acc_stderr": 0.02830457667314111,
258
+ "acc_norm": 0.42483660130718953,
259
+ "acc_norm_stderr": 0.02830457667314111
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.47,
263
+ "acc_stderr": 0.05016135580465919,
264
+ "acc_norm": 0.47,
265
+ "acc_norm_stderr": 0.05016135580465919
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.5950413223140496,
269
+ "acc_stderr": 0.04481137755942469,
270
+ "acc_norm": 0.5950413223140496,
271
+ "acc_norm_stderr": 0.04481137755942469
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.375,
275
+ "acc_stderr": 0.039397364351956274,
276
+ "acc_norm": 0.375,
277
+ "acc_norm_stderr": 0.039397364351956274
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.4215686274509804,
281
+ "acc_stderr": 0.01997742260022747,
282
+ "acc_norm": 0.4215686274509804,
283
+ "acc_norm_stderr": 0.01997742260022747
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3546099290780142,
287
+ "acc_stderr": 0.02853865002887864,
288
+ "acc_norm": 0.3546099290780142,
289
+ "acc_norm_stderr": 0.02853865002887864
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.2767857142857143,
293
+ "acc_stderr": 0.04246624336697624,
294
+ "acc_norm": 0.2767857142857143,
295
+ "acc_norm_stderr": 0.04246624336697624
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.33796296296296297,
299
+ "acc_stderr": 0.03225941352631295,
300
+ "acc_norm": 0.33796296296296297,
301
+ "acc_norm_stderr": 0.03225941352631295
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2435754189944134,
305
+ "acc_stderr": 0.01435591196476786,
306
+ "acc_norm": 0.2435754189944134,
307
+ "acc_norm_stderr": 0.01435591196476786
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.36,
311
+ "acc_stderr": 0.04824181513244218,
312
+ "acc_norm": 0.36,
313
+ "acc_norm_stderr": 0.04824181513244218
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.41,
317
+ "acc_stderr": 0.049431107042371025,
318
+ "acc_norm": 0.41,
319
+ "acc_norm_stderr": 0.049431107042371025
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3860294117647059,
323
+ "acc_stderr": 0.029573269134411124,
324
+ "acc_norm": 0.3860294117647059,
325
+ "acc_norm_stderr": 0.029573269134411124
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5061224489795918,
329
+ "acc_stderr": 0.03200682020163907,
330
+ "acc_norm": 0.5061224489795918,
331
+ "acc_norm_stderr": 0.03200682020163907
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6751054852320675,
335
+ "acc_stderr": 0.030486039389105303,
336
+ "acc_norm": 0.6751054852320675,
337
+ "acc_norm_stderr": 0.030486039389105303
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3774445893089961,
341
+ "acc_stderr": 0.012380680911165804,
342
+ "acc_norm": 0.3774445893089961,
343
+ "acc_norm_stderr": 0.012380680911165804
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.5245098039215687,
347
+ "acc_stderr": 0.035050931943487976,
348
+ "acc_norm": 0.5245098039215687,
349
+ "acc_norm_stderr": 0.035050931943487976
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5575757575757576,
353
+ "acc_stderr": 0.03878372113711275,
354
+ "acc_norm": 0.5575757575757576,
355
+ "acc_norm_stderr": 0.03878372113711275
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.31946144430844553,
359
+ "mc1_stderr": 0.0163226441829605,
360
+ "mc2": 0.4756188079524156,
361
+ "mc2_stderr": 0.015396392654893808
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5100354191263282,
365
+ "acc_stderr": 0.01718689128689406,
366
+ "acc_norm": 0.5832349468713105,
367
+ "acc_norm_stderr": 0.01695048914610882
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-dpo-13b-test3",
436
+ "model_sha": "d70fdfed2e0b43ac6715ee5ec24801fd2bd5c25d",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-dpo-13b-v1/result_2023-11-18 17:03:07.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.39505119453924914,
5
+ "acc_stderr": 0.014285898292938165,
6
+ "acc_norm": 0.45819112627986347,
7
+ "acc_norm_stderr": 0.014560220308714702
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.42561242780322645,
11
+ "acc_stderr": 0.004934250390879782,
12
+ "acc_norm": 0.569308902609042,
13
+ "acc_norm_stderr": 0.004941609820763589
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.47953216374269003,
17
+ "acc_stderr": 0.0383161053282193,
18
+ "acc_norm": 0.47953216374269003,
19
+ "acc_norm_stderr": 0.0383161053282193
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.4563106796116505,
23
+ "acc_stderr": 0.049318019942204146,
24
+ "acc_norm": 0.4563106796116505,
25
+ "acc_norm_stderr": 0.049318019942204146
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5210727969348659,
29
+ "acc_stderr": 0.01786407678621291,
30
+ "acc_norm": 0.5210727969348659,
31
+ "acc_norm_stderr": 0.01786407678621291
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4888888888888889,
35
+ "acc_stderr": 0.04318275491977976,
36
+ "acc_norm": 0.4888888888888889,
37
+ "acc_norm_stderr": 0.04318275491977976
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.3,
41
+ "acc_stderr": 0.046056618647183814,
42
+ "acc_norm": 0.3,
43
+ "acc_norm_stderr": 0.046056618647183814
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.40425531914893614,
47
+ "acc_stderr": 0.032081157507886836,
48
+ "acc_norm": 0.40425531914893614,
49
+ "acc_norm_stderr": 0.032081157507886836
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39759036144578314,
53
+ "acc_stderr": 0.03809973084540218,
54
+ "acc_norm": 0.39759036144578314,
55
+ "acc_norm_stderr": 0.03809973084540218
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4533762057877814,
59
+ "acc_stderr": 0.02827435985489424,
60
+ "acc_norm": 0.4533762057877814,
61
+ "acc_norm_stderr": 0.02827435985489424
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.47085201793721976,
65
+ "acc_stderr": 0.03350073248773403,
66
+ "acc_norm": 0.47085201793721976,
67
+ "acc_norm_stderr": 0.03350073248773403
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.42748091603053434,
71
+ "acc_stderr": 0.043389203057924,
72
+ "acc_norm": 0.42748091603053434,
73
+ "acc_norm_stderr": 0.043389203057924
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.42,
77
+ "acc_stderr": 0.049604496374885836,
78
+ "acc_norm": 0.42,
79
+ "acc_norm_stderr": 0.049604496374885836
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5151515151515151,
83
+ "acc_stderr": 0.0356071651653106,
84
+ "acc_norm": 0.5151515151515151,
85
+ "acc_norm_stderr": 0.0356071651653106
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.45517241379310347,
89
+ "acc_stderr": 0.04149886942192117,
90
+ "acc_norm": 0.45517241379310347,
91
+ "acc_norm_stderr": 0.04149886942192117
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.16666666666666666,
95
+ "acc_stderr": 0.03708284662416544,
96
+ "acc_norm": 0.16666666666666666,
97
+ "acc_norm_stderr": 0.03708284662416544
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.37815126050420167,
101
+ "acc_stderr": 0.031499305777849054,
102
+ "acc_norm": 0.37815126050420167,
103
+ "acc_norm_stderr": 0.031499305777849054
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.37435897435897436,
107
+ "acc_stderr": 0.024537591572830517,
108
+ "acc_norm": 0.37435897435897436,
109
+ "acc_norm_stderr": 0.024537591572830517
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.56,
113
+ "acc_stderr": 0.04988876515698589,
114
+ "acc_norm": 0.56,
115
+ "acc_norm_stderr": 0.04988876515698589
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.31,
119
+ "acc_stderr": 0.04648231987117316,
120
+ "acc_norm": 0.31,
121
+ "acc_norm_stderr": 0.04648231987117316
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5185185185185185,
125
+ "acc_stderr": 0.04830366024635331,
126
+ "acc_norm": 0.5185185185185185,
127
+ "acc_norm_stderr": 0.04830366024635331
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4236453201970443,
131
+ "acc_stderr": 0.034767257476490364,
132
+ "acc_norm": 0.4236453201970443,
133
+ "acc_norm_stderr": 0.034767257476490364
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.41935483870967744,
137
+ "acc_stderr": 0.02807158890109185,
138
+ "acc_norm": 0.41935483870967744,
139
+ "acc_norm_stderr": 0.02807158890109185
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6367521367521367,
143
+ "acc_stderr": 0.03150712523091265,
144
+ "acc_norm": 0.6367521367521367,
145
+ "acc_norm_stderr": 0.03150712523091265
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.41509433962264153,
149
+ "acc_stderr": 0.03032594578928611,
150
+ "acc_norm": 0.41509433962264153,
151
+ "acc_norm_stderr": 0.03032594578928611
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4818181818181818,
155
+ "acc_stderr": 0.04785964010794915,
156
+ "acc_norm": 0.4818181818181818,
157
+ "acc_norm_stderr": 0.04785964010794915
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.28888888888888886,
161
+ "acc_stderr": 0.027634907264178544,
162
+ "acc_norm": 0.28888888888888886,
163
+ "acc_norm_stderr": 0.027634907264178544
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.32450331125827814,
167
+ "acc_stderr": 0.038227469376587525,
168
+ "acc_norm": 0.32450331125827814,
169
+ "acc_norm_stderr": 0.038227469376587525
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5124378109452736,
173
+ "acc_stderr": 0.0353443984853958,
174
+ "acc_norm": 0.5124378109452736,
175
+ "acc_norm_stderr": 0.0353443984853958
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3815028901734104,
179
+ "acc_stderr": 0.03703851193099521,
180
+ "acc_norm": 0.3815028901734104,
181
+ "acc_norm_stderr": 0.03703851193099521
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.30952380952380953,
185
+ "acc_stderr": 0.023809523809523857,
186
+ "acc_norm": 0.30952380952380953,
187
+ "acc_norm_stderr": 0.023809523809523857
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3125,
191
+ "acc_stderr": 0.038760854559127644,
192
+ "acc_norm": 0.3125,
193
+ "acc_norm_stderr": 0.038760854559127644
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.27,
197
+ "acc_stderr": 0.044619604333847394,
198
+ "acc_norm": 0.27,
199
+ "acc_norm_stderr": 0.044619604333847394
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.59,
203
+ "acc_stderr": 0.04943110704237101,
204
+ "acc_norm": 0.59,
205
+ "acc_norm_stderr": 0.04943110704237101
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5,
209
+ "acc_stderr": 0.026919095102908273,
210
+ "acc_norm": 0.5,
211
+ "acc_norm_stderr": 0.026919095102908273
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4601226993865031,
215
+ "acc_stderr": 0.03915857291436971,
216
+ "acc_norm": 0.4601226993865031,
217
+ "acc_norm_stderr": 0.03915857291436971
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4444444444444444,
221
+ "acc_stderr": 0.027648477877413317,
222
+ "acc_norm": 0.4444444444444444,
223
+ "acc_norm_stderr": 0.027648477877413317
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.32,
227
+ "acc_stderr": 0.04688261722621504,
228
+ "acc_norm": 0.32,
229
+ "acc_norm_stderr": 0.04688261722621504
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5233160621761658,
233
+ "acc_stderr": 0.036045136724422014,
234
+ "acc_norm": 0.5233160621761658,
235
+ "acc_norm_stderr": 0.036045136724422014
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2894736842105263,
239
+ "acc_stderr": 0.04266339443159394,
240
+ "acc_norm": 0.2894736842105263,
241
+ "acc_norm_stderr": 0.04266339443159394
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.48990825688073397,
245
+ "acc_stderr": 0.021432956203453316,
246
+ "acc_norm": 0.48990825688073397,
247
+ "acc_norm_stderr": 0.021432956203453316
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.20634920634920634,
251
+ "acc_stderr": 0.0361960452412425,
252
+ "acc_norm": 0.20634920634920634,
253
+ "acc_norm_stderr": 0.0361960452412425
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4019607843137255,
257
+ "acc_stderr": 0.02807415894760066,
258
+ "acc_norm": 0.4019607843137255,
259
+ "acc_norm_stderr": 0.02807415894760066
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.45,
263
+ "acc_stderr": 0.05,
264
+ "acc_norm": 0.45,
265
+ "acc_norm_stderr": 0.05
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6115702479338843,
269
+ "acc_stderr": 0.04449270350068383,
270
+ "acc_norm": 0.6115702479338843,
271
+ "acc_norm_stderr": 0.04449270350068383
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3684210526315789,
275
+ "acc_stderr": 0.03925523381052932,
276
+ "acc_norm": 0.3684210526315789,
277
+ "acc_norm_stderr": 0.03925523381052932
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.4035947712418301,
281
+ "acc_stderr": 0.01984828016840117,
282
+ "acc_norm": 0.4035947712418301,
283
+ "acc_norm_stderr": 0.01984828016840117
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.2730496453900709,
287
+ "acc_stderr": 0.026577860943307854,
288
+ "acc_norm": 0.2730496453900709,
289
+ "acc_norm_stderr": 0.026577860943307854
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.26785714285714285,
293
+ "acc_stderr": 0.04203277291467762,
294
+ "acc_norm": 0.26785714285714285,
295
+ "acc_norm_stderr": 0.04203277291467762
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.24074074074074073,
299
+ "acc_stderr": 0.029157522184605593,
300
+ "acc_norm": 0.24074074074074073,
301
+ "acc_norm_stderr": 0.029157522184605593
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2424581005586592,
305
+ "acc_stderr": 0.01433352205921789,
306
+ "acc_norm": 0.2424581005586592,
307
+ "acc_norm_stderr": 0.01433352205921789
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.36,
311
+ "acc_stderr": 0.04824181513244218,
312
+ "acc_norm": 0.36,
313
+ "acc_norm_stderr": 0.04824181513244218
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.48,
317
+ "acc_stderr": 0.050211673156867795,
318
+ "acc_norm": 0.48,
319
+ "acc_norm_stderr": 0.050211673156867795
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3088235294117647,
323
+ "acc_stderr": 0.02806499816704009,
324
+ "acc_norm": 0.3088235294117647,
325
+ "acc_norm_stderr": 0.02806499816704009
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.3836734693877551,
329
+ "acc_stderr": 0.031130880396235922,
330
+ "acc_norm": 0.3836734693877551,
331
+ "acc_norm_stderr": 0.031130880396235922
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5654008438818565,
335
+ "acc_stderr": 0.03226759995510145,
336
+ "acc_norm": 0.5654008438818565,
337
+ "acc_norm_stderr": 0.03226759995510145
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.318122555410691,
341
+ "acc_stderr": 0.011895407281104074,
342
+ "acc_norm": 0.318122555410691,
343
+ "acc_norm_stderr": 0.011895407281104074
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4411764705882353,
347
+ "acc_stderr": 0.03484941514429231,
348
+ "acc_norm": 0.4411764705882353,
349
+ "acc_norm_stderr": 0.03484941514429231
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.6181818181818182,
353
+ "acc_stderr": 0.03793713171165634,
354
+ "acc_norm": 0.6181818181818182,
355
+ "acc_norm_stderr": 0.03793713171165634
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.3108935128518972,
359
+ "mc1_stderr": 0.016203316673559693,
360
+ "mc2": 0.474366186048088,
361
+ "mc2_stderr": 0.01540967506791855
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5041322314049587,
365
+ "acc_stderr": 0.017189767032130817,
366
+ "acc_norm": 0.5525383707201889,
367
+ "acc_norm_stderr": 0.017095190301500574
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-dpo-13b-v1",
436
+ "model_sha": "13d027c0a2069284308f4992d67a202ac2e50b22",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-dpo-13b-v2/result_2023-11-19 10:36:38.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.39505119453924914,
5
+ "acc_stderr": 0.014285898292938165,
6
+ "acc_norm": 0.46075085324232085,
7
+ "acc_norm_stderr": 0.014566303676636588
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.42640908185620396,
11
+ "acc_stderr": 0.004935439955031694,
12
+ "acc_norm": 0.5706034654451304,
13
+ "acc_norm_stderr": 0.0049397843114489855
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.47953216374269003,
17
+ "acc_stderr": 0.0383161053282193,
18
+ "acc_norm": 0.47953216374269003,
19
+ "acc_norm_stderr": 0.0383161053282193
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.44660194174757284,
23
+ "acc_stderr": 0.04922424153458934,
24
+ "acc_norm": 0.44660194174757284,
25
+ "acc_norm_stderr": 0.04922424153458934
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5197956577266922,
29
+ "acc_stderr": 0.017865944827291633,
30
+ "acc_norm": 0.5197956577266922,
31
+ "acc_norm_stderr": 0.017865944827291633
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4888888888888889,
35
+ "acc_stderr": 0.04318275491977976,
36
+ "acc_norm": 0.4888888888888889,
37
+ "acc_norm_stderr": 0.04318275491977976
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.3,
41
+ "acc_stderr": 0.046056618647183814,
42
+ "acc_norm": 0.3,
43
+ "acc_norm_stderr": 0.046056618647183814
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.4,
47
+ "acc_stderr": 0.03202563076101735,
48
+ "acc_norm": 0.4,
49
+ "acc_norm_stderr": 0.03202563076101735
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.4036144578313253,
53
+ "acc_stderr": 0.03819486140758398,
54
+ "acc_norm": 0.4036144578313253,
55
+ "acc_norm_stderr": 0.03819486140758398
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4533762057877814,
59
+ "acc_stderr": 0.028274359854894245,
60
+ "acc_norm": 0.4533762057877814,
61
+ "acc_norm_stderr": 0.028274359854894245
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.47533632286995514,
65
+ "acc_stderr": 0.03351695167652628,
66
+ "acc_norm": 0.47533632286995514,
67
+ "acc_norm_stderr": 0.03351695167652628
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.42748091603053434,
71
+ "acc_stderr": 0.043389203057924,
72
+ "acc_norm": 0.42748091603053434,
73
+ "acc_norm_stderr": 0.043389203057924
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.42,
77
+ "acc_stderr": 0.049604496374885836,
78
+ "acc_norm": 0.42,
79
+ "acc_norm_stderr": 0.049604496374885836
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5151515151515151,
83
+ "acc_stderr": 0.0356071651653106,
84
+ "acc_norm": 0.5151515151515151,
85
+ "acc_norm_stderr": 0.0356071651653106
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.45517241379310347,
89
+ "acc_stderr": 0.04149886942192117,
90
+ "acc_norm": 0.45517241379310347,
91
+ "acc_norm_stderr": 0.04149886942192117
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.16666666666666666,
95
+ "acc_stderr": 0.03708284662416544,
96
+ "acc_norm": 0.16666666666666666,
97
+ "acc_norm_stderr": 0.03708284662416544
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.37815126050420167,
101
+ "acc_stderr": 0.031499305777849054,
102
+ "acc_norm": 0.37815126050420167,
103
+ "acc_norm_stderr": 0.031499305777849054
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.3717948717948718,
107
+ "acc_stderr": 0.024503472557110946,
108
+ "acc_norm": 0.3717948717948718,
109
+ "acc_norm_stderr": 0.024503472557110946
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.56,
113
+ "acc_stderr": 0.04988876515698589,
114
+ "acc_norm": 0.56,
115
+ "acc_norm_stderr": 0.04988876515698589
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.31,
119
+ "acc_stderr": 0.04648231987117316,
120
+ "acc_norm": 0.31,
121
+ "acc_norm_stderr": 0.04648231987117316
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5185185185185185,
125
+ "acc_stderr": 0.04830366024635331,
126
+ "acc_norm": 0.5185185185185185,
127
+ "acc_norm_stderr": 0.04830366024635331
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4236453201970443,
131
+ "acc_stderr": 0.034767257476490364,
132
+ "acc_norm": 0.4236453201970443,
133
+ "acc_norm_stderr": 0.034767257476490364
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.41935483870967744,
137
+ "acc_stderr": 0.02807158890109185,
138
+ "acc_norm": 0.41935483870967744,
139
+ "acc_norm_stderr": 0.02807158890109185
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6324786324786325,
143
+ "acc_stderr": 0.03158539157745636,
144
+ "acc_norm": 0.6324786324786325,
145
+ "acc_norm_stderr": 0.03158539157745636
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.41509433962264153,
149
+ "acc_stderr": 0.03032594578928611,
150
+ "acc_norm": 0.41509433962264153,
151
+ "acc_norm_stderr": 0.03032594578928611
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4909090909090909,
155
+ "acc_stderr": 0.04788339768702861,
156
+ "acc_norm": 0.4909090909090909,
157
+ "acc_norm_stderr": 0.04788339768702861
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.29259259259259257,
161
+ "acc_stderr": 0.02773896963217609,
162
+ "acc_norm": 0.29259259259259257,
163
+ "acc_norm_stderr": 0.02773896963217609
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.32450331125827814,
167
+ "acc_stderr": 0.038227469376587525,
168
+ "acc_norm": 0.32450331125827814,
169
+ "acc_norm_stderr": 0.038227469376587525
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5124378109452736,
173
+ "acc_stderr": 0.0353443984853958,
174
+ "acc_norm": 0.5124378109452736,
175
+ "acc_norm_stderr": 0.0353443984853958
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3815028901734104,
179
+ "acc_stderr": 0.03703851193099521,
180
+ "acc_norm": 0.3815028901734104,
181
+ "acc_norm_stderr": 0.03703851193099521
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.30952380952380953,
185
+ "acc_stderr": 0.023809523809523857,
186
+ "acc_norm": 0.30952380952380953,
187
+ "acc_norm_stderr": 0.023809523809523857
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3194444444444444,
191
+ "acc_stderr": 0.03899073687357335,
192
+ "acc_norm": 0.3194444444444444,
193
+ "acc_norm_stderr": 0.03899073687357335
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.26,
197
+ "acc_stderr": 0.044084400227680794,
198
+ "acc_norm": 0.26,
199
+ "acc_norm_stderr": 0.044084400227680794
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.59,
203
+ "acc_stderr": 0.04943110704237101,
204
+ "acc_norm": 0.59,
205
+ "acc_norm_stderr": 0.04943110704237101
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5028901734104047,
209
+ "acc_stderr": 0.026918645383239015,
210
+ "acc_norm": 0.5028901734104047,
211
+ "acc_norm_stderr": 0.026918645383239015
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4601226993865031,
215
+ "acc_stderr": 0.03915857291436971,
216
+ "acc_norm": 0.4601226993865031,
217
+ "acc_norm_stderr": 0.03915857291436971
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.44753086419753085,
221
+ "acc_stderr": 0.02766713856942271,
222
+ "acc_norm": 0.44753086419753085,
223
+ "acc_norm_stderr": 0.02766713856942271
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.31,
227
+ "acc_stderr": 0.04648231987117316,
228
+ "acc_norm": 0.31,
229
+ "acc_norm_stderr": 0.04648231987117316
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5233160621761658,
233
+ "acc_stderr": 0.036045136724422014,
234
+ "acc_norm": 0.5233160621761658,
235
+ "acc_norm_stderr": 0.036045136724422014
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2894736842105263,
239
+ "acc_stderr": 0.04266339443159394,
240
+ "acc_norm": 0.2894736842105263,
241
+ "acc_norm_stderr": 0.04266339443159394
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.4917431192660551,
245
+ "acc_stderr": 0.021434399918214334,
246
+ "acc_norm": 0.4917431192660551,
247
+ "acc_norm_stderr": 0.021434399918214334
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.20634920634920634,
251
+ "acc_stderr": 0.0361960452412425,
252
+ "acc_norm": 0.20634920634920634,
253
+ "acc_norm_stderr": 0.0361960452412425
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.39869281045751637,
257
+ "acc_stderr": 0.02803609227389177,
258
+ "acc_norm": 0.39869281045751637,
259
+ "acc_norm_stderr": 0.02803609227389177
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.45,
263
+ "acc_stderr": 0.05,
264
+ "acc_norm": 0.45,
265
+ "acc_norm_stderr": 0.05
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6115702479338843,
269
+ "acc_stderr": 0.04449270350068383,
270
+ "acc_norm": 0.6115702479338843,
271
+ "acc_norm_stderr": 0.04449270350068383
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3684210526315789,
275
+ "acc_stderr": 0.03925523381052932,
276
+ "acc_norm": 0.3684210526315789,
277
+ "acc_norm_stderr": 0.03925523381052932
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.4035947712418301,
281
+ "acc_stderr": 0.01984828016840117,
282
+ "acc_norm": 0.4035947712418301,
283
+ "acc_norm_stderr": 0.01984828016840117
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.2730496453900709,
287
+ "acc_stderr": 0.026577860943307854,
288
+ "acc_norm": 0.2730496453900709,
289
+ "acc_norm_stderr": 0.026577860943307854
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.26785714285714285,
293
+ "acc_stderr": 0.04203277291467762,
294
+ "acc_norm": 0.26785714285714285,
295
+ "acc_norm_stderr": 0.04203277291467762
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.24074074074074073,
299
+ "acc_stderr": 0.029157522184605593,
300
+ "acc_norm": 0.24074074074074073,
301
+ "acc_norm_stderr": 0.029157522184605593
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2424581005586592,
305
+ "acc_stderr": 0.01433352205921789,
306
+ "acc_norm": 0.2424581005586592,
307
+ "acc_norm_stderr": 0.01433352205921789
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.36,
311
+ "acc_stderr": 0.04824181513244218,
312
+ "acc_norm": 0.36,
313
+ "acc_norm_stderr": 0.04824181513244218
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.48,
317
+ "acc_stderr": 0.050211673156867795,
318
+ "acc_norm": 0.48,
319
+ "acc_norm_stderr": 0.050211673156867795
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3125,
323
+ "acc_stderr": 0.02815637344037142,
324
+ "acc_norm": 0.3125,
325
+ "acc_norm_stderr": 0.02815637344037142
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.3836734693877551,
329
+ "acc_stderr": 0.031130880396235922,
330
+ "acc_norm": 0.3836734693877551,
331
+ "acc_norm_stderr": 0.031130880396235922
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.569620253164557,
335
+ "acc_stderr": 0.03223017195937598,
336
+ "acc_norm": 0.569620253164557,
337
+ "acc_norm_stderr": 0.03223017195937598
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.31747066492829207,
341
+ "acc_stderr": 0.011888892068809309,
342
+ "acc_norm": 0.31747066492829207,
343
+ "acc_norm_stderr": 0.011888892068809309
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4411764705882353,
347
+ "acc_stderr": 0.034849415144292316,
348
+ "acc_norm": 0.4411764705882353,
349
+ "acc_norm_stderr": 0.034849415144292316
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.6121212121212121,
353
+ "acc_stderr": 0.0380491365397101,
354
+ "acc_norm": 0.6121212121212121,
355
+ "acc_norm_stderr": 0.0380491365397101
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.3084455324357405,
359
+ "mc1_stderr": 0.01616803938315687,
360
+ "mc2": 0.47439440606323957,
361
+ "mc2_stderr": 0.015414552807155835
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5088547815820543,
365
+ "acc_stderr": 0.01718765819933674,
366
+ "acc_norm": 0.5548996458087367,
367
+ "acc_norm_stderr": 0.017086417431005464
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-dpo-13b-v2",
436
+ "model_sha": "6fd9c176286458a9e802d0955a243f7b538c8e1c",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-instruct-20B-v1/result_2023-11-05 16:07:53.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3984641638225256,
5
+ "acc_stderr": 0.014306946052735563,
6
+ "acc_norm": 0.4616040955631399,
7
+ "acc_norm_stderr": 0.01456824555029636
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.42162915753833896,
11
+ "acc_stderr": 0.004928105880776079,
12
+ "acc_norm": 0.5677155945030871,
13
+ "acc_norm_stderr": 0.004943809330692697
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5555555555555556,
17
+ "acc_stderr": 0.038110796698335316,
18
+ "acc_norm": 0.5555555555555556,
19
+ "acc_norm_stderr": 0.038110796698335316
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5436893203883495,
23
+ "acc_stderr": 0.049318019942204146,
24
+ "acc_norm": 0.5436893203883495,
25
+ "acc_norm_stderr": 0.049318019942204146
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5491698595146871,
29
+ "acc_stderr": 0.01779329757269903,
30
+ "acc_norm": 0.5491698595146871,
31
+ "acc_norm_stderr": 0.01779329757269903
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.5037037037037037,
35
+ "acc_stderr": 0.043192236258113324,
36
+ "acc_norm": 0.5037037037037037,
37
+ "acc_norm_stderr": 0.043192236258113324
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.28,
41
+ "acc_stderr": 0.045126085985421255,
42
+ "acc_norm": 0.28,
43
+ "acc_norm_stderr": 0.045126085985421255
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.3574468085106383,
47
+ "acc_stderr": 0.03132941789476425,
48
+ "acc_norm": 0.3574468085106383,
49
+ "acc_norm_stderr": 0.03132941789476425
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3855421686746988,
53
+ "acc_stderr": 0.03789134424611548,
54
+ "acc_norm": 0.3855421686746988,
55
+ "acc_norm_stderr": 0.03789134424611548
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.48231511254019294,
59
+ "acc_stderr": 0.02838032284907713,
60
+ "acc_norm": 0.48231511254019294,
61
+ "acc_norm_stderr": 0.02838032284907713
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.47533632286995514,
65
+ "acc_stderr": 0.03351695167652628,
66
+ "acc_norm": 0.47533632286995514,
67
+ "acc_norm_stderr": 0.03351695167652628
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.45038167938931295,
71
+ "acc_stderr": 0.04363643698524779,
72
+ "acc_norm": 0.45038167938931295,
73
+ "acc_norm_stderr": 0.04363643698524779
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.42,
77
+ "acc_stderr": 0.049604496374885836,
78
+ "acc_norm": 0.42,
79
+ "acc_norm_stderr": 0.049604496374885836
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5555555555555556,
83
+ "acc_stderr": 0.035402943770953675,
84
+ "acc_norm": 0.5555555555555556,
85
+ "acc_norm_stderr": 0.035402943770953675
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.3793103448275862,
89
+ "acc_stderr": 0.04043461861916748,
90
+ "acc_norm": 0.3793103448275862,
91
+ "acc_norm_stderr": 0.04043461861916748
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.17647058823529413,
95
+ "acc_stderr": 0.03793281185307811,
96
+ "acc_norm": 0.17647058823529413,
97
+ "acc_norm_stderr": 0.03793281185307811
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.4369747899159664,
101
+ "acc_stderr": 0.03221943636566196,
102
+ "acc_norm": 0.4369747899159664,
103
+ "acc_norm_stderr": 0.03221943636566196
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4205128205128205,
107
+ "acc_stderr": 0.025028610276710855,
108
+ "acc_norm": 0.4205128205128205,
109
+ "acc_norm_stderr": 0.025028610276710855
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.52,
113
+ "acc_stderr": 0.050211673156867795,
114
+ "acc_norm": 0.52,
115
+ "acc_norm_stderr": 0.050211673156867795
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.34,
119
+ "acc_stderr": 0.04760952285695235,
120
+ "acc_norm": 0.34,
121
+ "acc_norm_stderr": 0.04760952285695235
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.48148148148148145,
125
+ "acc_stderr": 0.04830366024635331,
126
+ "acc_norm": 0.48148148148148145,
127
+ "acc_norm_stderr": 0.04830366024635331
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.35960591133004927,
131
+ "acc_stderr": 0.033764582465095665,
132
+ "acc_norm": 0.35960591133004927,
133
+ "acc_norm_stderr": 0.033764582465095665
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.42258064516129035,
137
+ "acc_stderr": 0.02810096472427264,
138
+ "acc_norm": 0.42258064516129035,
139
+ "acc_norm_stderr": 0.02810096472427264
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.5982905982905983,
143
+ "acc_stderr": 0.03211693751051622,
144
+ "acc_norm": 0.5982905982905983,
145
+ "acc_norm_stderr": 0.03211693751051622
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4188679245283019,
149
+ "acc_stderr": 0.030365050829115208,
150
+ "acc_norm": 0.4188679245283019,
151
+ "acc_norm_stderr": 0.030365050829115208
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4727272727272727,
155
+ "acc_stderr": 0.04782001791380063,
156
+ "acc_norm": 0.4727272727272727,
157
+ "acc_norm_stderr": 0.04782001791380063
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.26296296296296295,
161
+ "acc_stderr": 0.02684205787383371,
162
+ "acc_norm": 0.26296296296296295,
163
+ "acc_norm_stderr": 0.02684205787383371
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.33112582781456956,
167
+ "acc_stderr": 0.038425817186598696,
168
+ "acc_norm": 0.33112582781456956,
169
+ "acc_norm_stderr": 0.038425817186598696
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5024875621890548,
173
+ "acc_stderr": 0.03535490150137289,
174
+ "acc_norm": 0.5024875621890548,
175
+ "acc_norm_stderr": 0.03535490150137289
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.4508670520231214,
179
+ "acc_stderr": 0.037940126746970296,
180
+ "acc_norm": 0.4508670520231214,
181
+ "acc_norm_stderr": 0.037940126746970296
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.291005291005291,
185
+ "acc_stderr": 0.023393826500484875,
186
+ "acc_norm": 0.291005291005291,
187
+ "acc_norm_stderr": 0.023393826500484875
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3541666666666667,
191
+ "acc_stderr": 0.039994111357535424,
192
+ "acc_norm": 0.3541666666666667,
193
+ "acc_norm_stderr": 0.039994111357535424
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.22,
197
+ "acc_stderr": 0.04163331998932269,
198
+ "acc_norm": 0.22,
199
+ "acc_norm_stderr": 0.04163331998932269
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.62,
203
+ "acc_stderr": 0.048783173121456344,
204
+ "acc_norm": 0.62,
205
+ "acc_norm_stderr": 0.048783173121456344
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5,
209
+ "acc_stderr": 0.026919095102908273,
210
+ "acc_norm": 0.5,
211
+ "acc_norm_stderr": 0.026919095102908273
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.5153374233128835,
215
+ "acc_stderr": 0.03926522378708843,
216
+ "acc_norm": 0.5153374233128835,
217
+ "acc_norm_stderr": 0.03926522378708843
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.5,
221
+ "acc_stderr": 0.02782074420373286,
222
+ "acc_norm": 0.5,
223
+ "acc_norm_stderr": 0.02782074420373286
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.34,
227
+ "acc_stderr": 0.04760952285695235,
228
+ "acc_norm": 0.34,
229
+ "acc_norm_stderr": 0.04760952285695235
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5181347150259067,
233
+ "acc_stderr": 0.03606065001832919,
234
+ "acc_norm": 0.5181347150259067,
235
+ "acc_norm_stderr": 0.03606065001832919
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2894736842105263,
239
+ "acc_stderr": 0.04266339443159394,
240
+ "acc_norm": 0.2894736842105263,
241
+ "acc_norm_stderr": 0.04266339443159394
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5431192660550459,
245
+ "acc_stderr": 0.021357458785226224,
246
+ "acc_norm": 0.5431192660550459,
247
+ "acc_norm_stderr": 0.021357458785226224
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.30158730158730157,
251
+ "acc_stderr": 0.04104947269903394,
252
+ "acc_norm": 0.30158730158730157,
253
+ "acc_norm_stderr": 0.04104947269903394
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.3954248366013072,
257
+ "acc_stderr": 0.02799672318063145,
258
+ "acc_norm": 0.3954248366013072,
259
+ "acc_norm_stderr": 0.02799672318063145
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.44,
263
+ "acc_stderr": 0.04988876515698589,
264
+ "acc_norm": 0.44,
265
+ "acc_norm_stderr": 0.04988876515698589
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6198347107438017,
269
+ "acc_stderr": 0.04431324501968431,
270
+ "acc_norm": 0.6198347107438017,
271
+ "acc_norm_stderr": 0.04431324501968431
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.42105263157894735,
275
+ "acc_stderr": 0.04017901275981748,
276
+ "acc_norm": 0.42105263157894735,
277
+ "acc_norm_stderr": 0.04017901275981748
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3937908496732026,
281
+ "acc_stderr": 0.01976621199107307,
282
+ "acc_norm": 0.3937908496732026,
283
+ "acc_norm_stderr": 0.01976621199107307
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3404255319148936,
287
+ "acc_stderr": 0.02826765748265015,
288
+ "acc_norm": 0.3404255319148936,
289
+ "acc_norm_stderr": 0.02826765748265015
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.2767857142857143,
293
+ "acc_stderr": 0.04246624336697624,
294
+ "acc_norm": 0.2767857142857143,
295
+ "acc_norm_stderr": 0.04246624336697624
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.41203703703703703,
299
+ "acc_stderr": 0.03356787758160835,
300
+ "acc_norm": 0.41203703703703703,
301
+ "acc_norm_stderr": 0.03356787758160835
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.29497206703910617,
305
+ "acc_stderr": 0.015251931579208185,
306
+ "acc_norm": 0.29497206703910617,
307
+ "acc_norm_stderr": 0.015251931579208185
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.37,
311
+ "acc_stderr": 0.048523658709391,
312
+ "acc_norm": 0.37,
313
+ "acc_norm_stderr": 0.048523658709391
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.44,
317
+ "acc_stderr": 0.04988876515698589,
318
+ "acc_norm": 0.44,
319
+ "acc_norm_stderr": 0.04988876515698589
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3860294117647059,
323
+ "acc_stderr": 0.029573269134411124,
324
+ "acc_norm": 0.3860294117647059,
325
+ "acc_norm_stderr": 0.029573269134411124
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.46938775510204084,
329
+ "acc_stderr": 0.031949171367580624,
330
+ "acc_norm": 0.46938775510204084,
331
+ "acc_norm_stderr": 0.031949171367580624
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6919831223628692,
335
+ "acc_stderr": 0.030052389335605695,
336
+ "acc_norm": 0.6919831223628692,
337
+ "acc_norm_stderr": 0.030052389335605695
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3683181225554107,
341
+ "acc_stderr": 0.012319403369564639,
342
+ "acc_norm": 0.3683181225554107,
343
+ "acc_norm_stderr": 0.012319403369564639
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.5245098039215687,
347
+ "acc_stderr": 0.03505093194348798,
348
+ "acc_norm": 0.5245098039215687,
349
+ "acc_norm_stderr": 0.03505093194348798
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5757575757575758,
353
+ "acc_stderr": 0.03859268142070262,
354
+ "acc_norm": 0.5757575757575758,
355
+ "acc_norm_stderr": 0.03859268142070262
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2962056303549572,
359
+ "mc1_stderr": 0.015983595101811392,
360
+ "mc2": 0.458694749783158,
361
+ "mc2_stderr": 0.015135220490705375
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.45336481700118064,
365
+ "acc_stderr": 0.017115418225226862,
366
+ "acc_norm": 0.564344746162928,
367
+ "acc_norm_stderr": 0.017047415229476313
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-instruct-20B-v1",
436
+ "model_sha": "4de05113ecc02aa2da28893d8e2827912ebe0d20",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-instruct-v2-13b/result_2023-10-18 01:40:38.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3848122866894198,
5
+ "acc_stderr": 0.014218371065251095,
6
+ "acc_norm": 0.4402730375426621,
7
+ "acc_norm_stderr": 0.014506769524804243
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.4190400318661621,
11
+ "acc_stderr": 0.0049239357498424945,
12
+ "acc_norm": 0.5560645289782912,
13
+ "acc_norm_stderr": 0.004958314114266494
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5614035087719298,
17
+ "acc_stderr": 0.038057975055904594,
18
+ "acc_norm": 0.5614035087719298,
19
+ "acc_norm_stderr": 0.038057975055904594
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.6019417475728155,
23
+ "acc_stderr": 0.04846748253977238,
24
+ "acc_norm": 0.6019417475728155,
25
+ "acc_norm_stderr": 0.04846748253977238
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5287356321839081,
29
+ "acc_stderr": 0.017850410794380173,
30
+ "acc_norm": 0.5287356321839081,
31
+ "acc_norm_stderr": 0.017850410794380173
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4222222222222222,
35
+ "acc_stderr": 0.04266763404099582,
36
+ "acc_norm": 0.4222222222222222,
37
+ "acc_norm_stderr": 0.04266763404099582
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.28,
41
+ "acc_stderr": 0.045126085985421276,
42
+ "acc_norm": 0.28,
43
+ "acc_norm_stderr": 0.045126085985421276
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.3404255319148936,
47
+ "acc_stderr": 0.030976692998534443,
48
+ "acc_norm": 0.3404255319148936,
49
+ "acc_norm_stderr": 0.030976692998534443
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39759036144578314,
53
+ "acc_stderr": 0.038099730845402184,
54
+ "acc_norm": 0.39759036144578314,
55
+ "acc_norm_stderr": 0.038099730845402184
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5273311897106109,
59
+ "acc_stderr": 0.028355633568328188,
60
+ "acc_norm": 0.5273311897106109,
61
+ "acc_norm_stderr": 0.028355633568328188
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.4798206278026906,
65
+ "acc_stderr": 0.033530461674123,
66
+ "acc_norm": 0.4798206278026906,
67
+ "acc_norm_stderr": 0.033530461674123
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.5114503816793893,
71
+ "acc_stderr": 0.043841400240780176,
72
+ "acc_norm": 0.5114503816793893,
73
+ "acc_norm_stderr": 0.043841400240780176
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.39,
77
+ "acc_stderr": 0.04902071300001975,
78
+ "acc_norm": 0.39,
79
+ "acc_norm_stderr": 0.04902071300001975
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5353535353535354,
83
+ "acc_stderr": 0.03553436368828061,
84
+ "acc_norm": 0.5353535353535354,
85
+ "acc_norm_stderr": 0.03553436368828061
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4827586206896552,
89
+ "acc_stderr": 0.04164188720169377,
90
+ "acc_norm": 0.4827586206896552,
91
+ "acc_norm_stderr": 0.04164188720169377
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.23529411764705882,
95
+ "acc_stderr": 0.04220773659171452,
96
+ "acc_norm": 0.23529411764705882,
97
+ "acc_norm_stderr": 0.04220773659171452
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.49159663865546216,
101
+ "acc_stderr": 0.03247390276569669,
102
+ "acc_norm": 0.49159663865546216,
103
+ "acc_norm_stderr": 0.03247390276569669
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.46153846153846156,
107
+ "acc_stderr": 0.025275892070240634,
108
+ "acc_norm": 0.46153846153846156,
109
+ "acc_norm_stderr": 0.025275892070240634
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.45,
113
+ "acc_stderr": 0.05,
114
+ "acc_norm": 0.45,
115
+ "acc_norm_stderr": 0.05
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.37,
119
+ "acc_stderr": 0.048523658709391,
120
+ "acc_norm": 0.37,
121
+ "acc_norm_stderr": 0.048523658709391
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5277777777777778,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.5277777777777778,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3399014778325123,
131
+ "acc_stderr": 0.033327690684107895,
132
+ "acc_norm": 0.3399014778325123,
133
+ "acc_norm_stderr": 0.033327690684107895
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4806451612903226,
137
+ "acc_stderr": 0.0284226874043121,
138
+ "acc_norm": 0.4806451612903226,
139
+ "acc_norm_stderr": 0.0284226874043121
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6752136752136753,
143
+ "acc_stderr": 0.03067902276549883,
144
+ "acc_norm": 0.6752136752136753,
145
+ "acc_norm_stderr": 0.03067902276549883
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.47547169811320755,
149
+ "acc_stderr": 0.030735822206205615,
150
+ "acc_norm": 0.47547169811320755,
151
+ "acc_norm_stderr": 0.030735822206205615
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5454545454545454,
155
+ "acc_stderr": 0.04769300568972745,
156
+ "acc_norm": 0.5454545454545454,
157
+ "acc_norm_stderr": 0.04769300568972745
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2777777777777778,
161
+ "acc_stderr": 0.027309140588230172,
162
+ "acc_norm": 0.2777777777777778,
163
+ "acc_norm_stderr": 0.027309140588230172
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.33112582781456956,
167
+ "acc_stderr": 0.038425817186598696,
168
+ "acc_norm": 0.33112582781456956,
169
+ "acc_norm_stderr": 0.038425817186598696
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5671641791044776,
173
+ "acc_stderr": 0.03503490923673282,
174
+ "acc_norm": 0.5671641791044776,
175
+ "acc_norm_stderr": 0.03503490923673282
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.4393063583815029,
179
+ "acc_stderr": 0.037842719328874674,
180
+ "acc_norm": 0.4393063583815029,
181
+ "acc_norm_stderr": 0.037842719328874674
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.3333333333333333,
185
+ "acc_stderr": 0.0242785680243077,
186
+ "acc_norm": 0.3333333333333333,
187
+ "acc_norm_stderr": 0.0242785680243077
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4027777777777778,
191
+ "acc_stderr": 0.04101405519842425,
192
+ "acc_norm": 0.4027777777777778,
193
+ "acc_norm_stderr": 0.04101405519842425
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.34,
197
+ "acc_stderr": 0.047609522856952344,
198
+ "acc_norm": 0.34,
199
+ "acc_norm_stderr": 0.047609522856952344
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.63,
203
+ "acc_stderr": 0.04852365870939099,
204
+ "acc_norm": 0.63,
205
+ "acc_norm_stderr": 0.04852365870939099
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.49421965317919075,
209
+ "acc_stderr": 0.026917296179149116,
210
+ "acc_norm": 0.49421965317919075,
211
+ "acc_norm_stderr": 0.026917296179149116
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.48466257668711654,
215
+ "acc_stderr": 0.039265223787088445,
216
+ "acc_norm": 0.48466257668711654,
217
+ "acc_norm_stderr": 0.039265223787088445
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.49074074074074076,
221
+ "acc_stderr": 0.027815973433878014,
222
+ "acc_norm": 0.49074074074074076,
223
+ "acc_norm_stderr": 0.027815973433878014
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.3,
227
+ "acc_stderr": 0.04605661864718381,
228
+ "acc_norm": 0.3,
229
+ "acc_norm_stderr": 0.04605661864718381
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5440414507772021,
233
+ "acc_stderr": 0.03594413711272436,
234
+ "acc_norm": 0.5440414507772021,
235
+ "acc_norm_stderr": 0.03594413711272436
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2543859649122807,
239
+ "acc_stderr": 0.040969851398436716,
240
+ "acc_norm": 0.2543859649122807,
241
+ "acc_norm_stderr": 0.040969851398436716
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5559633027522936,
245
+ "acc_stderr": 0.021302621211654518,
246
+ "acc_norm": 0.5559633027522936,
247
+ "acc_norm_stderr": 0.021302621211654518
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.30158730158730157,
251
+ "acc_stderr": 0.04104947269903394,
252
+ "acc_norm": 0.30158730158730157,
253
+ "acc_norm_stderr": 0.04104947269903394
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4673202614379085,
257
+ "acc_stderr": 0.028568699752225875,
258
+ "acc_norm": 0.4673202614379085,
259
+ "acc_norm_stderr": 0.028568699752225875
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.51,
263
+ "acc_stderr": 0.05024183937956912,
264
+ "acc_norm": 0.51,
265
+ "acc_norm_stderr": 0.05024183937956912
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.7272727272727273,
269
+ "acc_stderr": 0.04065578140908705,
270
+ "acc_norm": 0.7272727272727273,
271
+ "acc_norm_stderr": 0.04065578140908705
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.47368421052631576,
275
+ "acc_stderr": 0.04063302731486671,
276
+ "acc_norm": 0.47368421052631576,
277
+ "acc_norm_stderr": 0.04063302731486671
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3741830065359477,
281
+ "acc_stderr": 0.019576953122088833,
282
+ "acc_norm": 0.3741830065359477,
283
+ "acc_norm_stderr": 0.019576953122088833
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.31560283687943264,
287
+ "acc_stderr": 0.027724989449509314,
288
+ "acc_norm": 0.31560283687943264,
289
+ "acc_norm_stderr": 0.027724989449509314
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.22321428571428573,
293
+ "acc_stderr": 0.039523019677025116,
294
+ "acc_norm": 0.22321428571428573,
295
+ "acc_norm_stderr": 0.039523019677025116
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.4074074074074074,
299
+ "acc_stderr": 0.033509916046960436,
300
+ "acc_norm": 0.4074074074074074,
301
+ "acc_norm_stderr": 0.033509916046960436
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.264804469273743,
305
+ "acc_stderr": 0.014756906483260664,
306
+ "acc_norm": 0.264804469273743,
307
+ "acc_norm_stderr": 0.014756906483260664
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.35,
311
+ "acc_stderr": 0.0479372485441102,
312
+ "acc_norm": 0.35,
313
+ "acc_norm_stderr": 0.0479372485441102
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.39,
317
+ "acc_stderr": 0.04902071300001975,
318
+ "acc_norm": 0.39,
319
+ "acc_norm_stderr": 0.04902071300001975
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.39705882352941174,
323
+ "acc_stderr": 0.029722152099280058,
324
+ "acc_norm": 0.39705882352941174,
325
+ "acc_norm_stderr": 0.029722152099280058
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.49387755102040815,
329
+ "acc_stderr": 0.032006820201639086,
330
+ "acc_norm": 0.49387755102040815,
331
+ "acc_norm_stderr": 0.032006820201639086
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5738396624472574,
335
+ "acc_stderr": 0.03219035703131774,
336
+ "acc_norm": 0.5738396624472574,
337
+ "acc_norm_stderr": 0.03219035703131774
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.35071707953063885,
341
+ "acc_stderr": 0.012187773370741518,
342
+ "acc_norm": 0.35071707953063885,
343
+ "acc_norm_stderr": 0.012187773370741518
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4950980392156863,
347
+ "acc_stderr": 0.03509143375606786,
348
+ "acc_norm": 0.4950980392156863,
349
+ "acc_norm_stderr": 0.03509143375606786
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5696969696969697,
353
+ "acc_stderr": 0.03866225962879077,
354
+ "acc_norm": 0.5696969696969697,
355
+ "acc_norm_stderr": 0.03866225962879077
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.27539779681762544,
359
+ "mc1_stderr": 0.015638135667775523,
360
+ "mc2": 0.44227632802507094,
361
+ "mc2_stderr": 0.015242459306682204
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5088547815820543,
365
+ "acc_stderr": 0.017187658199336743,
366
+ "acc_norm": 0.5608028335301063,
367
+ "acc_norm_stderr": 0.017062775744780705
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-instruct-v2-13b",
436
+ "model_sha": "9f429309fc6b939d08c659ab4666f6e80324dcd1",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-instruct-v3-13B/result_2023-11-01 18:54:40.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3967576791808874,
5
+ "acc_stderr": 0.014296513020180646,
6
+ "acc_norm": 0.454778156996587,
7
+ "acc_norm_stderr": 0.014551507060836355
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.4164509061939853,
11
+ "acc_stderr": 0.004919626380645517,
12
+ "acc_norm": 0.5536745668193587,
13
+ "acc_norm_stderr": 0.004960947388535101
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5380116959064327,
17
+ "acc_stderr": 0.038237270928823064,
18
+ "acc_norm": 0.5380116959064327,
19
+ "acc_norm_stderr": 0.038237270928823064
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.6019417475728155,
23
+ "acc_stderr": 0.048467482539772386,
24
+ "acc_norm": 0.6019417475728155,
25
+ "acc_norm_stderr": 0.048467482539772386
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5159642401021711,
29
+ "acc_stderr": 0.017870847506081738,
30
+ "acc_norm": 0.5159642401021711,
31
+ "acc_norm_stderr": 0.017870847506081738
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.42962962962962964,
35
+ "acc_stderr": 0.04276349494376599,
36
+ "acc_norm": 0.42962962962962964,
37
+ "acc_norm_stderr": 0.04276349494376599
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.32,
41
+ "acc_stderr": 0.046882617226215034,
42
+ "acc_norm": 0.32,
43
+ "acc_norm_stderr": 0.046882617226215034
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.37446808510638296,
47
+ "acc_stderr": 0.031639106653672915,
48
+ "acc_norm": 0.37446808510638296,
49
+ "acc_norm_stderr": 0.031639106653672915
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.45180722891566266,
53
+ "acc_stderr": 0.03874371556587953,
54
+ "acc_norm": 0.45180722891566266,
55
+ "acc_norm_stderr": 0.03874371556587953
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5627009646302251,
59
+ "acc_stderr": 0.0281739177617629,
60
+ "acc_norm": 0.5627009646302251,
61
+ "acc_norm_stderr": 0.0281739177617629
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.3811659192825112,
65
+ "acc_stderr": 0.032596251184168284,
66
+ "acc_norm": 0.3811659192825112,
67
+ "acc_norm_stderr": 0.032596251184168284
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.46564885496183206,
71
+ "acc_stderr": 0.043749285605997376,
72
+ "acc_norm": 0.46564885496183206,
73
+ "acc_norm_stderr": 0.043749285605997376
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.42,
77
+ "acc_stderr": 0.049604496374885836,
78
+ "acc_norm": 0.42,
79
+ "acc_norm_stderr": 0.049604496374885836
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5757575757575758,
83
+ "acc_stderr": 0.035212249088415866,
84
+ "acc_norm": 0.5757575757575758,
85
+ "acc_norm_stderr": 0.035212249088415866
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.45517241379310347,
89
+ "acc_stderr": 0.04149886942192117,
90
+ "acc_norm": 0.45517241379310347,
91
+ "acc_norm_stderr": 0.04149886942192117
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.18627450980392157,
95
+ "acc_stderr": 0.038739587141493524,
96
+ "acc_norm": 0.18627450980392157,
97
+ "acc_norm_stderr": 0.038739587141493524
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.4495798319327731,
101
+ "acc_stderr": 0.03231293497137707,
102
+ "acc_norm": 0.4495798319327731,
103
+ "acc_norm_stderr": 0.03231293497137707
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4641025641025641,
107
+ "acc_stderr": 0.025285585990017834,
108
+ "acc_norm": 0.4641025641025641,
109
+ "acc_norm_stderr": 0.025285585990017834
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.46,
113
+ "acc_stderr": 0.05009082659620333,
114
+ "acc_norm": 0.46,
115
+ "acc_norm_stderr": 0.05009082659620333
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.35,
119
+ "acc_stderr": 0.04793724854411019,
120
+ "acc_norm": 0.35,
121
+ "acc_norm_stderr": 0.04793724854411019
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5277777777777778,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.5277777777777778,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.35467980295566504,
131
+ "acc_stderr": 0.03366124489051449,
132
+ "acc_norm": 0.35467980295566504,
133
+ "acc_norm_stderr": 0.03366124489051449
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.5096774193548387,
137
+ "acc_stderr": 0.02843867799890955,
138
+ "acc_norm": 0.5096774193548387,
139
+ "acc_norm_stderr": 0.02843867799890955
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6837606837606838,
143
+ "acc_stderr": 0.030463656747340254,
144
+ "acc_norm": 0.6837606837606838,
145
+ "acc_norm_stderr": 0.030463656747340254
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4641509433962264,
149
+ "acc_stderr": 0.030693675018458003,
150
+ "acc_norm": 0.4641509433962264,
151
+ "acc_norm_stderr": 0.030693675018458003
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.44545454545454544,
155
+ "acc_stderr": 0.047605488214603246,
156
+ "acc_norm": 0.44545454545454544,
157
+ "acc_norm_stderr": 0.047605488214603246
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2777777777777778,
161
+ "acc_stderr": 0.02730914058823017,
162
+ "acc_norm": 0.2777777777777778,
163
+ "acc_norm_stderr": 0.02730914058823017
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.304635761589404,
167
+ "acc_stderr": 0.03757949922943343,
168
+ "acc_norm": 0.304635761589404,
169
+ "acc_norm_stderr": 0.03757949922943343
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5920398009950248,
173
+ "acc_stderr": 0.03475116365194092,
174
+ "acc_norm": 0.5920398009950248,
175
+ "acc_norm_stderr": 0.03475116365194092
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3988439306358382,
179
+ "acc_stderr": 0.03733626655383509,
180
+ "acc_norm": 0.3988439306358382,
181
+ "acc_norm_stderr": 0.03733626655383509
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2857142857142857,
185
+ "acc_stderr": 0.02326651221373056,
186
+ "acc_norm": 0.2857142857142857,
187
+ "acc_norm_stderr": 0.02326651221373056
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4027777777777778,
191
+ "acc_stderr": 0.04101405519842424,
192
+ "acc_norm": 0.4027777777777778,
193
+ "acc_norm_stderr": 0.04101405519842424
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.36,
197
+ "acc_stderr": 0.04824181513244218,
198
+ "acc_norm": 0.36,
199
+ "acc_norm_stderr": 0.04824181513244218
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.64,
203
+ "acc_stderr": 0.04824181513244218,
204
+ "acc_norm": 0.64,
205
+ "acc_norm_stderr": 0.04824181513244218
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5028901734104047,
209
+ "acc_stderr": 0.02691864538323901,
210
+ "acc_norm": 0.5028901734104047,
211
+ "acc_norm_stderr": 0.02691864538323901
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4723926380368098,
215
+ "acc_stderr": 0.03922378290610991,
216
+ "acc_norm": 0.4723926380368098,
217
+ "acc_norm_stderr": 0.03922378290610991
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4783950617283951,
221
+ "acc_stderr": 0.02779476010500874,
222
+ "acc_norm": 0.4783950617283951,
223
+ "acc_norm_stderr": 0.02779476010500874
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.28,
227
+ "acc_stderr": 0.045126085985421296,
228
+ "acc_norm": 0.28,
229
+ "acc_norm_stderr": 0.045126085985421296
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.538860103626943,
233
+ "acc_stderr": 0.035975244117345775,
234
+ "acc_norm": 0.538860103626943,
235
+ "acc_norm_stderr": 0.035975244117345775
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2807017543859649,
239
+ "acc_stderr": 0.042270544512321984,
240
+ "acc_norm": 0.2807017543859649,
241
+ "acc_norm_stderr": 0.042270544512321984
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5229357798165137,
245
+ "acc_stderr": 0.0214147570581755,
246
+ "acc_norm": 0.5229357798165137,
247
+ "acc_norm_stderr": 0.0214147570581755
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3412698412698413,
251
+ "acc_stderr": 0.04240799327574925,
252
+ "acc_norm": 0.3412698412698413,
253
+ "acc_norm_stderr": 0.04240799327574925
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.477124183006536,
257
+ "acc_stderr": 0.028599936776089775,
258
+ "acc_norm": 0.477124183006536,
259
+ "acc_norm_stderr": 0.028599936776089775
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.45,
263
+ "acc_stderr": 0.05,
264
+ "acc_norm": 0.45,
265
+ "acc_norm_stderr": 0.05
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6776859504132231,
269
+ "acc_stderr": 0.04266416363352167,
270
+ "acc_norm": 0.6776859504132231,
271
+ "acc_norm_stderr": 0.04266416363352167
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.46710526315789475,
275
+ "acc_stderr": 0.040601270352363966,
276
+ "acc_norm": 0.46710526315789475,
277
+ "acc_norm_stderr": 0.040601270352363966
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3660130718954248,
281
+ "acc_stderr": 0.019488025745529672,
282
+ "acc_norm": 0.3660130718954248,
283
+ "acc_norm_stderr": 0.019488025745529672
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.35106382978723405,
287
+ "acc_stderr": 0.028473501272963775,
288
+ "acc_norm": 0.35106382978723405,
289
+ "acc_norm_stderr": 0.028473501272963775
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25,
293
+ "acc_stderr": 0.04109974682633932,
294
+ "acc_norm": 0.25,
295
+ "acc_norm_stderr": 0.04109974682633932
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.375,
299
+ "acc_stderr": 0.033016908987210894,
300
+ "acc_norm": 0.375,
301
+ "acc_norm_stderr": 0.033016908987210894
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2547486033519553,
305
+ "acc_stderr": 0.014572650383409167,
306
+ "acc_norm": 0.2547486033519553,
307
+ "acc_norm_stderr": 0.014572650383409167
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.45,
311
+ "acc_stderr": 0.05,
312
+ "acc_norm": 0.45,
313
+ "acc_norm_stderr": 0.05
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.42,
317
+ "acc_stderr": 0.04960449637488584,
318
+ "acc_norm": 0.42,
319
+ "acc_norm_stderr": 0.04960449637488584
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.45955882352941174,
323
+ "acc_stderr": 0.03027332507734575,
324
+ "acc_norm": 0.45955882352941174,
325
+ "acc_norm_stderr": 0.03027332507734575
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4897959183673469,
329
+ "acc_stderr": 0.03200255347893782,
330
+ "acc_norm": 0.4897959183673469,
331
+ "acc_norm_stderr": 0.03200255347893782
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6033755274261603,
335
+ "acc_stderr": 0.03184399873811225,
336
+ "acc_norm": 0.6033755274261603,
337
+ "acc_norm_stderr": 0.03184399873811225
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3285528031290743,
341
+ "acc_stderr": 0.011996027247502912,
342
+ "acc_norm": 0.3285528031290743,
343
+ "acc_norm_stderr": 0.011996027247502912
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.49019607843137253,
347
+ "acc_stderr": 0.035086373586305716,
348
+ "acc_norm": 0.49019607843137253,
349
+ "acc_norm_stderr": 0.035086373586305716
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5575757575757576,
353
+ "acc_stderr": 0.03878372113711275,
354
+ "acc_norm": 0.5575757575757576,
355
+ "acc_norm_stderr": 0.03878372113711275
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.27906976744186046,
359
+ "mc1_stderr": 0.01570210709062789,
360
+ "mc2": 0.4486611820923937,
361
+ "mc2_stderr": 0.015416976946375454
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.48406139315230223,
365
+ "acc_stderr": 0.017181617837190192,
366
+ "acc_norm": 0.5619834710743802,
367
+ "acc_norm_stderr": 0.01705775370216029
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-instruct-v3-13b",
436
+ "model_sha": "199c2113f09f153bce1ad7aac35e6e756a99b89b",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-instruct-v3-13B/result_2023-11-02 01:15:35.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3967576791808874,
5
+ "acc_stderr": 0.014296513020180646,
6
+ "acc_norm": 0.454778156996587,
7
+ "acc_norm_stderr": 0.014551507060836355
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.4164509061939853,
11
+ "acc_stderr": 0.004919626380645517,
12
+ "acc_norm": 0.5536745668193587,
13
+ "acc_norm_stderr": 0.004960947388535101
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5380116959064327,
17
+ "acc_stderr": 0.038237270928823064,
18
+ "acc_norm": 0.5380116959064327,
19
+ "acc_norm_stderr": 0.038237270928823064
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.6019417475728155,
23
+ "acc_stderr": 0.048467482539772386,
24
+ "acc_norm": 0.6019417475728155,
25
+ "acc_norm_stderr": 0.048467482539772386
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5159642401021711,
29
+ "acc_stderr": 0.017870847506081738,
30
+ "acc_norm": 0.5159642401021711,
31
+ "acc_norm_stderr": 0.017870847506081738
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.42962962962962964,
35
+ "acc_stderr": 0.04276349494376599,
36
+ "acc_norm": 0.42962962962962964,
37
+ "acc_norm_stderr": 0.04276349494376599
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.32,
41
+ "acc_stderr": 0.046882617226215034,
42
+ "acc_norm": 0.32,
43
+ "acc_norm_stderr": 0.046882617226215034
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.37446808510638296,
47
+ "acc_stderr": 0.031639106653672915,
48
+ "acc_norm": 0.37446808510638296,
49
+ "acc_norm_stderr": 0.031639106653672915
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.45180722891566266,
53
+ "acc_stderr": 0.03874371556587953,
54
+ "acc_norm": 0.45180722891566266,
55
+ "acc_norm_stderr": 0.03874371556587953
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5627009646302251,
59
+ "acc_stderr": 0.0281739177617629,
60
+ "acc_norm": 0.5627009646302251,
61
+ "acc_norm_stderr": 0.0281739177617629
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.3811659192825112,
65
+ "acc_stderr": 0.032596251184168284,
66
+ "acc_norm": 0.3811659192825112,
67
+ "acc_norm_stderr": 0.032596251184168284
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.46564885496183206,
71
+ "acc_stderr": 0.043749285605997376,
72
+ "acc_norm": 0.46564885496183206,
73
+ "acc_norm_stderr": 0.043749285605997376
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.42,
77
+ "acc_stderr": 0.049604496374885836,
78
+ "acc_norm": 0.42,
79
+ "acc_norm_stderr": 0.049604496374885836
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5757575757575758,
83
+ "acc_stderr": 0.035212249088415866,
84
+ "acc_norm": 0.5757575757575758,
85
+ "acc_norm_stderr": 0.035212249088415866
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.45517241379310347,
89
+ "acc_stderr": 0.04149886942192117,
90
+ "acc_norm": 0.45517241379310347,
91
+ "acc_norm_stderr": 0.04149886942192117
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.18627450980392157,
95
+ "acc_stderr": 0.038739587141493524,
96
+ "acc_norm": 0.18627450980392157,
97
+ "acc_norm_stderr": 0.038739587141493524
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.4495798319327731,
101
+ "acc_stderr": 0.03231293497137707,
102
+ "acc_norm": 0.4495798319327731,
103
+ "acc_norm_stderr": 0.03231293497137707
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4641025641025641,
107
+ "acc_stderr": 0.025285585990017834,
108
+ "acc_norm": 0.4641025641025641,
109
+ "acc_norm_stderr": 0.025285585990017834
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.46,
113
+ "acc_stderr": 0.05009082659620333,
114
+ "acc_norm": 0.46,
115
+ "acc_norm_stderr": 0.05009082659620333
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.35,
119
+ "acc_stderr": 0.04793724854411019,
120
+ "acc_norm": 0.35,
121
+ "acc_norm_stderr": 0.04793724854411019
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5277777777777778,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.5277777777777778,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.35467980295566504,
131
+ "acc_stderr": 0.03366124489051449,
132
+ "acc_norm": 0.35467980295566504,
133
+ "acc_norm_stderr": 0.03366124489051449
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.5096774193548387,
137
+ "acc_stderr": 0.02843867799890955,
138
+ "acc_norm": 0.5096774193548387,
139
+ "acc_norm_stderr": 0.02843867799890955
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6837606837606838,
143
+ "acc_stderr": 0.030463656747340254,
144
+ "acc_norm": 0.6837606837606838,
145
+ "acc_norm_stderr": 0.030463656747340254
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4641509433962264,
149
+ "acc_stderr": 0.030693675018458003,
150
+ "acc_norm": 0.4641509433962264,
151
+ "acc_norm_stderr": 0.030693675018458003
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.44545454545454544,
155
+ "acc_stderr": 0.047605488214603246,
156
+ "acc_norm": 0.44545454545454544,
157
+ "acc_norm_stderr": 0.047605488214603246
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2777777777777778,
161
+ "acc_stderr": 0.02730914058823017,
162
+ "acc_norm": 0.2777777777777778,
163
+ "acc_norm_stderr": 0.02730914058823017
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.304635761589404,
167
+ "acc_stderr": 0.03757949922943343,
168
+ "acc_norm": 0.304635761589404,
169
+ "acc_norm_stderr": 0.03757949922943343
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5920398009950248,
173
+ "acc_stderr": 0.03475116365194092,
174
+ "acc_norm": 0.5920398009950248,
175
+ "acc_norm_stderr": 0.03475116365194092
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3988439306358382,
179
+ "acc_stderr": 0.03733626655383509,
180
+ "acc_norm": 0.3988439306358382,
181
+ "acc_norm_stderr": 0.03733626655383509
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2857142857142857,
185
+ "acc_stderr": 0.02326651221373056,
186
+ "acc_norm": 0.2857142857142857,
187
+ "acc_norm_stderr": 0.02326651221373056
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4027777777777778,
191
+ "acc_stderr": 0.04101405519842424,
192
+ "acc_norm": 0.4027777777777778,
193
+ "acc_norm_stderr": 0.04101405519842424
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.36,
197
+ "acc_stderr": 0.04824181513244218,
198
+ "acc_norm": 0.36,
199
+ "acc_norm_stderr": 0.04824181513244218
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.64,
203
+ "acc_stderr": 0.04824181513244218,
204
+ "acc_norm": 0.64,
205
+ "acc_norm_stderr": 0.04824181513244218
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5028901734104047,
209
+ "acc_stderr": 0.02691864538323901,
210
+ "acc_norm": 0.5028901734104047,
211
+ "acc_norm_stderr": 0.02691864538323901
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4723926380368098,
215
+ "acc_stderr": 0.03922378290610991,
216
+ "acc_norm": 0.4723926380368098,
217
+ "acc_norm_stderr": 0.03922378290610991
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4783950617283951,
221
+ "acc_stderr": 0.02779476010500874,
222
+ "acc_norm": 0.4783950617283951,
223
+ "acc_norm_stderr": 0.02779476010500874
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.28,
227
+ "acc_stderr": 0.045126085985421296,
228
+ "acc_norm": 0.28,
229
+ "acc_norm_stderr": 0.045126085985421296
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.538860103626943,
233
+ "acc_stderr": 0.035975244117345775,
234
+ "acc_norm": 0.538860103626943,
235
+ "acc_norm_stderr": 0.035975244117345775
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2807017543859649,
239
+ "acc_stderr": 0.042270544512321984,
240
+ "acc_norm": 0.2807017543859649,
241
+ "acc_norm_stderr": 0.042270544512321984
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5229357798165137,
245
+ "acc_stderr": 0.0214147570581755,
246
+ "acc_norm": 0.5229357798165137,
247
+ "acc_norm_stderr": 0.0214147570581755
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3412698412698413,
251
+ "acc_stderr": 0.04240799327574925,
252
+ "acc_norm": 0.3412698412698413,
253
+ "acc_norm_stderr": 0.04240799327574925
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.477124183006536,
257
+ "acc_stderr": 0.028599936776089775,
258
+ "acc_norm": 0.477124183006536,
259
+ "acc_norm_stderr": 0.028599936776089775
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.45,
263
+ "acc_stderr": 0.05,
264
+ "acc_norm": 0.45,
265
+ "acc_norm_stderr": 0.05
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6776859504132231,
269
+ "acc_stderr": 0.04266416363352167,
270
+ "acc_norm": 0.6776859504132231,
271
+ "acc_norm_stderr": 0.04266416363352167
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.46710526315789475,
275
+ "acc_stderr": 0.040601270352363966,
276
+ "acc_norm": 0.46710526315789475,
277
+ "acc_norm_stderr": 0.040601270352363966
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3660130718954248,
281
+ "acc_stderr": 0.019488025745529672,
282
+ "acc_norm": 0.3660130718954248,
283
+ "acc_norm_stderr": 0.019488025745529672
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.35106382978723405,
287
+ "acc_stderr": 0.028473501272963775,
288
+ "acc_norm": 0.35106382978723405,
289
+ "acc_norm_stderr": 0.028473501272963775
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25,
293
+ "acc_stderr": 0.04109974682633932,
294
+ "acc_norm": 0.25,
295
+ "acc_norm_stderr": 0.04109974682633932
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.375,
299
+ "acc_stderr": 0.033016908987210894,
300
+ "acc_norm": 0.375,
301
+ "acc_norm_stderr": 0.033016908987210894
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2547486033519553,
305
+ "acc_stderr": 0.014572650383409167,
306
+ "acc_norm": 0.2547486033519553,
307
+ "acc_norm_stderr": 0.014572650383409167
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.45,
311
+ "acc_stderr": 0.05,
312
+ "acc_norm": 0.45,
313
+ "acc_norm_stderr": 0.05
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.42,
317
+ "acc_stderr": 0.04960449637488584,
318
+ "acc_norm": 0.42,
319
+ "acc_norm_stderr": 0.04960449637488584
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.45955882352941174,
323
+ "acc_stderr": 0.03027332507734575,
324
+ "acc_norm": 0.45955882352941174,
325
+ "acc_norm_stderr": 0.03027332507734575
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4897959183673469,
329
+ "acc_stderr": 0.03200255347893782,
330
+ "acc_norm": 0.4897959183673469,
331
+ "acc_norm_stderr": 0.03200255347893782
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6033755274261603,
335
+ "acc_stderr": 0.03184399873811225,
336
+ "acc_norm": 0.6033755274261603,
337
+ "acc_norm_stderr": 0.03184399873811225
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3285528031290743,
341
+ "acc_stderr": 0.011996027247502912,
342
+ "acc_norm": 0.3285528031290743,
343
+ "acc_norm_stderr": 0.011996027247502912
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.49019607843137253,
347
+ "acc_stderr": 0.035086373586305716,
348
+ "acc_norm": 0.49019607843137253,
349
+ "acc_norm_stderr": 0.035086373586305716
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5575757575757576,
353
+ "acc_stderr": 0.03878372113711275,
354
+ "acc_norm": 0.5575757575757576,
355
+ "acc_norm_stderr": 0.03878372113711275
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.27906976744186046,
359
+ "mc1_stderr": 0.01570210709062789,
360
+ "mc2": 0.44866578973581106,
361
+ "mc2_stderr": 0.015416926437342405
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.48406139315230223,
365
+ "acc_stderr": 0.017181617837190192,
366
+ "acc_norm": 0.5619834710743802,
367
+ "acc_norm_stderr": 0.01705775370216029
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-instruct-v3-13B",
436
+ "model_sha": "199c2113f09f153bce1ad7aac35e6e756a99b89b",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-instruct-v4-13B/result_2023-11-09 09:34:14.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.4087030716723549,
5
+ "acc_stderr": 0.014365750345427006,
6
+ "acc_norm": 0.4564846416382253,
7
+ "acc_norm_stderr": 0.01455594976049644
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.43168691495717987,
11
+ "acc_stderr": 0.0049429906231311166,
12
+ "acc_norm": 0.5795658235411273,
13
+ "acc_norm_stderr": 0.0049261984839487115
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5321637426900585,
17
+ "acc_stderr": 0.038268824176603704,
18
+ "acc_norm": 0.5321637426900585,
19
+ "acc_norm_stderr": 0.038268824176603704
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.49514563106796117,
23
+ "acc_stderr": 0.049505043821289195,
24
+ "acc_norm": 0.49514563106796117,
25
+ "acc_norm_stderr": 0.049505043821289195
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5555555555555556,
29
+ "acc_stderr": 0.017769250583533246,
30
+ "acc_norm": 0.5555555555555556,
31
+ "acc_norm_stderr": 0.017769250583533246
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4074074074074074,
35
+ "acc_stderr": 0.0424463323835323,
36
+ "acc_norm": 0.4074074074074074,
37
+ "acc_norm_stderr": 0.0424463323835323
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.27,
41
+ "acc_stderr": 0.04461960433384741,
42
+ "acc_norm": 0.27,
43
+ "acc_norm_stderr": 0.04461960433384741
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.4,
47
+ "acc_stderr": 0.03202563076101735,
48
+ "acc_norm": 0.4,
49
+ "acc_norm_stderr": 0.03202563076101735
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.4457831325301205,
53
+ "acc_stderr": 0.038695433234721015,
54
+ "acc_norm": 0.4457831325301205,
55
+ "acc_norm_stderr": 0.038695433234721015
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4855305466237942,
59
+ "acc_stderr": 0.028386198084177673,
60
+ "acc_norm": 0.4855305466237942,
61
+ "acc_norm_stderr": 0.028386198084177673
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.5022421524663677,
65
+ "acc_stderr": 0.033557465352232634,
66
+ "acc_norm": 0.5022421524663677,
67
+ "acc_norm_stderr": 0.033557465352232634
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4732824427480916,
71
+ "acc_stderr": 0.04379024936553894,
72
+ "acc_norm": 0.4732824427480916,
73
+ "acc_norm_stderr": 0.04379024936553894
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.39,
77
+ "acc_stderr": 0.04902071300001975,
78
+ "acc_norm": 0.39,
79
+ "acc_norm_stderr": 0.04902071300001975
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5909090909090909,
83
+ "acc_stderr": 0.035029757994130085,
84
+ "acc_norm": 0.5909090909090909,
85
+ "acc_norm_stderr": 0.035029757994130085
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.3724137931034483,
89
+ "acc_stderr": 0.0402873153294756,
90
+ "acc_norm": 0.3724137931034483,
91
+ "acc_norm_stderr": 0.0402873153294756
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.22549019607843138,
95
+ "acc_stderr": 0.04158307533083286,
96
+ "acc_norm": 0.22549019607843138,
97
+ "acc_norm_stderr": 0.04158307533083286
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.453781512605042,
101
+ "acc_stderr": 0.03233943468182088,
102
+ "acc_norm": 0.453781512605042,
103
+ "acc_norm_stderr": 0.03233943468182088
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.45897435897435895,
107
+ "acc_stderr": 0.025265525491284295,
108
+ "acc_norm": 0.45897435897435895,
109
+ "acc_norm_stderr": 0.025265525491284295
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.54,
113
+ "acc_stderr": 0.05009082659620332,
114
+ "acc_norm": 0.54,
115
+ "acc_norm_stderr": 0.05009082659620332
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.36,
119
+ "acc_stderr": 0.048241815132442176,
120
+ "acc_norm": 0.36,
121
+ "acc_norm_stderr": 0.048241815132442176
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5462962962962963,
125
+ "acc_stderr": 0.048129173245368216,
126
+ "acc_norm": 0.5462962962962963,
127
+ "acc_norm_stderr": 0.048129173245368216
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.35467980295566504,
131
+ "acc_stderr": 0.03366124489051449,
132
+ "acc_norm": 0.35467980295566504,
133
+ "acc_norm_stderr": 0.03366124489051449
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.45806451612903226,
137
+ "acc_stderr": 0.028343787250540636,
138
+ "acc_norm": 0.45806451612903226,
139
+ "acc_norm_stderr": 0.028343787250540636
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6495726495726496,
143
+ "acc_stderr": 0.0312561082442188,
144
+ "acc_norm": 0.6495726495726496,
145
+ "acc_norm_stderr": 0.0312561082442188
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4490566037735849,
149
+ "acc_stderr": 0.030612730713641092,
150
+ "acc_norm": 0.4490566037735849,
151
+ "acc_norm_stderr": 0.030612730713641092
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.6,
155
+ "acc_stderr": 0.0469237132203465,
156
+ "acc_norm": 0.6,
157
+ "acc_norm_stderr": 0.0469237132203465
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2518518518518518,
161
+ "acc_stderr": 0.026466117538959916,
162
+ "acc_norm": 0.2518518518518518,
163
+ "acc_norm_stderr": 0.026466117538959916
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.31125827814569534,
167
+ "acc_stderr": 0.03780445850526733,
168
+ "acc_norm": 0.31125827814569534,
169
+ "acc_norm_stderr": 0.03780445850526733
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5671641791044776,
173
+ "acc_stderr": 0.03503490923673281,
174
+ "acc_norm": 0.5671641791044776,
175
+ "acc_norm_stderr": 0.03503490923673281
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.4161849710982659,
179
+ "acc_stderr": 0.037585177754049466,
180
+ "acc_norm": 0.4161849710982659,
181
+ "acc_norm_stderr": 0.037585177754049466
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.30423280423280424,
185
+ "acc_stderr": 0.023695415009463087,
186
+ "acc_norm": 0.30423280423280424,
187
+ "acc_norm_stderr": 0.023695415009463087
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3958333333333333,
191
+ "acc_stderr": 0.04089465449325583,
192
+ "acc_norm": 0.3958333333333333,
193
+ "acc_norm_stderr": 0.04089465449325583
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.31,
197
+ "acc_stderr": 0.04648231987117316,
198
+ "acc_norm": 0.31,
199
+ "acc_norm_stderr": 0.04648231987117316
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.64,
203
+ "acc_stderr": 0.048241815132442176,
204
+ "acc_norm": 0.64,
205
+ "acc_norm_stderr": 0.048241815132442176
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.49710982658959535,
209
+ "acc_stderr": 0.02691864538323901,
210
+ "acc_norm": 0.49710982658959535,
211
+ "acc_norm_stderr": 0.02691864538323901
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.5766871165644172,
215
+ "acc_stderr": 0.03881891213334383,
216
+ "acc_norm": 0.5766871165644172,
217
+ "acc_norm_stderr": 0.03881891213334383
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.49691358024691357,
221
+ "acc_stderr": 0.027820214158594377,
222
+ "acc_norm": 0.49691358024691357,
223
+ "acc_norm_stderr": 0.027820214158594377
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.36,
227
+ "acc_stderr": 0.04824181513244218,
228
+ "acc_norm": 0.36,
229
+ "acc_norm_stderr": 0.04824181513244218
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5440414507772021,
233
+ "acc_stderr": 0.03594413711272436,
234
+ "acc_norm": 0.5440414507772021,
235
+ "acc_norm_stderr": 0.03594413711272436
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.21052631578947367,
239
+ "acc_stderr": 0.03835153954399419,
240
+ "acc_norm": 0.21052631578947367,
241
+ "acc_norm_stderr": 0.03835153954399419
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5871559633027523,
245
+ "acc_stderr": 0.02110912813341391,
246
+ "acc_norm": 0.5871559633027523,
247
+ "acc_norm_stderr": 0.02110912813341391
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3333333333333333,
251
+ "acc_stderr": 0.04216370213557835,
252
+ "acc_norm": 0.3333333333333333,
253
+ "acc_norm_stderr": 0.04216370213557835
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4215686274509804,
257
+ "acc_stderr": 0.02827549015679143,
258
+ "acc_norm": 0.4215686274509804,
259
+ "acc_norm_stderr": 0.02827549015679143
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.47,
263
+ "acc_stderr": 0.05016135580465919,
264
+ "acc_norm": 0.47,
265
+ "acc_norm_stderr": 0.05016135580465919
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.5950413223140496,
269
+ "acc_stderr": 0.04481137755942469,
270
+ "acc_norm": 0.5950413223140496,
271
+ "acc_norm_stderr": 0.04481137755942469
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.375,
275
+ "acc_stderr": 0.039397364351956274,
276
+ "acc_norm": 0.375,
277
+ "acc_norm_stderr": 0.039397364351956274
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.42810457516339867,
281
+ "acc_stderr": 0.0200176292142131,
282
+ "acc_norm": 0.42810457516339867,
283
+ "acc_norm_stderr": 0.0200176292142131
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3404255319148936,
287
+ "acc_stderr": 0.028267657482650144,
288
+ "acc_norm": 0.3404255319148936,
289
+ "acc_norm_stderr": 0.028267657482650144
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.2767857142857143,
293
+ "acc_stderr": 0.04246624336697625,
294
+ "acc_norm": 0.2767857142857143,
295
+ "acc_norm_stderr": 0.04246624336697625
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.33796296296296297,
299
+ "acc_stderr": 0.03225941352631295,
300
+ "acc_norm": 0.33796296296296297,
301
+ "acc_norm_stderr": 0.03225941352631295
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2435754189944134,
305
+ "acc_stderr": 0.01435591196476786,
306
+ "acc_norm": 0.2435754189944134,
307
+ "acc_norm_stderr": 0.01435591196476786
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.38,
311
+ "acc_stderr": 0.048783173121456316,
312
+ "acc_norm": 0.38,
313
+ "acc_norm_stderr": 0.048783173121456316
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.39,
317
+ "acc_stderr": 0.04902071300001974,
318
+ "acc_norm": 0.39,
319
+ "acc_norm_stderr": 0.04902071300001974
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.38235294117647056,
323
+ "acc_stderr": 0.02952009569768775,
324
+ "acc_norm": 0.38235294117647056,
325
+ "acc_norm_stderr": 0.02952009569768775
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5020408163265306,
329
+ "acc_stderr": 0.0320089533497105,
330
+ "acc_norm": 0.5020408163265306,
331
+ "acc_norm_stderr": 0.0320089533497105
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6751054852320675,
335
+ "acc_stderr": 0.030486039389105303,
336
+ "acc_norm": 0.6751054852320675,
337
+ "acc_norm_stderr": 0.030486039389105303
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3741851368970013,
341
+ "acc_stderr": 0.012359335618172063,
342
+ "acc_norm": 0.3741851368970013,
343
+ "acc_norm_stderr": 0.012359335618172063
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.5294117647058824,
347
+ "acc_stderr": 0.03503235296367993,
348
+ "acc_norm": 0.5294117647058824,
349
+ "acc_norm_stderr": 0.03503235296367993
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5575757575757576,
353
+ "acc_stderr": 0.03878372113711275,
354
+ "acc_norm": 0.5575757575757576,
355
+ "acc_norm_stderr": 0.03878372113711275
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.3243574051407589,
359
+ "mc1_stderr": 0.01638797677964794,
360
+ "mc2": 0.4753344144954286,
361
+ "mc2_stderr": 0.015470233894001158
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.500590318772137,
365
+ "acc_stderr": 0.01719034212344859,
366
+ "acc_norm": 0.5726092089728453,
367
+ "acc_norm_stderr": 0.017008129844823156
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-instruct-v4-13B",
436
+ "model_sha": "fabf605d23d96e548908ffe9f0ad49dae01c46f8",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-instruct-v5-13B/result_2023-11-16 17:37:54.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3822525597269625,
5
+ "acc_stderr": 0.014200454049979272,
6
+ "acc_norm": 0.43600682593856654,
7
+ "acc_norm_stderr": 0.014491225699230918
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.4221270663214499,
11
+ "acc_stderr": 0.00492889189587429,
12
+ "acc_norm": 0.5567616012746465,
13
+ "acc_norm_stderr": 0.004957524197900418
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4619883040935672,
17
+ "acc_stderr": 0.03823727092882307,
18
+ "acc_norm": 0.4619883040935672,
19
+ "acc_norm_stderr": 0.03823727092882307
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.49514563106796117,
23
+ "acc_stderr": 0.04950504382128921,
24
+ "acc_norm": 0.49514563106796117,
25
+ "acc_norm_stderr": 0.04950504382128921
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5274584929757343,
29
+ "acc_stderr": 0.017852981266633944,
30
+ "acc_norm": 0.5274584929757343,
31
+ "acc_norm_stderr": 0.017852981266633944
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4740740740740741,
35
+ "acc_stderr": 0.04313531696750573,
36
+ "acc_norm": 0.4740740740740741,
37
+ "acc_norm_stderr": 0.04313531696750573
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.29,
41
+ "acc_stderr": 0.045604802157206824,
42
+ "acc_norm": 0.29,
43
+ "acc_norm_stderr": 0.045604802157206824
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.40425531914893614,
47
+ "acc_stderr": 0.03208115750788684,
48
+ "acc_norm": 0.40425531914893614,
49
+ "acc_norm_stderr": 0.03208115750788684
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39156626506024095,
53
+ "acc_stderr": 0.03799857454479637,
54
+ "acc_norm": 0.39156626506024095,
55
+ "acc_norm_stderr": 0.03799857454479637
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.47266881028938906,
59
+ "acc_stderr": 0.028355633568328174,
60
+ "acc_norm": 0.47266881028938906,
61
+ "acc_norm_stderr": 0.028355633568328174
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.47533632286995514,
65
+ "acc_stderr": 0.03351695167652628,
66
+ "acc_norm": 0.47533632286995514,
67
+ "acc_norm_stderr": 0.03351695167652628
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4580152671755725,
71
+ "acc_stderr": 0.04369802690578756,
72
+ "acc_norm": 0.4580152671755725,
73
+ "acc_norm_stderr": 0.04369802690578756
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.42,
77
+ "acc_stderr": 0.049604496374885836,
78
+ "acc_norm": 0.42,
79
+ "acc_norm_stderr": 0.049604496374885836
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.51010101010101,
83
+ "acc_stderr": 0.035616254886737454,
84
+ "acc_norm": 0.51010101010101,
85
+ "acc_norm_stderr": 0.035616254886737454
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4896551724137931,
89
+ "acc_stderr": 0.04165774775728763,
90
+ "acc_norm": 0.4896551724137931,
91
+ "acc_norm_stderr": 0.04165774775728763
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.20588235294117646,
95
+ "acc_stderr": 0.04023382273617747,
96
+ "acc_norm": 0.20588235294117646,
97
+ "acc_norm_stderr": 0.04023382273617747
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.40756302521008403,
101
+ "acc_stderr": 0.031918633744784645,
102
+ "acc_norm": 0.40756302521008403,
103
+ "acc_norm_stderr": 0.031918633744784645
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.39487179487179486,
107
+ "acc_stderr": 0.024784316942156367,
108
+ "acc_norm": 0.39487179487179486,
109
+ "acc_norm_stderr": 0.024784316942156367
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.53,
113
+ "acc_stderr": 0.05016135580465919,
114
+ "acc_norm": 0.53,
115
+ "acc_norm_stderr": 0.05016135580465919
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.33,
119
+ "acc_stderr": 0.047258156262526045,
120
+ "acc_norm": 0.33,
121
+ "acc_norm_stderr": 0.047258156262526045
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5277777777777778,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.5277777777777778,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.39408866995073893,
131
+ "acc_stderr": 0.03438157967036545,
132
+ "acc_norm": 0.39408866995073893,
133
+ "acc_norm_stderr": 0.03438157967036545
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.44193548387096776,
137
+ "acc_stderr": 0.028251557906849734,
138
+ "acc_norm": 0.44193548387096776,
139
+ "acc_norm_stderr": 0.028251557906849734
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6324786324786325,
143
+ "acc_stderr": 0.031585391577456365,
144
+ "acc_norm": 0.6324786324786325,
145
+ "acc_norm_stderr": 0.031585391577456365
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4188679245283019,
149
+ "acc_stderr": 0.030365050829115208,
150
+ "acc_norm": 0.4188679245283019,
151
+ "acc_norm_stderr": 0.030365050829115208
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4090909090909091,
155
+ "acc_stderr": 0.04709306978661896,
156
+ "acc_norm": 0.4090909090909091,
157
+ "acc_norm_stderr": 0.04709306978661896
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2740740740740741,
161
+ "acc_stderr": 0.027195934804085626,
162
+ "acc_norm": 0.2740740740740741,
163
+ "acc_norm_stderr": 0.027195934804085626
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2913907284768212,
167
+ "acc_stderr": 0.03710185726119995,
168
+ "acc_norm": 0.2913907284768212,
169
+ "acc_norm_stderr": 0.03710185726119995
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5522388059701493,
173
+ "acc_stderr": 0.03516184772952167,
174
+ "acc_norm": 0.5522388059701493,
175
+ "acc_norm_stderr": 0.03516184772952167
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3815028901734104,
179
+ "acc_stderr": 0.03703851193099521,
180
+ "acc_norm": 0.3815028901734104,
181
+ "acc_norm_stderr": 0.03703851193099521
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2804232804232804,
185
+ "acc_stderr": 0.02313528797432563,
186
+ "acc_norm": 0.2804232804232804,
187
+ "acc_norm_stderr": 0.02313528797432563
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3472222222222222,
191
+ "acc_stderr": 0.039812405437178615,
192
+ "acc_norm": 0.3472222222222222,
193
+ "acc_norm_stderr": 0.039812405437178615
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.3,
197
+ "acc_stderr": 0.046056618647183814,
198
+ "acc_norm": 0.3,
199
+ "acc_norm_stderr": 0.046056618647183814
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.61,
203
+ "acc_stderr": 0.04902071300001975,
204
+ "acc_norm": 0.61,
205
+ "acc_norm_stderr": 0.04902071300001975
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5115606936416185,
209
+ "acc_stderr": 0.026911898686377906,
210
+ "acc_norm": 0.5115606936416185,
211
+ "acc_norm_stderr": 0.026911898686377906
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.44785276073619634,
215
+ "acc_stderr": 0.03906947479456601,
216
+ "acc_norm": 0.44785276073619634,
217
+ "acc_norm_stderr": 0.03906947479456601
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.43209876543209874,
221
+ "acc_stderr": 0.02756301097160667,
222
+ "acc_norm": 0.43209876543209874,
223
+ "acc_norm_stderr": 0.02756301097160667
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.3,
227
+ "acc_stderr": 0.046056618647183814,
228
+ "acc_norm": 0.3,
229
+ "acc_norm_stderr": 0.046056618647183814
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.538860103626943,
233
+ "acc_stderr": 0.035975244117345775,
234
+ "acc_norm": 0.538860103626943,
235
+ "acc_norm_stderr": 0.035975244117345775
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2982456140350877,
239
+ "acc_stderr": 0.04303684033537316,
240
+ "acc_norm": 0.2982456140350877,
241
+ "acc_norm_stderr": 0.04303684033537316
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5009174311926605,
245
+ "acc_stderr": 0.021437287056051215,
246
+ "acc_norm": 0.5009174311926605,
247
+ "acc_norm_stderr": 0.021437287056051215
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.2619047619047619,
251
+ "acc_stderr": 0.0393253768039287,
252
+ "acc_norm": 0.2619047619047619,
253
+ "acc_norm_stderr": 0.0393253768039287
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.434640522875817,
257
+ "acc_stderr": 0.028384256704883037,
258
+ "acc_norm": 0.434640522875817,
259
+ "acc_norm_stderr": 0.028384256704883037
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.42,
263
+ "acc_stderr": 0.049604496374885836,
264
+ "acc_norm": 0.42,
265
+ "acc_norm_stderr": 0.049604496374885836
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.5950413223140496,
269
+ "acc_stderr": 0.04481137755942469,
270
+ "acc_norm": 0.5950413223140496,
271
+ "acc_norm_stderr": 0.04481137755942469
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3684210526315789,
275
+ "acc_stderr": 0.03925523381052932,
276
+ "acc_norm": 0.3684210526315789,
277
+ "acc_norm_stderr": 0.03925523381052932
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.380718954248366,
281
+ "acc_stderr": 0.019643801557924806,
282
+ "acc_norm": 0.380718954248366,
283
+ "acc_norm_stderr": 0.019643801557924806
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.31560283687943264,
287
+ "acc_stderr": 0.027724989449509314,
288
+ "acc_norm": 0.31560283687943264,
289
+ "acc_norm_stderr": 0.027724989449509314
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.2767857142857143,
293
+ "acc_stderr": 0.042466243366976235,
294
+ "acc_norm": 0.2767857142857143,
295
+ "acc_norm_stderr": 0.042466243366976235
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.26851851851851855,
299
+ "acc_stderr": 0.030225226160012386,
300
+ "acc_norm": 0.26851851851851855,
301
+ "acc_norm_stderr": 0.030225226160012386
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2446927374301676,
305
+ "acc_stderr": 0.014378169884098424,
306
+ "acc_norm": 0.2446927374301676,
307
+ "acc_norm_stderr": 0.014378169884098424
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.37,
311
+ "acc_stderr": 0.04852365870939099,
312
+ "acc_norm": 0.37,
313
+ "acc_norm_stderr": 0.04852365870939099
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.47,
317
+ "acc_stderr": 0.05016135580465919,
318
+ "acc_norm": 0.47,
319
+ "acc_norm_stderr": 0.05016135580465919
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3382352941176471,
323
+ "acc_stderr": 0.028739328513983576,
324
+ "acc_norm": 0.3382352941176471,
325
+ "acc_norm_stderr": 0.028739328513983576
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.40816326530612246,
329
+ "acc_stderr": 0.03146465712827424,
330
+ "acc_norm": 0.40816326530612246,
331
+ "acc_norm_stderr": 0.03146465712827424
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.569620253164557,
335
+ "acc_stderr": 0.03223017195937598,
336
+ "acc_norm": 0.569620253164557,
337
+ "acc_norm_stderr": 0.03223017195937598
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.33116036505867014,
341
+ "acc_stderr": 0.01202012819598575,
342
+ "acc_norm": 0.33116036505867014,
343
+ "acc_norm_stderr": 0.01202012819598575
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.45098039215686275,
347
+ "acc_stderr": 0.03492406104163614,
348
+ "acc_norm": 0.45098039215686275,
349
+ "acc_norm_stderr": 0.03492406104163614
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5636363636363636,
353
+ "acc_stderr": 0.03872592983524754,
354
+ "acc_norm": 0.5636363636363636,
355
+ "acc_norm_stderr": 0.03872592983524754
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.29253365973072215,
359
+ "mc1_stderr": 0.015925597445286165,
360
+ "mc2": 0.4591418911312825,
361
+ "mc2_stderr": 0.015363002653584545
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.4911452184179457,
365
+ "acc_stderr": 0.017187658199336736,
366
+ "acc_norm": 0.5454545454545454,
367
+ "acc_norm_stderr": 0.017119172208061504
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-instruct-v5-13B",
436
+ "model_sha": "e625b2673e2a0839e7d3fc0f2a844e9966404678",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-instruct-v6-13B/result_2023-11-27 02:12:26.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3916382252559727,
5
+ "acc_stderr": 0.014264122124938215,
6
+ "acc_norm": 0.45051194539249145,
7
+ "acc_norm_stderr": 0.014539646098471627
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.4272057359091814,
11
+ "acc_stderr": 0.004936616428922639,
12
+ "acc_norm": 0.5610436168094005,
13
+ "acc_norm_stderr": 0.004952454721934797
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4678362573099415,
17
+ "acc_stderr": 0.03826882417660369,
18
+ "acc_norm": 0.4678362573099415,
19
+ "acc_norm_stderr": 0.03826882417660369
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.4854368932038835,
23
+ "acc_stderr": 0.04948637324026637,
24
+ "acc_norm": 0.4854368932038835,
25
+ "acc_norm_stderr": 0.04948637324026637
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5210727969348659,
29
+ "acc_stderr": 0.017864076786212903,
30
+ "acc_norm": 0.5210727969348659,
31
+ "acc_norm_stderr": 0.017864076786212903
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4148148148148148,
35
+ "acc_stderr": 0.04256193767901407,
36
+ "acc_norm": 0.4148148148148148,
37
+ "acc_norm_stderr": 0.04256193767901407
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.29,
41
+ "acc_stderr": 0.045604802157206824,
42
+ "acc_norm": 0.29,
43
+ "acc_norm_stderr": 0.045604802157206824
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.39574468085106385,
47
+ "acc_stderr": 0.03196758697835362,
48
+ "acc_norm": 0.39574468085106385,
49
+ "acc_norm_stderr": 0.03196758697835362
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.42771084337349397,
53
+ "acc_stderr": 0.03851597683718533,
54
+ "acc_norm": 0.42771084337349397,
55
+ "acc_norm_stderr": 0.03851597683718533
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4565916398713826,
59
+ "acc_stderr": 0.028290869054197604,
60
+ "acc_norm": 0.4565916398713826,
61
+ "acc_norm_stderr": 0.028290869054197604
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.47533632286995514,
65
+ "acc_stderr": 0.03351695167652629,
66
+ "acc_norm": 0.47533632286995514,
67
+ "acc_norm_stderr": 0.03351695167652629
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4580152671755725,
71
+ "acc_stderr": 0.04369802690578756,
72
+ "acc_norm": 0.4580152671755725,
73
+ "acc_norm_stderr": 0.04369802690578756
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.37,
77
+ "acc_stderr": 0.04852365870939098,
78
+ "acc_norm": 0.37,
79
+ "acc_norm_stderr": 0.04852365870939098
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5202020202020202,
83
+ "acc_stderr": 0.03559443565563919,
84
+ "acc_norm": 0.5202020202020202,
85
+ "acc_norm_stderr": 0.03559443565563919
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4068965517241379,
89
+ "acc_stderr": 0.04093793981266237,
90
+ "acc_norm": 0.4068965517241379,
91
+ "acc_norm_stderr": 0.04093793981266237
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.23529411764705882,
95
+ "acc_stderr": 0.04220773659171453,
96
+ "acc_norm": 0.23529411764705882,
97
+ "acc_norm_stderr": 0.04220773659171453
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.40756302521008403,
101
+ "acc_stderr": 0.031918633744784645,
102
+ "acc_norm": 0.40756302521008403,
103
+ "acc_norm_stderr": 0.031918633744784645
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4230769230769231,
107
+ "acc_stderr": 0.02504919787604233,
108
+ "acc_norm": 0.4230769230769231,
109
+ "acc_norm_stderr": 0.02504919787604233
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.57,
113
+ "acc_stderr": 0.049756985195624284,
114
+ "acc_norm": 0.57,
115
+ "acc_norm_stderr": 0.049756985195624284
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.35,
119
+ "acc_stderr": 0.0479372485441102,
120
+ "acc_norm": 0.35,
121
+ "acc_norm_stderr": 0.0479372485441102
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.48148148148148145,
125
+ "acc_stderr": 0.04830366024635331,
126
+ "acc_norm": 0.48148148148148145,
127
+ "acc_norm_stderr": 0.04830366024635331
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.39408866995073893,
131
+ "acc_stderr": 0.034381579670365446,
132
+ "acc_norm": 0.39408866995073893,
133
+ "acc_norm_stderr": 0.034381579670365446
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.45483870967741935,
137
+ "acc_stderr": 0.028327743091561056,
138
+ "acc_norm": 0.45483870967741935,
139
+ "acc_norm_stderr": 0.028327743091561056
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6410256410256411,
143
+ "acc_stderr": 0.03142616993791924,
144
+ "acc_norm": 0.6410256410256411,
145
+ "acc_norm_stderr": 0.03142616993791924
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.45660377358490567,
149
+ "acc_stderr": 0.03065674869673943,
150
+ "acc_norm": 0.45660377358490567,
151
+ "acc_norm_stderr": 0.03065674869673943
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4727272727272727,
155
+ "acc_stderr": 0.04782001791380063,
156
+ "acc_norm": 0.4727272727272727,
157
+ "acc_norm_stderr": 0.04782001791380063
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.24814814814814815,
161
+ "acc_stderr": 0.0263357394040558,
162
+ "acc_norm": 0.24814814814814815,
163
+ "acc_norm_stderr": 0.0263357394040558
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.31788079470198677,
167
+ "acc_stderr": 0.03802039760107903,
168
+ "acc_norm": 0.31788079470198677,
169
+ "acc_norm_stderr": 0.03802039760107903
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.572139303482587,
173
+ "acc_stderr": 0.03498541988407795,
174
+ "acc_norm": 0.572139303482587,
175
+ "acc_norm_stderr": 0.03498541988407795
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3699421965317919,
179
+ "acc_stderr": 0.03681229633394319,
180
+ "acc_norm": 0.3699421965317919,
181
+ "acc_norm_stderr": 0.03681229633394319
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2724867724867725,
185
+ "acc_stderr": 0.022930973071633345,
186
+ "acc_norm": 0.2724867724867725,
187
+ "acc_norm_stderr": 0.022930973071633345
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3263888888888889,
191
+ "acc_stderr": 0.03921067198982266,
192
+ "acc_norm": 0.3263888888888889,
193
+ "acc_norm_stderr": 0.03921067198982266
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.36,
197
+ "acc_stderr": 0.04824181513244218,
198
+ "acc_norm": 0.36,
199
+ "acc_norm_stderr": 0.04824181513244218
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.59,
203
+ "acc_stderr": 0.04943110704237102,
204
+ "acc_norm": 0.59,
205
+ "acc_norm_stderr": 0.04943110704237102
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.4421965317919075,
209
+ "acc_stderr": 0.026738603643807403,
210
+ "acc_norm": 0.4421965317919075,
211
+ "acc_norm_stderr": 0.026738603643807403
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.44785276073619634,
215
+ "acc_stderr": 0.03906947479456601,
216
+ "acc_norm": 0.44785276073619634,
217
+ "acc_norm_stderr": 0.03906947479456601
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4567901234567901,
221
+ "acc_stderr": 0.027716661650194038,
222
+ "acc_norm": 0.4567901234567901,
223
+ "acc_norm_stderr": 0.027716661650194038
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.3,
227
+ "acc_stderr": 0.046056618647183814,
228
+ "acc_norm": 0.3,
229
+ "acc_norm_stderr": 0.046056618647183814
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.49740932642487046,
233
+ "acc_stderr": 0.03608390745384487,
234
+ "acc_norm": 0.49740932642487046,
235
+ "acc_norm_stderr": 0.03608390745384487
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.20175438596491227,
239
+ "acc_stderr": 0.03775205013583639,
240
+ "acc_norm": 0.20175438596491227,
241
+ "acc_norm_stderr": 0.03775205013583639
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.4935779816513762,
245
+ "acc_stderr": 0.021435554820013077,
246
+ "acc_norm": 0.4935779816513762,
247
+ "acc_norm_stderr": 0.021435554820013077
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.30952380952380953,
251
+ "acc_stderr": 0.04134913018303316,
252
+ "acc_norm": 0.30952380952380953,
253
+ "acc_norm_stderr": 0.04134913018303316
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.42810457516339867,
257
+ "acc_stderr": 0.02833239748366427,
258
+ "acc_norm": 0.42810457516339867,
259
+ "acc_norm_stderr": 0.02833239748366427
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.43,
263
+ "acc_stderr": 0.049756985195624284,
264
+ "acc_norm": 0.43,
265
+ "acc_norm_stderr": 0.049756985195624284
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.5950413223140496,
269
+ "acc_stderr": 0.04481137755942469,
270
+ "acc_norm": 0.5950413223140496,
271
+ "acc_norm_stderr": 0.04481137755942469
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3881578947368421,
275
+ "acc_stderr": 0.03965842097512744,
276
+ "acc_norm": 0.3881578947368421,
277
+ "acc_norm_stderr": 0.03965842097512744
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3660130718954248,
281
+ "acc_stderr": 0.019488025745529672,
282
+ "acc_norm": 0.3660130718954248,
283
+ "acc_norm_stderr": 0.019488025745529672
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.2801418439716312,
287
+ "acc_stderr": 0.02678917235114024,
288
+ "acc_norm": 0.2801418439716312,
289
+ "acc_norm_stderr": 0.02678917235114024
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25892857142857145,
293
+ "acc_stderr": 0.04157751539865629,
294
+ "acc_norm": 0.25892857142857145,
295
+ "acc_norm_stderr": 0.04157751539865629
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3101851851851852,
299
+ "acc_stderr": 0.031546962856566295,
300
+ "acc_norm": 0.3101851851851852,
301
+ "acc_norm_stderr": 0.031546962856566295
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2446927374301676,
305
+ "acc_stderr": 0.014378169884098424,
306
+ "acc_norm": 0.2446927374301676,
307
+ "acc_norm_stderr": 0.014378169884098424
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.33,
311
+ "acc_stderr": 0.047258156262526045,
312
+ "acc_norm": 0.33,
313
+ "acc_norm_stderr": 0.047258156262526045
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.44,
317
+ "acc_stderr": 0.049888765156985884,
318
+ "acc_norm": 0.44,
319
+ "acc_norm_stderr": 0.049888765156985884
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.33088235294117646,
323
+ "acc_stderr": 0.028582709753898428,
324
+ "acc_norm": 0.33088235294117646,
325
+ "acc_norm_stderr": 0.028582709753898428
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.46938775510204084,
329
+ "acc_stderr": 0.031949171367580624,
330
+ "acc_norm": 0.46938775510204084,
331
+ "acc_norm_stderr": 0.031949171367580624
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5485232067510548,
335
+ "acc_stderr": 0.032393600173974704,
336
+ "acc_norm": 0.5485232067510548,
337
+ "acc_norm_stderr": 0.032393600173974704
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.31486310299869624,
341
+ "acc_stderr": 0.011862561755715945,
342
+ "acc_norm": 0.31486310299869624,
343
+ "acc_norm_stderr": 0.011862561755715945
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4411764705882353,
347
+ "acc_stderr": 0.034849415144292316,
348
+ "acc_norm": 0.4411764705882353,
349
+ "acc_norm_stderr": 0.034849415144292316
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5393939393939394,
353
+ "acc_stderr": 0.03892207016552012,
354
+ "acc_norm": 0.5393939393939394,
355
+ "acc_norm_stderr": 0.03892207016552012
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.3023255813953488,
359
+ "mc1_stderr": 0.01607750926613303,
360
+ "mc2": 0.45710797981768625,
361
+ "mc2_stderr": 0.015464643764155465
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.46871310507674147,
365
+ "acc_stderr": 0.017156666859785456,
366
+ "acc_norm": 0.5619834710743802,
367
+ "acc_norm_stderr": 0.017057753702160294
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-instruct-v6-13B",
436
+ "model_sha": "f24326c48f4edb60bc3bdc186b65e0fcb9254c1e",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-instruct-v7-13B/result_2023-12-03 17:19:11.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3916382252559727,
5
+ "acc_stderr": 0.014264122124938213,
6
+ "acc_norm": 0.4564846416382253,
7
+ "acc_norm_stderr": 0.014555949760496435
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.42272455686118304,
11
+ "acc_stderr": 0.00492982833760698,
12
+ "acc_norm": 0.5592511451902011,
13
+ "acc_norm_stderr": 0.004954622308739005
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.47953216374269003,
17
+ "acc_stderr": 0.0383161053282193,
18
+ "acc_norm": 0.47953216374269003,
19
+ "acc_norm_stderr": 0.0383161053282193
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.46601941747572817,
23
+ "acc_stderr": 0.04939291447273481,
24
+ "acc_norm": 0.46601941747572817,
25
+ "acc_norm_stderr": 0.04939291447273481
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5146871008939975,
29
+ "acc_stderr": 0.017872248024429122,
30
+ "acc_norm": 0.5146871008939975,
31
+ "acc_norm_stderr": 0.017872248024429122
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.45185185185185184,
35
+ "acc_stderr": 0.04299268905480863,
36
+ "acc_norm": 0.45185185185185184,
37
+ "acc_norm_stderr": 0.04299268905480863
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.29,
41
+ "acc_stderr": 0.045604802157206824,
42
+ "acc_norm": 0.29,
43
+ "acc_norm_stderr": 0.045604802157206824
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.4297872340425532,
47
+ "acc_stderr": 0.03236214467715564,
48
+ "acc_norm": 0.4297872340425532,
49
+ "acc_norm_stderr": 0.03236214467715564
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.40963855421686746,
53
+ "acc_stderr": 0.038284011150790206,
54
+ "acc_norm": 0.40963855421686746,
55
+ "acc_norm_stderr": 0.038284011150790206
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4887459807073955,
59
+ "acc_stderr": 0.028390897396863537,
60
+ "acc_norm": 0.4887459807073955,
61
+ "acc_norm_stderr": 0.028390897396863537
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.515695067264574,
65
+ "acc_stderr": 0.0335412657542081,
66
+ "acc_norm": 0.515695067264574,
67
+ "acc_norm_stderr": 0.0335412657542081
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4351145038167939,
71
+ "acc_stderr": 0.04348208051644858,
72
+ "acc_norm": 0.4351145038167939,
73
+ "acc_norm_stderr": 0.04348208051644858
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.39,
77
+ "acc_stderr": 0.04902071300001974,
78
+ "acc_norm": 0.39,
79
+ "acc_norm_stderr": 0.04902071300001974
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5050505050505051,
83
+ "acc_stderr": 0.035621707606254015,
84
+ "acc_norm": 0.5050505050505051,
85
+ "acc_norm_stderr": 0.035621707606254015
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.35172413793103446,
89
+ "acc_stderr": 0.03979236637497411,
90
+ "acc_norm": 0.35172413793103446,
91
+ "acc_norm_stderr": 0.03979236637497411
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.21568627450980393,
95
+ "acc_stderr": 0.04092563958237655,
96
+ "acc_norm": 0.21568627450980393,
97
+ "acc_norm_stderr": 0.04092563958237655
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.3697478991596639,
101
+ "acc_stderr": 0.031357095996135904,
102
+ "acc_norm": 0.3697478991596639,
103
+ "acc_norm_stderr": 0.031357095996135904
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.38974358974358975,
107
+ "acc_stderr": 0.024726967886647074,
108
+ "acc_norm": 0.38974358974358975,
109
+ "acc_norm_stderr": 0.024726967886647074
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.52,
113
+ "acc_stderr": 0.050211673156867795,
114
+ "acc_norm": 0.52,
115
+ "acc_norm_stderr": 0.050211673156867795
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.36,
119
+ "acc_stderr": 0.04824181513244218,
120
+ "acc_norm": 0.36,
121
+ "acc_norm_stderr": 0.04824181513244218
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5185185185185185,
125
+ "acc_stderr": 0.04830366024635331,
126
+ "acc_norm": 0.5185185185185185,
127
+ "acc_norm_stderr": 0.04830366024635331
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3842364532019704,
131
+ "acc_stderr": 0.0342239856565755,
132
+ "acc_norm": 0.3842364532019704,
133
+ "acc_norm_stderr": 0.0342239856565755
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.43548387096774194,
137
+ "acc_stderr": 0.028206225591502737,
138
+ "acc_norm": 0.43548387096774194,
139
+ "acc_norm_stderr": 0.028206225591502737
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6410256410256411,
143
+ "acc_stderr": 0.03142616993791924,
144
+ "acc_norm": 0.6410256410256411,
145
+ "acc_norm_stderr": 0.03142616993791924
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.43018867924528303,
149
+ "acc_stderr": 0.030471445867183238,
150
+ "acc_norm": 0.43018867924528303,
151
+ "acc_norm_stderr": 0.030471445867183238
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.44545454545454544,
155
+ "acc_stderr": 0.047605488214603246,
156
+ "acc_norm": 0.44545454545454544,
157
+ "acc_norm_stderr": 0.047605488214603246
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.24444444444444444,
161
+ "acc_stderr": 0.026202766534652148,
162
+ "acc_norm": 0.24444444444444444,
163
+ "acc_norm_stderr": 0.026202766534652148
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2582781456953642,
167
+ "acc_stderr": 0.035737053147634576,
168
+ "acc_norm": 0.2582781456953642,
169
+ "acc_norm_stderr": 0.035737053147634576
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.527363184079602,
173
+ "acc_stderr": 0.03530235517334682,
174
+ "acc_norm": 0.527363184079602,
175
+ "acc_norm_stderr": 0.03530235517334682
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3815028901734104,
179
+ "acc_stderr": 0.03703851193099522,
180
+ "acc_norm": 0.3815028901734104,
181
+ "acc_norm_stderr": 0.03703851193099522
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2698412698412698,
185
+ "acc_stderr": 0.022860838309232072,
186
+ "acc_norm": 0.2698412698412698,
187
+ "acc_norm_stderr": 0.022860838309232072
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3333333333333333,
191
+ "acc_stderr": 0.039420826399272135,
192
+ "acc_norm": 0.3333333333333333,
193
+ "acc_norm_stderr": 0.039420826399272135
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.33,
197
+ "acc_stderr": 0.04725815626252606,
198
+ "acc_norm": 0.33,
199
+ "acc_norm_stderr": 0.04725815626252606
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.61,
203
+ "acc_stderr": 0.04902071300001975,
204
+ "acc_norm": 0.61,
205
+ "acc_norm_stderr": 0.04902071300001975
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.48265895953757226,
209
+ "acc_stderr": 0.026902900458666647,
210
+ "acc_norm": 0.48265895953757226,
211
+ "acc_norm_stderr": 0.026902900458666647
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.44171779141104295,
215
+ "acc_stderr": 0.03901591825836184,
216
+ "acc_norm": 0.44171779141104295,
217
+ "acc_norm_stderr": 0.03901591825836184
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.44753086419753085,
221
+ "acc_stderr": 0.027667138569422708,
222
+ "acc_norm": 0.44753086419753085,
223
+ "acc_norm_stderr": 0.027667138569422708
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.29,
227
+ "acc_stderr": 0.045604802157206845,
228
+ "acc_norm": 0.29,
229
+ "acc_norm_stderr": 0.045604802157206845
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.49740932642487046,
233
+ "acc_stderr": 0.03608390745384486,
234
+ "acc_norm": 0.49740932642487046,
235
+ "acc_norm_stderr": 0.03608390745384486
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2719298245614035,
239
+ "acc_stderr": 0.041857744240220575,
240
+ "acc_norm": 0.2719298245614035,
241
+ "acc_norm_stderr": 0.041857744240220575
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.48807339449541287,
245
+ "acc_stderr": 0.021431223617362233,
246
+ "acc_norm": 0.48807339449541287,
247
+ "acc_norm_stderr": 0.021431223617362233
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.23809523809523808,
251
+ "acc_stderr": 0.038095238095238106,
252
+ "acc_norm": 0.23809523809523808,
253
+ "acc_norm_stderr": 0.038095238095238106
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4019607843137255,
257
+ "acc_stderr": 0.028074158947600663,
258
+ "acc_norm": 0.4019607843137255,
259
+ "acc_norm_stderr": 0.028074158947600663
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.43,
263
+ "acc_stderr": 0.049756985195624284,
264
+ "acc_norm": 0.43,
265
+ "acc_norm_stderr": 0.049756985195624284
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.5371900826446281,
269
+ "acc_stderr": 0.04551711196104218,
270
+ "acc_norm": 0.5371900826446281,
271
+ "acc_norm_stderr": 0.04551711196104218
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3355263157894737,
275
+ "acc_stderr": 0.038424985593952694,
276
+ "acc_norm": 0.3355263157894737,
277
+ "acc_norm_stderr": 0.038424985593952694
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.37254901960784315,
281
+ "acc_stderr": 0.019559646809215923,
282
+ "acc_norm": 0.37254901960784315,
283
+ "acc_norm_stderr": 0.019559646809215923
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.30141843971631205,
287
+ "acc_stderr": 0.02737412888263115,
288
+ "acc_norm": 0.30141843971631205,
289
+ "acc_norm_stderr": 0.02737412888263115
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25892857142857145,
293
+ "acc_stderr": 0.04157751539865629,
294
+ "acc_norm": 0.25892857142857145,
295
+ "acc_norm_stderr": 0.04157751539865629
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.2777777777777778,
299
+ "acc_stderr": 0.030546745264953195,
300
+ "acc_norm": 0.2777777777777778,
301
+ "acc_norm_stderr": 0.030546745264953195
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2435754189944134,
305
+ "acc_stderr": 0.014355911964767857,
306
+ "acc_norm": 0.2435754189944134,
307
+ "acc_norm_stderr": 0.014355911964767857
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.33,
311
+ "acc_stderr": 0.04725815626252605,
312
+ "acc_norm": 0.33,
313
+ "acc_norm_stderr": 0.04725815626252605
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.43,
317
+ "acc_stderr": 0.049756985195624284,
318
+ "acc_norm": 0.43,
319
+ "acc_norm_stderr": 0.049756985195624284
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.33088235294117646,
323
+ "acc_stderr": 0.02858270975389843,
324
+ "acc_norm": 0.33088235294117646,
325
+ "acc_norm_stderr": 0.02858270975389843
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.44081632653061226,
329
+ "acc_stderr": 0.03178419114175364,
330
+ "acc_norm": 0.44081632653061226,
331
+ "acc_norm_stderr": 0.03178419114175364
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5569620253164557,
335
+ "acc_stderr": 0.03233532777533484,
336
+ "acc_norm": 0.5569620253164557,
337
+ "acc_norm_stderr": 0.03233532777533484
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.32529335071707954,
341
+ "acc_stderr": 0.011965311536571528,
342
+ "acc_norm": 0.32529335071707954,
343
+ "acc_norm_stderr": 0.011965311536571528
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4362745098039216,
347
+ "acc_stderr": 0.03480693138457039,
348
+ "acc_norm": 0.4362745098039216,
349
+ "acc_norm_stderr": 0.03480693138457039
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5333333333333333,
353
+ "acc_stderr": 0.03895658065271846,
354
+ "acc_norm": 0.5333333333333333,
355
+ "acc_norm_stderr": 0.03895658065271846
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.3084455324357405,
359
+ "mc1_stderr": 0.01616803938315687,
360
+ "mc2": 0.4532384559135145,
361
+ "mc2_stderr": 0.015485047009493541
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.4793388429752066,
365
+ "acc_stderr": 0.017175671279836446,
366
+ "acc_norm": 0.5301062573789846,
367
+ "acc_norm_stderr": 0.017159163590170223
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-instruct-v7-13B",
436
+ "model_sha": "40b2f1775ec5f92bfa8191fda6bb5f7c78564b3c",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-platypus-13B-v2/result_2023-10-30 01:04:14.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3890784982935154,
5
+ "acc_stderr": 0.014247309976045607,
6
+ "acc_norm": 0.4496587030716723,
7
+ "acc_norm_stderr": 0.01453714444428474
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.42013543118900615,
11
+ "acc_stderr": 0.00492571700809971,
12
+ "acc_norm": 0.5487950607448715,
13
+ "acc_norm_stderr": 0.004965963647210315
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5555555555555556,
17
+ "acc_stderr": 0.03811079669833531,
18
+ "acc_norm": 0.5555555555555556,
19
+ "acc_norm_stderr": 0.03811079669833531
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5145631067961165,
23
+ "acc_stderr": 0.04948637324026637,
24
+ "acc_norm": 0.5145631067961165,
25
+ "acc_norm_stderr": 0.04948637324026637
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4878671775223499,
29
+ "acc_stderr": 0.01787469866749135,
30
+ "acc_norm": 0.4878671775223499,
31
+ "acc_norm_stderr": 0.01787469866749135
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.37037037037037035,
35
+ "acc_stderr": 0.04171654161354544,
36
+ "acc_norm": 0.37037037037037035,
37
+ "acc_norm_stderr": 0.04171654161354544
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.24,
41
+ "acc_stderr": 0.04292346959909284,
42
+ "acc_norm": 0.24,
43
+ "acc_norm_stderr": 0.04292346959909284
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.3276595744680851,
47
+ "acc_stderr": 0.030683020843231015,
48
+ "acc_norm": 0.3276595744680851,
49
+ "acc_norm_stderr": 0.030683020843231015
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3795180722891566,
53
+ "acc_stderr": 0.03777798822748018,
54
+ "acc_norm": 0.3795180722891566,
55
+ "acc_norm_stderr": 0.03777798822748018
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5048231511254019,
59
+ "acc_stderr": 0.02839677044411129,
60
+ "acc_norm": 0.5048231511254019,
61
+ "acc_norm_stderr": 0.02839677044411129
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.45739910313901344,
65
+ "acc_stderr": 0.033435777055830646,
66
+ "acc_norm": 0.45739910313901344,
67
+ "acc_norm_stderr": 0.033435777055830646
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.48091603053435117,
71
+ "acc_stderr": 0.04382094705550988,
72
+ "acc_norm": 0.48091603053435117,
73
+ "acc_norm_stderr": 0.04382094705550988
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.39,
77
+ "acc_stderr": 0.04902071300001975,
78
+ "acc_norm": 0.39,
79
+ "acc_norm_stderr": 0.04902071300001975
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5050505050505051,
83
+ "acc_stderr": 0.035621707606254015,
84
+ "acc_norm": 0.5050505050505051,
85
+ "acc_norm_stderr": 0.035621707606254015
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.3724137931034483,
89
+ "acc_stderr": 0.0402873153294756,
90
+ "acc_norm": 0.3724137931034483,
91
+ "acc_norm_stderr": 0.0402873153294756
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.22549019607843138,
95
+ "acc_stderr": 0.04158307533083286,
96
+ "acc_norm": 0.22549019607843138,
97
+ "acc_norm_stderr": 0.04158307533083286
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.46638655462184875,
101
+ "acc_stderr": 0.03240501447690071,
102
+ "acc_norm": 0.46638655462184875,
103
+ "acc_norm_stderr": 0.03240501447690071
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4307692307692308,
107
+ "acc_stderr": 0.025106820660539753,
108
+ "acc_norm": 0.4307692307692308,
109
+ "acc_norm_stderr": 0.025106820660539753
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.54,
113
+ "acc_stderr": 0.05009082659620332,
114
+ "acc_norm": 0.54,
115
+ "acc_norm_stderr": 0.05009082659620332
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.31,
119
+ "acc_stderr": 0.04648231987117316,
120
+ "acc_norm": 0.31,
121
+ "acc_norm_stderr": 0.04648231987117316
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.49074074074074076,
125
+ "acc_stderr": 0.04832853553437055,
126
+ "acc_norm": 0.49074074074074076,
127
+ "acc_norm_stderr": 0.04832853553437055
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3399014778325123,
131
+ "acc_stderr": 0.033327690684107895,
132
+ "acc_norm": 0.3399014778325123,
133
+ "acc_norm_stderr": 0.033327690684107895
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.44516129032258067,
137
+ "acc_stderr": 0.02827241018621491,
138
+ "acc_norm": 0.44516129032258067,
139
+ "acc_norm_stderr": 0.02827241018621491
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6452991452991453,
143
+ "acc_stderr": 0.03134250486245402,
144
+ "acc_norm": 0.6452991452991453,
145
+ "acc_norm_stderr": 0.03134250486245402
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.39245283018867927,
149
+ "acc_stderr": 0.030052580579557845,
150
+ "acc_norm": 0.39245283018867927,
151
+ "acc_norm_stderr": 0.030052580579557845
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4,
155
+ "acc_stderr": 0.0469237132203465,
156
+ "acc_norm": 0.4,
157
+ "acc_norm_stderr": 0.0469237132203465
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.22592592592592592,
161
+ "acc_stderr": 0.025497532639609553,
162
+ "acc_norm": 0.22592592592592592,
163
+ "acc_norm_stderr": 0.025497532639609553
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2847682119205298,
167
+ "acc_stderr": 0.03684881521389023,
168
+ "acc_norm": 0.2847682119205298,
169
+ "acc_norm_stderr": 0.03684881521389023
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5970149253731343,
173
+ "acc_stderr": 0.034683432951111266,
174
+ "acc_norm": 0.5970149253731343,
175
+ "acc_norm_stderr": 0.034683432951111266
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3988439306358382,
179
+ "acc_stderr": 0.037336266553835096,
180
+ "acc_norm": 0.3988439306358382,
181
+ "acc_norm_stderr": 0.037336266553835096
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.31216931216931215,
185
+ "acc_stderr": 0.023865206836972585,
186
+ "acc_norm": 0.31216931216931215,
187
+ "acc_norm_stderr": 0.023865206836972585
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3402777777777778,
191
+ "acc_stderr": 0.03962135573486219,
192
+ "acc_norm": 0.3402777777777778,
193
+ "acc_norm_stderr": 0.03962135573486219
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.31,
197
+ "acc_stderr": 0.04648231987117316,
198
+ "acc_norm": 0.31,
199
+ "acc_norm_stderr": 0.04648231987117316
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.64,
203
+ "acc_stderr": 0.04824181513244218,
204
+ "acc_norm": 0.64,
205
+ "acc_norm_stderr": 0.04824181513244218
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5,
209
+ "acc_stderr": 0.026919095102908273,
210
+ "acc_norm": 0.5,
211
+ "acc_norm_stderr": 0.026919095102908273
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.44171779141104295,
215
+ "acc_stderr": 0.03901591825836185,
216
+ "acc_norm": 0.44171779141104295,
217
+ "acc_norm_stderr": 0.03901591825836185
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.46296296296296297,
221
+ "acc_stderr": 0.027744313443376536,
222
+ "acc_norm": 0.46296296296296297,
223
+ "acc_norm_stderr": 0.027744313443376536
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.36,
227
+ "acc_stderr": 0.04824181513244218,
228
+ "acc_norm": 0.36,
229
+ "acc_norm_stderr": 0.04824181513244218
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.49740932642487046,
233
+ "acc_stderr": 0.03608390745384487,
234
+ "acc_norm": 0.49740932642487046,
235
+ "acc_norm_stderr": 0.03608390745384487
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.21929824561403508,
239
+ "acc_stderr": 0.03892431106518754,
240
+ "acc_norm": 0.21929824561403508,
241
+ "acc_norm_stderr": 0.03892431106518754
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5009174311926605,
245
+ "acc_stderr": 0.021437287056051215,
246
+ "acc_norm": 0.5009174311926605,
247
+ "acc_norm_stderr": 0.021437287056051215
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.25396825396825395,
251
+ "acc_stderr": 0.038932596106046734,
252
+ "acc_norm": 0.25396825396825395,
253
+ "acc_norm_stderr": 0.038932596106046734
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.46078431372549017,
257
+ "acc_stderr": 0.028541722692618874,
258
+ "acc_norm": 0.46078431372549017,
259
+ "acc_norm_stderr": 0.028541722692618874
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.51,
263
+ "acc_stderr": 0.05024183937956911,
264
+ "acc_norm": 0.51,
265
+ "acc_norm_stderr": 0.05024183937956911
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6528925619834711,
269
+ "acc_stderr": 0.04345724570292534,
270
+ "acc_norm": 0.6528925619834711,
271
+ "acc_norm_stderr": 0.04345724570292534
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.48026315789473684,
275
+ "acc_stderr": 0.040657710025626057,
276
+ "acc_norm": 0.48026315789473684,
277
+ "acc_norm_stderr": 0.040657710025626057
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.35784313725490197,
281
+ "acc_stderr": 0.019393058402355435,
282
+ "acc_norm": 0.35784313725490197,
283
+ "acc_norm_stderr": 0.019393058402355435
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3191489361702128,
287
+ "acc_stderr": 0.027807990141320203,
288
+ "acc_norm": 0.3191489361702128,
289
+ "acc_norm_stderr": 0.027807990141320203
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.22321428571428573,
293
+ "acc_stderr": 0.039523019677025116,
294
+ "acc_norm": 0.22321428571428573,
295
+ "acc_norm_stderr": 0.039523019677025116
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.27314814814814814,
299
+ "acc_stderr": 0.030388051301678116,
300
+ "acc_norm": 0.27314814814814814,
301
+ "acc_norm_stderr": 0.030388051301678116
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2424581005586592,
305
+ "acc_stderr": 0.01433352205921789,
306
+ "acc_norm": 0.2424581005586592,
307
+ "acc_norm_stderr": 0.01433352205921789
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.36,
311
+ "acc_stderr": 0.048241815132442176,
312
+ "acc_norm": 0.36,
313
+ "acc_norm_stderr": 0.048241815132442176
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.44,
317
+ "acc_stderr": 0.04988876515698589,
318
+ "acc_norm": 0.44,
319
+ "acc_norm_stderr": 0.04988876515698589
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.2757352941176471,
323
+ "acc_stderr": 0.027146271936625166,
324
+ "acc_norm": 0.2757352941176471,
325
+ "acc_norm_stderr": 0.027146271936625166
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4489795918367347,
329
+ "acc_stderr": 0.03184213866687579,
330
+ "acc_norm": 0.4489795918367347,
331
+ "acc_norm_stderr": 0.03184213866687579
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.569620253164557,
335
+ "acc_stderr": 0.03223017195937598,
336
+ "acc_norm": 0.569620253164557,
337
+ "acc_norm_stderr": 0.03223017195937598
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3135593220338983,
341
+ "acc_stderr": 0.011849234291459313,
342
+ "acc_norm": 0.3135593220338983,
343
+ "acc_norm_stderr": 0.011849234291459313
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4950980392156863,
347
+ "acc_stderr": 0.03509143375606785,
348
+ "acc_norm": 0.4950980392156863,
349
+ "acc_norm_stderr": 0.03509143375606785
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5515151515151515,
353
+ "acc_stderr": 0.03883565977956928,
354
+ "acc_norm": 0.5515151515151515,
355
+ "acc_norm_stderr": 0.03883565977956928
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.26805385556915545,
359
+ "mc1_stderr": 0.015506204722834562,
360
+ "mc2": 0.44689474709496685,
361
+ "mc2_stderr": 0.015256070107718848
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5478158205430933,
365
+ "acc_stderr": 0.017111567130916796,
366
+ "acc_norm": 0.5962219598583235,
367
+ "acc_norm_stderr": 0.016869031540298632
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-platypus-13B-v2",
436
+ "model_sha": "1b4eb6319be99c113d17778ce2737acffe2a0fee",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/GenAI-llama2-ko-en-platypus-13B/result_2023-10-29 00:31:00.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3839590443686007,
5
+ "acc_stderr": 0.01421244498065189,
6
+ "acc_norm": 0.4522184300341297,
7
+ "acc_norm_stderr": 0.014544519880633832
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.4166500697072296,
11
+ "acc_stderr": 0.00491996282220832,
12
+ "acc_norm": 0.5524795857398924,
13
+ "acc_norm_stderr": 0.004962220512548352
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5614035087719298,
17
+ "acc_stderr": 0.038057975055904594,
18
+ "acc_norm": 0.5614035087719298,
19
+ "acc_norm_stderr": 0.038057975055904594
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.49514563106796117,
23
+ "acc_stderr": 0.049505043821289195,
24
+ "acc_norm": 0.49514563106796117,
25
+ "acc_norm_stderr": 0.049505043821289195
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5351213282247765,
29
+ "acc_stderr": 0.017835798806290642,
30
+ "acc_norm": 0.5351213282247765,
31
+ "acc_norm_stderr": 0.017835798806290642
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.34814814814814815,
35
+ "acc_stderr": 0.041153246103369526,
36
+ "acc_norm": 0.34814814814814815,
37
+ "acc_norm_stderr": 0.041153246103369526
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.24,
41
+ "acc_stderr": 0.042923469599092816,
42
+ "acc_norm": 0.24,
43
+ "acc_norm_stderr": 0.042923469599092816
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.31063829787234043,
47
+ "acc_stderr": 0.03025123757921317,
48
+ "acc_norm": 0.31063829787234043,
49
+ "acc_norm_stderr": 0.03025123757921317
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.42168674698795183,
53
+ "acc_stderr": 0.038444531817709175,
54
+ "acc_norm": 0.42168674698795183,
55
+ "acc_norm_stderr": 0.038444531817709175
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5048231511254019,
59
+ "acc_stderr": 0.02839677044411129,
60
+ "acc_norm": 0.5048231511254019,
61
+ "acc_norm_stderr": 0.02839677044411129
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.4484304932735426,
65
+ "acc_stderr": 0.03337883736255099,
66
+ "acc_norm": 0.4484304932735426,
67
+ "acc_norm_stderr": 0.03337883736255099
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.5267175572519084,
71
+ "acc_stderr": 0.04379024936553894,
72
+ "acc_norm": 0.5267175572519084,
73
+ "acc_norm_stderr": 0.04379024936553894
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.43,
77
+ "acc_stderr": 0.04975698519562429,
78
+ "acc_norm": 0.43,
79
+ "acc_norm_stderr": 0.04975698519562429
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.494949494949495,
83
+ "acc_stderr": 0.035621707606254015,
84
+ "acc_norm": 0.494949494949495,
85
+ "acc_norm_stderr": 0.035621707606254015
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4068965517241379,
89
+ "acc_stderr": 0.04093793981266236,
90
+ "acc_norm": 0.4068965517241379,
91
+ "acc_norm_stderr": 0.04093793981266236
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.17647058823529413,
95
+ "acc_stderr": 0.0379328118530781,
96
+ "acc_norm": 0.17647058823529413,
97
+ "acc_norm_stderr": 0.0379328118530781
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.4957983193277311,
101
+ "acc_stderr": 0.03247734334448111,
102
+ "acc_norm": 0.4957983193277311,
103
+ "acc_norm_stderr": 0.03247734334448111
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4230769230769231,
107
+ "acc_stderr": 0.025049197876042328,
108
+ "acc_norm": 0.4230769230769231,
109
+ "acc_norm_stderr": 0.025049197876042328
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.46,
113
+ "acc_stderr": 0.05009082659620332,
114
+ "acc_norm": 0.46,
115
+ "acc_norm_stderr": 0.05009082659620332
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.28,
119
+ "acc_stderr": 0.04512608598542128,
120
+ "acc_norm": 0.28,
121
+ "acc_norm_stderr": 0.04512608598542128
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.46296296296296297,
125
+ "acc_stderr": 0.04820403072760628,
126
+ "acc_norm": 0.46296296296296297,
127
+ "acc_norm_stderr": 0.04820403072760628
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.35960591133004927,
131
+ "acc_stderr": 0.03376458246509568,
132
+ "acc_norm": 0.35960591133004927,
133
+ "acc_norm_stderr": 0.03376458246509568
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.43548387096774194,
137
+ "acc_stderr": 0.02820622559150275,
138
+ "acc_norm": 0.43548387096774194,
139
+ "acc_norm_stderr": 0.02820622559150275
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6581196581196581,
143
+ "acc_stderr": 0.03107502852650775,
144
+ "acc_norm": 0.6581196581196581,
145
+ "acc_norm_stderr": 0.03107502852650775
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4490566037735849,
149
+ "acc_stderr": 0.030612730713641095,
150
+ "acc_norm": 0.4490566037735849,
151
+ "acc_norm_stderr": 0.030612730713641095
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.41818181818181815,
155
+ "acc_stderr": 0.04724577405731572,
156
+ "acc_norm": 0.41818181818181815,
157
+ "acc_norm_stderr": 0.04724577405731572
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.23333333333333334,
161
+ "acc_stderr": 0.02578787422095932,
162
+ "acc_norm": 0.23333333333333334,
163
+ "acc_norm_stderr": 0.02578787422095932
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2052980132450331,
167
+ "acc_stderr": 0.032979866484738336,
168
+ "acc_norm": 0.2052980132450331,
169
+ "acc_norm_stderr": 0.032979866484738336
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5771144278606966,
173
+ "acc_stderr": 0.034932317774212816,
174
+ "acc_norm": 0.5771144278606966,
175
+ "acc_norm_stderr": 0.034932317774212816
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3699421965317919,
179
+ "acc_stderr": 0.036812296333943194,
180
+ "acc_norm": 0.3699421965317919,
181
+ "acc_norm_stderr": 0.036812296333943194
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2830687830687831,
185
+ "acc_stderr": 0.023201392938194978,
186
+ "acc_norm": 0.2830687830687831,
187
+ "acc_norm_stderr": 0.023201392938194978
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4166666666666667,
191
+ "acc_stderr": 0.04122728707651282,
192
+ "acc_norm": 0.4166666666666667,
193
+ "acc_norm_stderr": 0.04122728707651282
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.28,
197
+ "acc_stderr": 0.04512608598542128,
198
+ "acc_norm": 0.28,
199
+ "acc_norm_stderr": 0.04512608598542128
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.66,
203
+ "acc_stderr": 0.04760952285695237,
204
+ "acc_norm": 0.66,
205
+ "acc_norm_stderr": 0.04760952285695237
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5057803468208093,
209
+ "acc_stderr": 0.02691729617914911,
210
+ "acc_norm": 0.5057803468208093,
211
+ "acc_norm_stderr": 0.02691729617914911
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.44785276073619634,
215
+ "acc_stderr": 0.03906947479456602,
216
+ "acc_norm": 0.44785276073619634,
217
+ "acc_norm_stderr": 0.03906947479456602
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4722222222222222,
221
+ "acc_stderr": 0.027777777777777797,
222
+ "acc_norm": 0.4722222222222222,
223
+ "acc_norm_stderr": 0.027777777777777797
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.31,
227
+ "acc_stderr": 0.04648231987117316,
228
+ "acc_norm": 0.31,
229
+ "acc_norm_stderr": 0.04648231987117316
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5647668393782384,
233
+ "acc_stderr": 0.03578038165008586,
234
+ "acc_norm": 0.5647668393782384,
235
+ "acc_norm_stderr": 0.03578038165008586
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.23684210526315788,
239
+ "acc_stderr": 0.03999423879281335,
240
+ "acc_norm": 0.23684210526315788,
241
+ "acc_norm_stderr": 0.03999423879281335
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5522935779816514,
245
+ "acc_stderr": 0.02131975496242546,
246
+ "acc_norm": 0.5522935779816514,
247
+ "acc_norm_stderr": 0.02131975496242546
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.2777777777777778,
251
+ "acc_stderr": 0.040061680838488774,
252
+ "acc_norm": 0.2777777777777778,
253
+ "acc_norm_stderr": 0.040061680838488774
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.45751633986928103,
257
+ "acc_stderr": 0.02852638345214264,
258
+ "acc_norm": 0.45751633986928103,
259
+ "acc_norm_stderr": 0.02852638345214264
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.48,
263
+ "acc_stderr": 0.050211673156867795,
264
+ "acc_norm": 0.48,
265
+ "acc_norm_stderr": 0.050211673156867795
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6859504132231405,
269
+ "acc_stderr": 0.04236964753041018,
270
+ "acc_norm": 0.6859504132231405,
271
+ "acc_norm_stderr": 0.04236964753041018
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.45394736842105265,
275
+ "acc_stderr": 0.04051646342874141,
276
+ "acc_norm": 0.45394736842105265,
277
+ "acc_norm_stderr": 0.04051646342874141
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3839869281045752,
281
+ "acc_stderr": 0.019675808135281525,
282
+ "acc_norm": 0.3839869281045752,
283
+ "acc_norm_stderr": 0.019675808135281525
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.35815602836879434,
287
+ "acc_stderr": 0.02860208586275942,
288
+ "acc_norm": 0.35815602836879434,
289
+ "acc_norm_stderr": 0.02860208586275942
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.2857142857142857,
293
+ "acc_stderr": 0.042878587513404544,
294
+ "acc_norm": 0.2857142857142857,
295
+ "acc_norm_stderr": 0.042878587513404544
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.25462962962962965,
299
+ "acc_stderr": 0.02971127586000534,
300
+ "acc_norm": 0.25462962962962965,
301
+ "acc_norm_stderr": 0.02971127586000534
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.293854748603352,
305
+ "acc_stderr": 0.015235075776719616,
306
+ "acc_norm": 0.293854748603352,
307
+ "acc_norm_stderr": 0.015235075776719616
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.32,
311
+ "acc_stderr": 0.046882617226215034,
312
+ "acc_norm": 0.32,
313
+ "acc_norm_stderr": 0.046882617226215034
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.41,
317
+ "acc_stderr": 0.049431107042371025,
318
+ "acc_norm": 0.41,
319
+ "acc_norm_stderr": 0.049431107042371025
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.27941176470588236,
323
+ "acc_stderr": 0.02725720260611495,
324
+ "acc_norm": 0.27941176470588236,
325
+ "acc_norm_stderr": 0.02725720260611495
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4775510204081633,
329
+ "acc_stderr": 0.031976941187136725,
330
+ "acc_norm": 0.4775510204081633,
331
+ "acc_norm_stderr": 0.031976941187136725
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6075949367088608,
335
+ "acc_stderr": 0.0317847187456473,
336
+ "acc_norm": 0.6075949367088608,
337
+ "acc_norm_stderr": 0.0317847187456473
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.33116036505867014,
341
+ "acc_stderr": 0.01202012819598576,
342
+ "acc_norm": 0.33116036505867014,
343
+ "acc_norm_stderr": 0.01202012819598576
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.49019607843137253,
347
+ "acc_stderr": 0.03508637358630572,
348
+ "acc_norm": 0.49019607843137253,
349
+ "acc_norm_stderr": 0.03508637358630572
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5454545454545454,
353
+ "acc_stderr": 0.038881769216741004,
354
+ "acc_norm": 0.5454545454545454,
355
+ "acc_norm_stderr": 0.038881769216741004
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.27539779681762544,
359
+ "mc1_stderr": 0.015638135667775523,
360
+ "mc2": 0.4478444454695957,
361
+ "mc2_stderr": 0.015296142940086415
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5454545454545454,
365
+ "acc_stderr": 0.017119172208061504,
366
+ "acc_norm": 0.5938606847697757,
367
+ "acc_norm_stderr": 0.016884749503191396
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/GenAI-llama2-ko-en-platypus-13B",
436
+ "model_sha": "61d276d0715184790bae2979744f1ae7c0f451c0",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/llama-2-ko-7b-instruct/result_2023-09-29 09:41:36.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3455631399317406,
5
+ "acc_stderr": 0.013896938461145678,
6
+ "acc_norm": 0.3839590443686007,
7
+ "acc_norm_stderr": 0.01421244498065189
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3975303724357698,
11
+ "acc_stderr": 0.004883871774350598,
12
+ "acc_norm": 0.5247958573989245,
13
+ "acc_norm_stderr": 0.004983641854351152
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.3684210526315789,
17
+ "acc_stderr": 0.036996580176568775,
18
+ "acc_norm": 0.3684210526315789,
19
+ "acc_norm_stderr": 0.036996580176568775
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.32038834951456313,
23
+ "acc_stderr": 0.0462028408228004,
24
+ "acc_norm": 0.32038834951456313,
25
+ "acc_norm_stderr": 0.0462028408228004
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.3831417624521073,
29
+ "acc_stderr": 0.01738477419488563,
30
+ "acc_norm": 0.3831417624521073,
31
+ "acc_norm_stderr": 0.01738477419488563
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.32592592592592595,
35
+ "acc_stderr": 0.040491220417025055,
36
+ "acc_norm": 0.32592592592592595,
37
+ "acc_norm_stderr": 0.040491220417025055
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.28,
41
+ "acc_stderr": 0.045126085985421255,
42
+ "acc_norm": 0.28,
43
+ "acc_norm_stderr": 0.045126085985421255
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.28085106382978725,
47
+ "acc_stderr": 0.02937917046412482,
48
+ "acc_norm": 0.28085106382978725,
49
+ "acc_norm_stderr": 0.02937917046412482
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3795180722891566,
53
+ "acc_stderr": 0.03777798822748017,
54
+ "acc_norm": 0.3795180722891566,
55
+ "acc_norm_stderr": 0.03777798822748017
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.35691318327974275,
59
+ "acc_stderr": 0.027210420375934012,
60
+ "acc_norm": 0.35691318327974275,
61
+ "acc_norm_stderr": 0.027210420375934012
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.4170403587443946,
65
+ "acc_stderr": 0.03309266936071721,
66
+ "acc_norm": 0.4170403587443946,
67
+ "acc_norm_stderr": 0.03309266936071721
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4198473282442748,
71
+ "acc_stderr": 0.043285772152629715,
72
+ "acc_norm": 0.4198473282442748,
73
+ "acc_norm_stderr": 0.043285772152629715
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.28,
77
+ "acc_stderr": 0.04512608598542127,
78
+ "acc_norm": 0.28,
79
+ "acc_norm_stderr": 0.04512608598542127
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.35858585858585856,
83
+ "acc_stderr": 0.034169036403915214,
84
+ "acc_norm": 0.35858585858585856,
85
+ "acc_norm_stderr": 0.034169036403915214
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.25517241379310346,
89
+ "acc_stderr": 0.03632984052707842,
90
+ "acc_norm": 0.25517241379310346,
91
+ "acc_norm_stderr": 0.03632984052707842
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.1568627450980392,
95
+ "acc_stderr": 0.03618664819936245,
96
+ "acc_norm": 0.1568627450980392,
97
+ "acc_norm_stderr": 0.03618664819936245
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.25630252100840334,
101
+ "acc_stderr": 0.028359620870533953,
102
+ "acc_norm": 0.25630252100840334,
103
+ "acc_norm_stderr": 0.028359620870533953
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.24871794871794872,
107
+ "acc_stderr": 0.021916957709213803,
108
+ "acc_norm": 0.24871794871794872,
109
+ "acc_norm_stderr": 0.021916957709213803
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.36,
113
+ "acc_stderr": 0.048241815132442176,
114
+ "acc_norm": 0.36,
115
+ "acc_norm_stderr": 0.048241815132442176
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.33,
119
+ "acc_stderr": 0.04725815626252605,
120
+ "acc_norm": 0.33,
121
+ "acc_norm_stderr": 0.04725815626252605
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.37037037037037035,
125
+ "acc_stderr": 0.04668408033024932,
126
+ "acc_norm": 0.37037037037037035,
127
+ "acc_norm_stderr": 0.04668408033024932
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.23645320197044334,
131
+ "acc_stderr": 0.029896114291733545,
132
+ "acc_norm": 0.23645320197044334,
133
+ "acc_norm_stderr": 0.029896114291733545
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.3161290322580645,
137
+ "acc_stderr": 0.026450874489042767,
138
+ "acc_norm": 0.3161290322580645,
139
+ "acc_norm_stderr": 0.026450874489042767
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.4358974358974359,
143
+ "acc_stderr": 0.03248577511578401,
144
+ "acc_norm": 0.4358974358974359,
145
+ "acc_norm_stderr": 0.03248577511578401
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.30943396226415093,
149
+ "acc_stderr": 0.028450154794118627,
150
+ "acc_norm": 0.30943396226415093,
151
+ "acc_norm_stderr": 0.028450154794118627
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.37272727272727274,
155
+ "acc_stderr": 0.04631381319425463,
156
+ "acc_norm": 0.37272727272727274,
157
+ "acc_norm_stderr": 0.04631381319425463
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.25555555555555554,
161
+ "acc_stderr": 0.026593939101844072,
162
+ "acc_norm": 0.25555555555555554,
163
+ "acc_norm_stderr": 0.026593939101844072
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.23841059602649006,
167
+ "acc_stderr": 0.0347918557259966,
168
+ "acc_norm": 0.23841059602649006,
169
+ "acc_norm_stderr": 0.0347918557259966
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.3283582089552239,
173
+ "acc_stderr": 0.033206858897443244,
174
+ "acc_norm": 0.3283582089552239,
175
+ "acc_norm_stderr": 0.033206858897443244
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.23121387283236994,
179
+ "acc_stderr": 0.0321473730202947,
180
+ "acc_norm": 0.23121387283236994,
181
+ "acc_norm_stderr": 0.0321473730202947
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.23809523809523808,
185
+ "acc_stderr": 0.021935878081184756,
186
+ "acc_norm": 0.23809523809523808,
187
+ "acc_norm_stderr": 0.021935878081184756
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.2777777777777778,
191
+ "acc_stderr": 0.03745554791462457,
192
+ "acc_norm": 0.2777777777777778,
193
+ "acc_norm_stderr": 0.03745554791462457
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.21,
197
+ "acc_stderr": 0.040936018074033256,
198
+ "acc_norm": 0.21,
199
+ "acc_norm_stderr": 0.040936018074033256
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.37,
203
+ "acc_stderr": 0.04852365870939099,
204
+ "acc_norm": 0.37,
205
+ "acc_norm_stderr": 0.04852365870939099
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.33236994219653176,
209
+ "acc_stderr": 0.025361168749688225,
210
+ "acc_norm": 0.33236994219653176,
211
+ "acc_norm_stderr": 0.025361168749688225
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.25766871165644173,
215
+ "acc_stderr": 0.03436150827846917,
216
+ "acc_norm": 0.25766871165644173,
217
+ "acc_norm_stderr": 0.03436150827846917
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.33641975308641975,
221
+ "acc_stderr": 0.026289734945952926,
222
+ "acc_norm": 0.33641975308641975,
223
+ "acc_norm_stderr": 0.026289734945952926
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.28,
227
+ "acc_stderr": 0.04512608598542129,
228
+ "acc_norm": 0.28,
229
+ "acc_norm_stderr": 0.04512608598542129
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.32642487046632124,
233
+ "acc_stderr": 0.033840286211432945,
234
+ "acc_norm": 0.32642487046632124,
235
+ "acc_norm_stderr": 0.033840286211432945
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.30701754385964913,
239
+ "acc_stderr": 0.0433913832257986,
240
+ "acc_norm": 0.30701754385964913,
241
+ "acc_norm_stderr": 0.0433913832257986
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.3174311926605505,
245
+ "acc_stderr": 0.019957152198460497,
246
+ "acc_norm": 0.3174311926605505,
247
+ "acc_norm_stderr": 0.019957152198460497
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.1746031746031746,
251
+ "acc_stderr": 0.03395490020856111,
252
+ "acc_norm": 0.1746031746031746,
253
+ "acc_norm_stderr": 0.03395490020856111
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.35294117647058826,
257
+ "acc_stderr": 0.02736359328468495,
258
+ "acc_norm": 0.35294117647058826,
259
+ "acc_norm_stderr": 0.02736359328468495
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.25,
263
+ "acc_stderr": 0.04351941398892446,
264
+ "acc_norm": 0.25,
265
+ "acc_norm_stderr": 0.04351941398892446
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.36363636363636365,
269
+ "acc_stderr": 0.043913262867240704,
270
+ "acc_norm": 0.36363636363636365,
271
+ "acc_norm_stderr": 0.043913262867240704
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3026315789473684,
275
+ "acc_stderr": 0.037385206761196686,
276
+ "acc_norm": 0.3026315789473684,
277
+ "acc_norm_stderr": 0.037385206761196686
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.2679738562091503,
281
+ "acc_stderr": 0.017917974069594726,
282
+ "acc_norm": 0.2679738562091503,
283
+ "acc_norm_stderr": 0.017917974069594726
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.2624113475177305,
287
+ "acc_stderr": 0.02624492034984301,
288
+ "acc_norm": 0.2624113475177305,
289
+ "acc_norm_stderr": 0.02624492034984301
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.33035714285714285,
293
+ "acc_stderr": 0.04464285714285712,
294
+ "acc_norm": 0.33035714285714285,
295
+ "acc_norm_stderr": 0.04464285714285712
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.24074074074074073,
299
+ "acc_stderr": 0.029157522184605586,
300
+ "acc_norm": 0.24074074074074073,
301
+ "acc_norm_stderr": 0.029157522184605586
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2424581005586592,
305
+ "acc_stderr": 0.01433352205921789,
306
+ "acc_norm": 0.2424581005586592,
307
+ "acc_norm_stderr": 0.01433352205921789
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.32,
311
+ "acc_stderr": 0.04688261722621504,
312
+ "acc_norm": 0.32,
313
+ "acc_norm_stderr": 0.04688261722621504
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.21,
317
+ "acc_stderr": 0.040936018074033256,
318
+ "acc_norm": 0.21,
319
+ "acc_norm_stderr": 0.040936018074033256
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.35294117647058826,
323
+ "acc_stderr": 0.0290294228156814,
324
+ "acc_norm": 0.35294117647058826,
325
+ "acc_norm_stderr": 0.0290294228156814
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.27755102040816326,
329
+ "acc_stderr": 0.028666857790274648,
330
+ "acc_norm": 0.27755102040816326,
331
+ "acc_norm_stderr": 0.028666857790274648
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.35864978902953587,
335
+ "acc_stderr": 0.031219569445301847,
336
+ "acc_norm": 0.35864978902953587,
337
+ "acc_norm_stderr": 0.031219569445301847
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.27249022164276404,
341
+ "acc_stderr": 0.01137165829431153,
342
+ "acc_norm": 0.27249022164276404,
343
+ "acc_norm_stderr": 0.01137165829431153
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.2696078431372549,
347
+ "acc_stderr": 0.031145570659486782,
348
+ "acc_norm": 0.2696078431372549,
349
+ "acc_norm_stderr": 0.031145570659486782
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.3090909090909091,
353
+ "acc_stderr": 0.036085410115739666,
354
+ "acc_norm": 0.3090909090909091,
355
+ "acc_norm_stderr": 0.036085410115739666
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.24969400244798043,
359
+ "mc1_stderr": 0.015152286907148125,
360
+ "mc2": 0.39805148377575406,
361
+ "mc2_stderr": 0.015027401787198838
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.2396694214876033,
365
+ "acc_stderr": 0.014676495332267253,
366
+ "acc_norm": 0.31286894923258557,
367
+ "acc_norm_stderr": 0.015941010118302654
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/llama-2-ko-7b-instruct",
436
+ "model_sha": "3c590472282b5de4c76d846153db5f41b82c1b62",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/llama-2-ko-7b-instruction-v3/result_2023-10-01 18:41:33.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3293515358361775,
5
+ "acc_stderr": 0.013734057652635474,
6
+ "acc_norm": 0.386518771331058,
7
+ "acc_norm_stderr": 0.014230084761910474
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3835889265086636,
11
+ "acc_stderr": 0.00485265887677539,
12
+ "acc_norm": 0.5022903804023103,
13
+ "acc_norm_stderr": 0.004989729059957435
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.391812865497076,
17
+ "acc_stderr": 0.03743979825926401,
18
+ "acc_norm": 0.391812865497076,
19
+ "acc_norm_stderr": 0.03743979825926401
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.2815533980582524,
23
+ "acc_stderr": 0.04453254836326466,
24
+ "acc_norm": 0.2815533980582524,
25
+ "acc_norm_stderr": 0.04453254836326466
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.367816091954023,
29
+ "acc_stderr": 0.01724382889184626,
30
+ "acc_norm": 0.367816091954023,
31
+ "acc_norm_stderr": 0.01724382889184626
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.34074074074074073,
35
+ "acc_stderr": 0.04094376269996795,
36
+ "acc_norm": 0.34074074074074073,
37
+ "acc_norm_stderr": 0.04094376269996795
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.3,
41
+ "acc_stderr": 0.046056618647183814,
42
+ "acc_norm": 0.3,
43
+ "acc_norm_stderr": 0.046056618647183814
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.2851063829787234,
47
+ "acc_stderr": 0.02951319662553935,
48
+ "acc_norm": 0.2851063829787234,
49
+ "acc_norm_stderr": 0.02951319662553935
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3132530120481928,
53
+ "acc_stderr": 0.03610805018031024,
54
+ "acc_norm": 0.3132530120481928,
55
+ "acc_norm_stderr": 0.03610805018031024
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.3858520900321543,
59
+ "acc_stderr": 0.027648149599751464,
60
+ "acc_norm": 0.3858520900321543,
61
+ "acc_norm_stderr": 0.027648149599751464
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.43946188340807174,
65
+ "acc_stderr": 0.03331092511038179,
66
+ "acc_norm": 0.43946188340807174,
67
+ "acc_norm_stderr": 0.03331092511038179
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4351145038167939,
71
+ "acc_stderr": 0.04348208051644858,
72
+ "acc_norm": 0.4351145038167939,
73
+ "acc_norm_stderr": 0.04348208051644858
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.3,
77
+ "acc_stderr": 0.046056618647183814,
78
+ "acc_norm": 0.3,
79
+ "acc_norm_stderr": 0.046056618647183814
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.31313131313131315,
83
+ "acc_stderr": 0.03304205087813653,
84
+ "acc_norm": 0.31313131313131315,
85
+ "acc_norm_stderr": 0.03304205087813653
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.3586206896551724,
89
+ "acc_stderr": 0.039966295748767186,
90
+ "acc_norm": 0.3586206896551724,
91
+ "acc_norm_stderr": 0.039966295748767186
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.20588235294117646,
95
+ "acc_stderr": 0.04023382273617747,
96
+ "acc_norm": 0.20588235294117646,
97
+ "acc_norm_stderr": 0.04023382273617747
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.3697478991596639,
101
+ "acc_stderr": 0.031357095996135904,
102
+ "acc_norm": 0.3697478991596639,
103
+ "acc_norm_stderr": 0.031357095996135904
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.2794871794871795,
107
+ "acc_stderr": 0.022752388839776823,
108
+ "acc_norm": 0.2794871794871795,
109
+ "acc_norm_stderr": 0.022752388839776823
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.4,
113
+ "acc_stderr": 0.049236596391733084,
114
+ "acc_norm": 0.4,
115
+ "acc_norm_stderr": 0.049236596391733084
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.2,
119
+ "acc_stderr": 0.04020151261036846,
120
+ "acc_norm": 0.2,
121
+ "acc_norm_stderr": 0.04020151261036846
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.37962962962962965,
125
+ "acc_stderr": 0.04691521224077742,
126
+ "acc_norm": 0.37962962962962965,
127
+ "acc_norm_stderr": 0.04691521224077742
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.22660098522167488,
131
+ "acc_stderr": 0.02945486383529298,
132
+ "acc_norm": 0.22660098522167488,
133
+ "acc_norm_stderr": 0.02945486383529298
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.3225806451612903,
137
+ "acc_stderr": 0.026593084516572267,
138
+ "acc_norm": 0.3225806451612903,
139
+ "acc_norm_stderr": 0.026593084516572267
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.5,
143
+ "acc_stderr": 0.03275608910402091,
144
+ "acc_norm": 0.5,
145
+ "acc_norm_stderr": 0.03275608910402091
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.3471698113207547,
149
+ "acc_stderr": 0.029300101705549652,
150
+ "acc_norm": 0.3471698113207547,
151
+ "acc_norm_stderr": 0.029300101705549652
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.36363636363636365,
155
+ "acc_stderr": 0.04607582090719976,
156
+ "acc_norm": 0.36363636363636365,
157
+ "acc_norm_stderr": 0.04607582090719976
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.23703703703703705,
161
+ "acc_stderr": 0.025928876132766118,
162
+ "acc_norm": 0.23703703703703705,
163
+ "acc_norm_stderr": 0.025928876132766118
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2119205298013245,
167
+ "acc_stderr": 0.03336767086567977,
168
+ "acc_norm": 0.2119205298013245,
169
+ "acc_norm_stderr": 0.03336767086567977
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.39800995024875624,
173
+ "acc_stderr": 0.034611994290400135,
174
+ "acc_norm": 0.39800995024875624,
175
+ "acc_norm_stderr": 0.034611994290400135
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.2774566473988439,
179
+ "acc_stderr": 0.03414014007044036,
180
+ "acc_norm": 0.2774566473988439,
181
+ "acc_norm_stderr": 0.03414014007044036
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.24867724867724866,
185
+ "acc_stderr": 0.022261817692400175,
186
+ "acc_norm": 0.24867724867724866,
187
+ "acc_norm_stderr": 0.022261817692400175
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.2916666666666667,
191
+ "acc_stderr": 0.03800968060554859,
192
+ "acc_norm": 0.2916666666666667,
193
+ "acc_norm_stderr": 0.03800968060554859
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.21,
197
+ "acc_stderr": 0.040936018074033256,
198
+ "acc_norm": 0.21,
199
+ "acc_norm_stderr": 0.040936018074033256
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.44,
203
+ "acc_stderr": 0.04988876515698589,
204
+ "acc_norm": 0.44,
205
+ "acc_norm_stderr": 0.04988876515698589
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.38439306358381503,
209
+ "acc_stderr": 0.026189666966272035,
210
+ "acc_norm": 0.38439306358381503,
211
+ "acc_norm_stderr": 0.026189666966272035
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.3496932515337423,
215
+ "acc_stderr": 0.03746668325470022,
216
+ "acc_norm": 0.3496932515337423,
217
+ "acc_norm_stderr": 0.03746668325470022
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.3611111111111111,
221
+ "acc_stderr": 0.026725868809100793,
222
+ "acc_norm": 0.3611111111111111,
223
+ "acc_norm_stderr": 0.026725868809100793
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.28,
227
+ "acc_stderr": 0.04512608598542128,
228
+ "acc_norm": 0.28,
229
+ "acc_norm_stderr": 0.04512608598542128
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.35751295336787564,
233
+ "acc_stderr": 0.03458816042181005,
234
+ "acc_norm": 0.35751295336787564,
235
+ "acc_norm_stderr": 0.03458816042181005
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.23684210526315788,
239
+ "acc_stderr": 0.03999423879281336,
240
+ "acc_norm": 0.23684210526315788,
241
+ "acc_norm_stderr": 0.03999423879281336
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.3394495412844037,
245
+ "acc_stderr": 0.02030210934266235,
246
+ "acc_norm": 0.3394495412844037,
247
+ "acc_norm_stderr": 0.02030210934266235
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.2619047619047619,
251
+ "acc_stderr": 0.039325376803928704,
252
+ "acc_norm": 0.2619047619047619,
253
+ "acc_norm_stderr": 0.039325376803928704
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.3660130718954248,
257
+ "acc_stderr": 0.027582811415159607,
258
+ "acc_norm": 0.3660130718954248,
259
+ "acc_norm_stderr": 0.027582811415159607
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.38,
263
+ "acc_stderr": 0.04878317312145632,
264
+ "acc_norm": 0.38,
265
+ "acc_norm_stderr": 0.04878317312145632
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.49586776859504134,
269
+ "acc_stderr": 0.04564198767432754,
270
+ "acc_norm": 0.49586776859504134,
271
+ "acc_norm_stderr": 0.04564198767432754
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.24342105263157895,
275
+ "acc_stderr": 0.034923496688842384,
276
+ "acc_norm": 0.24342105263157895,
277
+ "acc_norm_stderr": 0.034923496688842384
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3366013071895425,
281
+ "acc_stderr": 0.019117213911495165,
282
+ "acc_norm": 0.3366013071895425,
283
+ "acc_norm_stderr": 0.019117213911495165
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.29432624113475175,
287
+ "acc_stderr": 0.027187127011503796,
288
+ "acc_norm": 0.29432624113475175,
289
+ "acc_norm_stderr": 0.027187127011503796
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.29464285714285715,
293
+ "acc_stderr": 0.04327040932578728,
294
+ "acc_norm": 0.29464285714285715,
295
+ "acc_norm_stderr": 0.04327040932578728
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.2824074074074074,
299
+ "acc_stderr": 0.030701372111510927,
300
+ "acc_norm": 0.2824074074074074,
301
+ "acc_norm_stderr": 0.030701372111510927
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.25139664804469275,
305
+ "acc_stderr": 0.014508979453553977,
306
+ "acc_norm": 0.25139664804469275,
307
+ "acc_norm_stderr": 0.014508979453553977
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.32,
311
+ "acc_stderr": 0.04688261722621505,
312
+ "acc_norm": 0.32,
313
+ "acc_norm_stderr": 0.04688261722621505
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.31,
317
+ "acc_stderr": 0.04648231987117316,
318
+ "acc_norm": 0.31,
319
+ "acc_norm_stderr": 0.04648231987117316
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3786764705882353,
323
+ "acc_stderr": 0.029465133639776125,
324
+ "acc_norm": 0.3786764705882353,
325
+ "acc_norm_stderr": 0.029465133639776125
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.3510204081632653,
329
+ "acc_stderr": 0.03055531675557364,
330
+ "acc_norm": 0.3510204081632653,
331
+ "acc_norm_stderr": 0.03055531675557364
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.4767932489451477,
335
+ "acc_stderr": 0.032512152011410174,
336
+ "acc_norm": 0.4767932489451477,
337
+ "acc_norm_stderr": 0.032512152011410174
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3089960886571056,
341
+ "acc_stderr": 0.01180172977723925,
342
+ "acc_norm": 0.3089960886571056,
343
+ "acc_norm_stderr": 0.01180172977723925
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.31862745098039214,
347
+ "acc_stderr": 0.032702871814820816,
348
+ "acc_norm": 0.31862745098039214,
349
+ "acc_norm_stderr": 0.032702871814820816
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.3212121212121212,
353
+ "acc_stderr": 0.0364620496325381,
354
+ "acc_norm": 0.3212121212121212,
355
+ "acc_norm_stderr": 0.0364620496325381
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.25091799265605874,
359
+ "mc1_stderr": 0.01517698502770769,
360
+ "mc2": 0.38056097212603235,
361
+ "mc2_stderr": 0.014936929596682727
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.21605667060212513,
365
+ "acc_stderr": 0.014149496716043137,
366
+ "acc_norm": 0.29279811097992914,
367
+ "acc_norm_stderr": 0.015644823205401337
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/llama-2-ko-7b-instruction-v3",
436
+ "model_sha": "c0fea9cb31d4ae90aa2ed048f774a9000341b538",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/polyglot-ko-12.8b-instruct/result_2023-09-27 21:10:18.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3046075085324232,
5
+ "acc_stderr": 0.013449522109932492,
6
+ "acc_norm": 0.363481228668942,
7
+ "acc_norm_stderr": 0.014056207319068282
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3981278629755029,
11
+ "acc_stderr": 0.0048851164655502755,
12
+ "acc_norm": 0.5159330810595499,
13
+ "acc_norm_stderr": 0.004987247325495624
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.22807017543859648,
17
+ "acc_stderr": 0.03218093795602357,
18
+ "acc_norm": 0.22807017543859648,
19
+ "acc_norm_stderr": 0.03218093795602357
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.24271844660194175,
23
+ "acc_stderr": 0.04245022486384495,
24
+ "acc_norm": 0.24271844660194175,
25
+ "acc_norm_stderr": 0.04245022486384495
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.23754789272030652,
29
+ "acc_stderr": 0.015218733046150193,
30
+ "acc_norm": 0.23754789272030652,
31
+ "acc_norm_stderr": 0.015218733046150193
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.28888888888888886,
35
+ "acc_stderr": 0.039154506304142495,
36
+ "acc_norm": 0.28888888888888886,
37
+ "acc_norm_stderr": 0.039154506304142495
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.27,
41
+ "acc_stderr": 0.0446196043338474,
42
+ "acc_norm": 0.27,
43
+ "acc_norm_stderr": 0.0446196043338474
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.2297872340425532,
47
+ "acc_stderr": 0.02750175294441242,
48
+ "acc_norm": 0.2297872340425532,
49
+ "acc_norm_stderr": 0.02750175294441242
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.2469879518072289,
53
+ "acc_stderr": 0.03357351982064536,
54
+ "acc_norm": 0.2469879518072289,
55
+ "acc_norm_stderr": 0.03357351982064536
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.33440514469453375,
59
+ "acc_stderr": 0.026795422327893944,
60
+ "acc_norm": 0.33440514469453375,
61
+ "acc_norm_stderr": 0.026795422327893944
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.1031390134529148,
65
+ "acc_stderr": 0.020412564289839272,
66
+ "acc_norm": 0.1031390134529148,
67
+ "acc_norm_stderr": 0.020412564289839272
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.2748091603053435,
71
+ "acc_stderr": 0.039153454088478354,
72
+ "acc_norm": 0.2748091603053435,
73
+ "acc_norm_stderr": 0.039153454088478354
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.37,
77
+ "acc_stderr": 0.048523658709391,
78
+ "acc_norm": 0.37,
79
+ "acc_norm_stderr": 0.048523658709391
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.40404040404040403,
83
+ "acc_stderr": 0.03496130972056128,
84
+ "acc_norm": 0.40404040404040403,
85
+ "acc_norm_stderr": 0.03496130972056128
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.3310344827586207,
89
+ "acc_stderr": 0.03921545312467122,
90
+ "acc_norm": 0.3310344827586207,
91
+ "acc_norm_stderr": 0.03921545312467122
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.2647058823529412,
95
+ "acc_stderr": 0.04389869956808777,
96
+ "acc_norm": 0.2647058823529412,
97
+ "acc_norm_stderr": 0.04389869956808777
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.18907563025210083,
101
+ "acc_stderr": 0.02543511943810536,
102
+ "acc_norm": 0.18907563025210083,
103
+ "acc_norm_stderr": 0.02543511943810536
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.21794871794871795,
107
+ "acc_stderr": 0.020932445774463175,
108
+ "acc_norm": 0.21794871794871795,
109
+ "acc_norm_stderr": 0.020932445774463175
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.18,
113
+ "acc_stderr": 0.03861229196653694,
114
+ "acc_norm": 0.18,
115
+ "acc_norm_stderr": 0.03861229196653694
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.16,
119
+ "acc_stderr": 0.03684529491774708,
120
+ "acc_norm": 0.16,
121
+ "acc_norm_stderr": 0.03684529491774708
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.26851851851851855,
125
+ "acc_stderr": 0.04284467968052191,
126
+ "acc_norm": 0.26851851851851855,
127
+ "acc_norm_stderr": 0.04284467968052191
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.1921182266009852,
131
+ "acc_stderr": 0.027719315709614778,
132
+ "acc_norm": 0.1921182266009852,
133
+ "acc_norm_stderr": 0.027719315709614778
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.25483870967741934,
137
+ "acc_stderr": 0.024790118459332204,
138
+ "acc_norm": 0.25483870967741934,
139
+ "acc_norm_stderr": 0.024790118459332204
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.2905982905982906,
143
+ "acc_stderr": 0.029745048572674057,
144
+ "acc_norm": 0.2905982905982906,
145
+ "acc_norm_stderr": 0.029745048572674057
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.29056603773584905,
149
+ "acc_stderr": 0.027943219989337156,
150
+ "acc_norm": 0.29056603773584905,
151
+ "acc_norm_stderr": 0.027943219989337156
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.2636363636363636,
155
+ "acc_stderr": 0.04220224692971987,
156
+ "acc_norm": 0.2636363636363636,
157
+ "acc_norm_stderr": 0.04220224692971987
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.25925925925925924,
161
+ "acc_stderr": 0.026719240783712166,
162
+ "acc_norm": 0.25925925925925924,
163
+ "acc_norm_stderr": 0.026719240783712166
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2913907284768212,
167
+ "acc_stderr": 0.03710185726119995,
168
+ "acc_norm": 0.2913907284768212,
169
+ "acc_norm_stderr": 0.03710185726119995
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.19402985074626866,
173
+ "acc_stderr": 0.027962677604768914,
174
+ "acc_norm": 0.19402985074626866,
175
+ "acc_norm_stderr": 0.027962677604768914
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.2254335260115607,
179
+ "acc_stderr": 0.03186209851641143,
180
+ "acc_norm": 0.2254335260115607,
181
+ "acc_norm_stderr": 0.03186209851641143
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.24867724867724866,
185
+ "acc_stderr": 0.02226181769240018,
186
+ "acc_norm": 0.24867724867724866,
187
+ "acc_norm_stderr": 0.02226181769240018
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3125,
191
+ "acc_stderr": 0.038760854559127644,
192
+ "acc_norm": 0.3125,
193
+ "acc_norm_stderr": 0.038760854559127644
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.21,
197
+ "acc_stderr": 0.04093601807403326,
198
+ "acc_norm": 0.21,
199
+ "acc_norm_stderr": 0.04093601807403326
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.27,
203
+ "acc_stderr": 0.044619604333847394,
204
+ "acc_norm": 0.27,
205
+ "acc_norm_stderr": 0.044619604333847394
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.2774566473988439,
209
+ "acc_stderr": 0.024105712607754307,
210
+ "acc_norm": 0.2774566473988439,
211
+ "acc_norm_stderr": 0.024105712607754307
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.3067484662576687,
215
+ "acc_stderr": 0.036230899157241474,
216
+ "acc_norm": 0.3067484662576687,
217
+ "acc_norm_stderr": 0.036230899157241474
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.24074074074074073,
221
+ "acc_stderr": 0.023788583551658537,
222
+ "acc_norm": 0.24074074074074073,
223
+ "acc_norm_stderr": 0.023788583551658537
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.26,
227
+ "acc_stderr": 0.04408440022768077,
228
+ "acc_norm": 0.26,
229
+ "acc_norm_stderr": 0.04408440022768077
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.22797927461139897,
233
+ "acc_stderr": 0.030276909945178253,
234
+ "acc_norm": 0.22797927461139897,
235
+ "acc_norm_stderr": 0.030276909945178253
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.30701754385964913,
239
+ "acc_stderr": 0.043391383225798594,
240
+ "acc_norm": 0.30701754385964913,
241
+ "acc_norm_stderr": 0.043391383225798594
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.3376146788990826,
245
+ "acc_stderr": 0.02027526598663891,
246
+ "acc_norm": 0.3376146788990826,
247
+ "acc_norm_stderr": 0.02027526598663891
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.23809523809523808,
251
+ "acc_stderr": 0.038095238095238126,
252
+ "acc_norm": 0.23809523809523808,
253
+ "acc_norm_stderr": 0.038095238095238126
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.27450980392156865,
257
+ "acc_stderr": 0.025553169991826524,
258
+ "acc_norm": 0.27450980392156865,
259
+ "acc_norm_stderr": 0.025553169991826524
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.32,
263
+ "acc_stderr": 0.04688261722621505,
264
+ "acc_norm": 0.32,
265
+ "acc_norm_stderr": 0.04688261722621505
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.24793388429752067,
269
+ "acc_stderr": 0.03941897526516302,
270
+ "acc_norm": 0.24793388429752067,
271
+ "acc_norm_stderr": 0.03941897526516302
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.23684210526315788,
275
+ "acc_stderr": 0.03459777606810537,
276
+ "acc_norm": 0.23684210526315788,
277
+ "acc_norm_stderr": 0.03459777606810537
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.25980392156862747,
281
+ "acc_stderr": 0.01774089950917779,
282
+ "acc_norm": 0.25980392156862747,
283
+ "acc_norm_stderr": 0.01774089950917779
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.22340425531914893,
287
+ "acc_stderr": 0.024847921358063962,
288
+ "acc_norm": 0.22340425531914893,
289
+ "acc_norm_stderr": 0.024847921358063962
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25892857142857145,
293
+ "acc_stderr": 0.04157751539865629,
294
+ "acc_norm": 0.25892857142857145,
295
+ "acc_norm_stderr": 0.04157751539865629
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.24537037037037038,
299
+ "acc_stderr": 0.029346665094372937,
300
+ "acc_norm": 0.24537037037037038,
301
+ "acc_norm_stderr": 0.029346665094372937
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.27262569832402234,
305
+ "acc_stderr": 0.014893391735249608,
306
+ "acc_norm": 0.27262569832402234,
307
+ "acc_norm_stderr": 0.014893391735249608
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.32,
311
+ "acc_stderr": 0.046882617226215034,
312
+ "acc_norm": 0.32,
313
+ "acc_norm_stderr": 0.046882617226215034
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.31,
317
+ "acc_stderr": 0.04648231987117316,
318
+ "acc_norm": 0.31,
319
+ "acc_norm_stderr": 0.04648231987117316
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3786764705882353,
323
+ "acc_stderr": 0.02946513363977613,
324
+ "acc_norm": 0.3786764705882353,
325
+ "acc_norm_stderr": 0.02946513363977613
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.35918367346938773,
329
+ "acc_stderr": 0.03071356045510849,
330
+ "acc_norm": 0.35918367346938773,
331
+ "acc_norm_stderr": 0.03071356045510849
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.24472573839662448,
335
+ "acc_stderr": 0.027985699387036416,
336
+ "acc_norm": 0.24472573839662448,
337
+ "acc_norm_stderr": 0.027985699387036416
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.2588005215123859,
341
+ "acc_stderr": 0.011186109046564608,
342
+ "acc_norm": 0.2588005215123859,
343
+ "acc_norm_stderr": 0.011186109046564608
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.28431372549019607,
347
+ "acc_stderr": 0.031660096793998116,
348
+ "acc_norm": 0.28431372549019607,
349
+ "acc_norm_stderr": 0.031660096793998116
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.2,
353
+ "acc_stderr": 0.03123475237772118,
354
+ "acc_norm": 0.2,
355
+ "acc_norm_stderr": 0.03123475237772118
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.27906976744186046,
359
+ "mc1_stderr": 0.015702107090627887,
360
+ "mc2": 0.4515720476496737,
361
+ "mc2_stderr": 0.015493161984611252
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.2680047225501771,
365
+ "acc_stderr": 0.015227905796335147,
366
+ "acc_norm": 0.3707201889020071,
367
+ "acc_norm_stderr": 0.016605801289212598
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/polyglot-ko-12.8b-instruct",
436
+ "model_sha": "a8354bcedc167e8e1f7dac8a347bf4b61d9c9bf0",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42MARU/sitebunny-13b/result_2023-09-27 08:17:31.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3643344709897611,
5
+ "acc_stderr": 0.014063260279882417,
6
+ "acc_norm": 0.4112627986348123,
7
+ "acc_norm_stderr": 0.014379441068522084
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3732324238199562,
11
+ "acc_stderr": 0.004826746160830189,
12
+ "acc_norm": 0.4751045608444533,
13
+ "acc_norm_stderr": 0.004983592410934169
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4853801169590643,
17
+ "acc_stderr": 0.038331852752130205,
18
+ "acc_norm": 0.4853801169590643,
19
+ "acc_norm_stderr": 0.038331852752130205
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5631067961165048,
23
+ "acc_stderr": 0.04911147107365777,
24
+ "acc_norm": 0.5631067961165048,
25
+ "acc_norm_stderr": 0.04911147107365777
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4878671775223499,
29
+ "acc_stderr": 0.017874698667491355,
30
+ "acc_norm": 0.4878671775223499,
31
+ "acc_norm_stderr": 0.017874698667491355
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4148148148148148,
35
+ "acc_stderr": 0.042561937679014075,
36
+ "acc_norm": 0.4148148148148148,
37
+ "acc_norm_stderr": 0.042561937679014075
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.3,
41
+ "acc_stderr": 0.046056618647183814,
42
+ "acc_norm": 0.3,
43
+ "acc_norm_stderr": 0.046056618647183814
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.4,
47
+ "acc_stderr": 0.03202563076101735,
48
+ "acc_norm": 0.4,
49
+ "acc_norm_stderr": 0.03202563076101735
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3855421686746988,
53
+ "acc_stderr": 0.03789134424611548,
54
+ "acc_norm": 0.3855421686746988,
55
+ "acc_norm_stderr": 0.03789134424611548
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4533762057877814,
59
+ "acc_stderr": 0.028274359854894245,
60
+ "acc_norm": 0.4533762057877814,
61
+ "acc_norm_stderr": 0.028274359854894245
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.4663677130044843,
65
+ "acc_stderr": 0.033481800170603065,
66
+ "acc_norm": 0.4663677130044843,
67
+ "acc_norm_stderr": 0.033481800170603065
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.48091603053435117,
71
+ "acc_stderr": 0.04382094705550988,
72
+ "acc_norm": 0.48091603053435117,
73
+ "acc_norm_stderr": 0.04382094705550988
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.42,
77
+ "acc_stderr": 0.04960449637488583,
78
+ "acc_norm": 0.42,
79
+ "acc_norm_stderr": 0.04960449637488583
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5909090909090909,
83
+ "acc_stderr": 0.03502975799413008,
84
+ "acc_norm": 0.5909090909090909,
85
+ "acc_norm_stderr": 0.03502975799413008
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4413793103448276,
89
+ "acc_stderr": 0.04137931034482758,
90
+ "acc_norm": 0.4413793103448276,
91
+ "acc_norm_stderr": 0.04137931034482758
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.16666666666666666,
95
+ "acc_stderr": 0.03708284662416544,
96
+ "acc_norm": 0.16666666666666666,
97
+ "acc_norm_stderr": 0.03708284662416544
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.4495798319327731,
101
+ "acc_stderr": 0.03231293497137707,
102
+ "acc_norm": 0.4495798319327731,
103
+ "acc_norm_stderr": 0.03231293497137707
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4358974358974359,
107
+ "acc_stderr": 0.025141801511177498,
108
+ "acc_norm": 0.4358974358974359,
109
+ "acc_norm_stderr": 0.025141801511177498
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.46,
113
+ "acc_stderr": 0.05009082659620333,
114
+ "acc_norm": 0.46,
115
+ "acc_norm_stderr": 0.05009082659620333
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.42,
119
+ "acc_stderr": 0.049604496374885836,
120
+ "acc_norm": 0.42,
121
+ "acc_norm_stderr": 0.049604496374885836
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5277777777777778,
125
+ "acc_stderr": 0.048262172941398944,
126
+ "acc_norm": 0.5277777777777778,
127
+ "acc_norm_stderr": 0.048262172941398944
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3645320197044335,
131
+ "acc_stderr": 0.0338640574606209,
132
+ "acc_norm": 0.3645320197044335,
133
+ "acc_norm_stderr": 0.0338640574606209
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.47096774193548385,
137
+ "acc_stderr": 0.028396016402761005,
138
+ "acc_norm": 0.47096774193548385,
139
+ "acc_norm_stderr": 0.028396016402761005
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6282051282051282,
143
+ "acc_stderr": 0.03166098891888078,
144
+ "acc_norm": 0.6282051282051282,
145
+ "acc_norm_stderr": 0.03166098891888078
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4528301886792453,
149
+ "acc_stderr": 0.030635627957961823,
150
+ "acc_norm": 0.4528301886792453,
151
+ "acc_norm_stderr": 0.030635627957961823
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.509090909090909,
155
+ "acc_stderr": 0.04788339768702861,
156
+ "acc_norm": 0.509090909090909,
157
+ "acc_norm_stderr": 0.04788339768702861
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.3111111111111111,
161
+ "acc_stderr": 0.028226446749683526,
162
+ "acc_norm": 0.3111111111111111,
163
+ "acc_norm_stderr": 0.028226446749683526
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.31788079470198677,
167
+ "acc_stderr": 0.038020397601079024,
168
+ "acc_norm": 0.31788079470198677,
169
+ "acc_norm_stderr": 0.038020397601079024
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5771144278606966,
173
+ "acc_stderr": 0.034932317774212816,
174
+ "acc_norm": 0.5771144278606966,
175
+ "acc_norm_stderr": 0.034932317774212816
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3815028901734104,
179
+ "acc_stderr": 0.03703851193099521,
180
+ "acc_norm": 0.3815028901734104,
181
+ "acc_norm_stderr": 0.03703851193099521
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.3253968253968254,
185
+ "acc_stderr": 0.024130158299762613,
186
+ "acc_norm": 0.3253968253968254,
187
+ "acc_norm_stderr": 0.024130158299762613
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3819444444444444,
191
+ "acc_stderr": 0.040629907841466674,
192
+ "acc_norm": 0.3819444444444444,
193
+ "acc_norm_stderr": 0.040629907841466674
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.26,
197
+ "acc_stderr": 0.04408440022768079,
198
+ "acc_norm": 0.26,
199
+ "acc_norm_stderr": 0.04408440022768079
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.6,
203
+ "acc_stderr": 0.049236596391733084,
204
+ "acc_norm": 0.6,
205
+ "acc_norm_stderr": 0.049236596391733084
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5144508670520231,
209
+ "acc_stderr": 0.02690784985628254,
210
+ "acc_norm": 0.5144508670520231,
211
+ "acc_norm_stderr": 0.02690784985628254
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.44171779141104295,
215
+ "acc_stderr": 0.03901591825836184,
216
+ "acc_norm": 0.44171779141104295,
217
+ "acc_norm_stderr": 0.03901591825836184
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4444444444444444,
221
+ "acc_stderr": 0.02764847787741332,
222
+ "acc_norm": 0.4444444444444444,
223
+ "acc_norm_stderr": 0.02764847787741332
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.35,
227
+ "acc_stderr": 0.0479372485441102,
228
+ "acc_norm": 0.35,
229
+ "acc_norm_stderr": 0.0479372485441102
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5077720207253886,
233
+ "acc_stderr": 0.03608003225569654,
234
+ "acc_norm": 0.5077720207253886,
235
+ "acc_norm_stderr": 0.03608003225569654
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.19298245614035087,
239
+ "acc_stderr": 0.037124548537213684,
240
+ "acc_norm": 0.19298245614035087,
241
+ "acc_norm_stderr": 0.037124548537213684
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.47889908256880737,
245
+ "acc_stderr": 0.021418224754264643,
246
+ "acc_norm": 0.47889908256880737,
247
+ "acc_norm_stderr": 0.021418224754264643
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.373015873015873,
251
+ "acc_stderr": 0.04325506042017086,
252
+ "acc_norm": 0.373015873015873,
253
+ "acc_norm_stderr": 0.04325506042017086
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.43790849673202614,
257
+ "acc_stderr": 0.028408302020332687,
258
+ "acc_norm": 0.43790849673202614,
259
+ "acc_norm_stderr": 0.028408302020332687
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.42,
263
+ "acc_stderr": 0.04960449637488584,
264
+ "acc_norm": 0.42,
265
+ "acc_norm_stderr": 0.04960449637488584
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6859504132231405,
269
+ "acc_stderr": 0.042369647530410184,
270
+ "acc_norm": 0.6859504132231405,
271
+ "acc_norm_stderr": 0.042369647530410184
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.34868421052631576,
275
+ "acc_stderr": 0.03878139888797609,
276
+ "acc_norm": 0.34868421052631576,
277
+ "acc_norm_stderr": 0.03878139888797609
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3480392156862745,
281
+ "acc_stderr": 0.01927099870822398,
282
+ "acc_norm": 0.3480392156862745,
283
+ "acc_norm_stderr": 0.01927099870822398
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3475177304964539,
287
+ "acc_stderr": 0.028406627809590947,
288
+ "acc_norm": 0.3475177304964539,
289
+ "acc_norm_stderr": 0.028406627809590947
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.23214285714285715,
293
+ "acc_stderr": 0.040073418097558065,
294
+ "acc_norm": 0.23214285714285715,
295
+ "acc_norm_stderr": 0.040073418097558065
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3101851851851852,
299
+ "acc_stderr": 0.031546962856566295,
300
+ "acc_norm": 0.3101851851851852,
301
+ "acc_norm_stderr": 0.031546962856566295
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.3106145251396648,
305
+ "acc_stderr": 0.015476515438005566,
306
+ "acc_norm": 0.3106145251396648,
307
+ "acc_norm_stderr": 0.015476515438005566
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.41,
311
+ "acc_stderr": 0.049431107042371025,
312
+ "acc_norm": 0.41,
313
+ "acc_norm_stderr": 0.049431107042371025
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.44,
317
+ "acc_stderr": 0.04988876515698589,
318
+ "acc_norm": 0.44,
319
+ "acc_norm_stderr": 0.04988876515698589
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3235294117647059,
323
+ "acc_stderr": 0.02841820861940679,
324
+ "acc_norm": 0.3235294117647059,
325
+ "acc_norm_stderr": 0.02841820861940679
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5428571428571428,
329
+ "acc_stderr": 0.03189141832421396,
330
+ "acc_norm": 0.5428571428571428,
331
+ "acc_norm_stderr": 0.03189141832421396
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5780590717299579,
335
+ "acc_stderr": 0.032148146302403695,
336
+ "acc_norm": 0.5780590717299579,
337
+ "acc_norm_stderr": 0.032148146302403695
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.32333767926988266,
341
+ "acc_stderr": 0.011946565758447202,
342
+ "acc_norm": 0.32333767926988266,
343
+ "acc_norm_stderr": 0.011946565758447202
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.47549019607843135,
347
+ "acc_stderr": 0.035050931943487976,
348
+ "acc_norm": 0.47549019607843135,
349
+ "acc_norm_stderr": 0.035050931943487976
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5393939393939394,
353
+ "acc_stderr": 0.03892207016552012,
354
+ "acc_norm": 0.5393939393939394,
355
+ "acc_norm_stderr": 0.03892207016552012
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.35006119951040393,
359
+ "mc1_stderr": 0.01669794942015103,
360
+ "mc2": 0.5148844380994511,
361
+ "mc2_stderr": 0.015947695748354234
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.42857142857142855,
365
+ "acc_stderr": 0.017014038119297473,
366
+ "acc_norm": 0.44155844155844154,
367
+ "acc_norm_stderr": 0.0170725258755631
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42MARU/sitebunny-13b",
436
+ "model_sha": "15c8578d2be688d6b03ed2076658865bb8752673",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42dot/42dot_LLM-PLM-1.3B/result_2023-10-18 01:46:47.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.2636518771331058,
5
+ "acc_stderr": 0.01287592915129705,
6
+ "acc_norm": 0.32593856655290104,
7
+ "acc_norm_stderr": 0.013697432466693242
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3563035251941844,
11
+ "acc_stderr": 0.004779276329704052,
12
+ "acc_norm": 0.4473212507468632,
13
+ "acc_norm_stderr": 0.004962010338226348
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.23976608187134502,
17
+ "acc_stderr": 0.03274485211946956,
18
+ "acc_norm": 0.23976608187134502,
19
+ "acc_norm_stderr": 0.03274485211946956
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.1941747572815534,
23
+ "acc_stderr": 0.03916667762822584,
24
+ "acc_norm": 0.1941747572815534,
25
+ "acc_norm_stderr": 0.03916667762822584
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.22094508301404853,
29
+ "acc_stderr": 0.014836205167333574,
30
+ "acc_norm": 0.22094508301404853,
31
+ "acc_norm_stderr": 0.014836205167333574
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.18518518518518517,
35
+ "acc_stderr": 0.0335567721631314,
36
+ "acc_norm": 0.18518518518518517,
37
+ "acc_norm_stderr": 0.0335567721631314
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.27,
41
+ "acc_stderr": 0.04461960433384741,
42
+ "acc_norm": 0.27,
43
+ "acc_norm_stderr": 0.04461960433384741
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.2851063829787234,
47
+ "acc_stderr": 0.02951319662553935,
48
+ "acc_norm": 0.2851063829787234,
49
+ "acc_norm_stderr": 0.02951319662553935
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3072289156626506,
53
+ "acc_stderr": 0.03591566797824664,
54
+ "acc_norm": 0.3072289156626506,
55
+ "acc_norm_stderr": 0.03591566797824664
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.2347266881028939,
59
+ "acc_stderr": 0.024071805887677045,
60
+ "acc_norm": 0.2347266881028939,
61
+ "acc_norm_stderr": 0.024071805887677045
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.2242152466367713,
65
+ "acc_stderr": 0.027991534258519527,
66
+ "acc_norm": 0.2242152466367713,
67
+ "acc_norm_stderr": 0.027991534258519527
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.2900763358778626,
71
+ "acc_stderr": 0.03980066246467765,
72
+ "acc_norm": 0.2900763358778626,
73
+ "acc_norm_stderr": 0.03980066246467765
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.28,
77
+ "acc_stderr": 0.04512608598542127,
78
+ "acc_norm": 0.28,
79
+ "acc_norm_stderr": 0.04512608598542127
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.2474747474747475,
83
+ "acc_stderr": 0.030746300742124484,
84
+ "acc_norm": 0.2474747474747475,
85
+ "acc_norm_stderr": 0.030746300742124484
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.21379310344827587,
89
+ "acc_stderr": 0.034165204477475494,
90
+ "acc_norm": 0.21379310344827587,
91
+ "acc_norm_stderr": 0.034165204477475494
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.22549019607843138,
95
+ "acc_stderr": 0.041583075330832865,
96
+ "acc_norm": 0.22549019607843138,
97
+ "acc_norm_stderr": 0.041583075330832865
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.3235294117647059,
101
+ "acc_stderr": 0.030388353551886835,
102
+ "acc_norm": 0.3235294117647059,
103
+ "acc_norm_stderr": 0.030388353551886835
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.36153846153846153,
107
+ "acc_stderr": 0.02435958146539698,
108
+ "acc_norm": 0.36153846153846153,
109
+ "acc_norm_stderr": 0.02435958146539698
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.17,
113
+ "acc_stderr": 0.03775251680686371,
114
+ "acc_norm": 0.17,
115
+ "acc_norm_stderr": 0.03775251680686371
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.34,
119
+ "acc_stderr": 0.04760952285695235,
120
+ "acc_norm": 0.34,
121
+ "acc_norm_stderr": 0.04760952285695235
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.2037037037037037,
125
+ "acc_stderr": 0.038935425188248475,
126
+ "acc_norm": 0.2037037037037037,
127
+ "acc_norm_stderr": 0.038935425188248475
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.270935960591133,
131
+ "acc_stderr": 0.031270907132976984,
132
+ "acc_norm": 0.270935960591133,
133
+ "acc_norm_stderr": 0.031270907132976984
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.2870967741935484,
137
+ "acc_stderr": 0.025736542745594525,
138
+ "acc_norm": 0.2870967741935484,
139
+ "acc_norm_stderr": 0.025736542745594525
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.20085470085470086,
143
+ "acc_stderr": 0.02624677294689047,
144
+ "acc_norm": 0.20085470085470086,
145
+ "acc_norm_stderr": 0.02624677294689047
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.25660377358490566,
149
+ "acc_stderr": 0.026880647889051985,
150
+ "acc_norm": 0.25660377358490566,
151
+ "acc_norm_stderr": 0.026880647889051985
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.2636363636363636,
155
+ "acc_stderr": 0.04220224692971987,
156
+ "acc_norm": 0.2636363636363636,
157
+ "acc_norm_stderr": 0.04220224692971987
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.24814814814814815,
161
+ "acc_stderr": 0.0263357394040558,
162
+ "acc_norm": 0.24814814814814815,
163
+ "acc_norm_stderr": 0.0263357394040558
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.3509933774834437,
167
+ "acc_stderr": 0.03896981964257374,
168
+ "acc_norm": 0.3509933774834437,
169
+ "acc_norm_stderr": 0.03896981964257374
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.2935323383084577,
173
+ "acc_stderr": 0.03220024104534205,
174
+ "acc_norm": 0.2935323383084577,
175
+ "acc_norm_stderr": 0.03220024104534205
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.26011560693641617,
179
+ "acc_stderr": 0.03345036916788991,
180
+ "acc_norm": 0.26011560693641617,
181
+ "acc_norm_stderr": 0.03345036916788991
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.24338624338624337,
185
+ "acc_stderr": 0.022101128787415426,
186
+ "acc_norm": 0.24338624338624337,
187
+ "acc_norm_stderr": 0.022101128787415426
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.2569444444444444,
191
+ "acc_stderr": 0.03653946969442099,
192
+ "acc_norm": 0.2569444444444444,
193
+ "acc_norm_stderr": 0.03653946969442099
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.26,
197
+ "acc_stderr": 0.0440844002276808,
198
+ "acc_norm": 0.26,
199
+ "acc_norm_stderr": 0.0440844002276808
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.25,
203
+ "acc_stderr": 0.04351941398892446,
204
+ "acc_norm": 0.25,
205
+ "acc_norm_stderr": 0.04351941398892446
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.2254335260115607,
209
+ "acc_stderr": 0.022497230190967547,
210
+ "acc_norm": 0.2254335260115607,
211
+ "acc_norm_stderr": 0.022497230190967547
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.26993865030674846,
215
+ "acc_stderr": 0.034878251684978906,
216
+ "acc_norm": 0.26993865030674846,
217
+ "acc_norm_stderr": 0.034878251684978906
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.28703703703703703,
221
+ "acc_stderr": 0.025171041915309684,
222
+ "acc_norm": 0.28703703703703703,
223
+ "acc_norm_stderr": 0.025171041915309684
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.3,
227
+ "acc_stderr": 0.046056618647183814,
228
+ "acc_norm": 0.3,
229
+ "acc_norm_stderr": 0.046056618647183814
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.35751295336787564,
233
+ "acc_stderr": 0.03458816042181006,
234
+ "acc_norm": 0.35751295336787564,
235
+ "acc_norm_stderr": 0.03458816042181006
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.30701754385964913,
239
+ "acc_stderr": 0.043391383225798594,
240
+ "acc_norm": 0.30701754385964913,
241
+ "acc_norm_stderr": 0.043391383225798594
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.24954128440366974,
245
+ "acc_stderr": 0.018553897629501614,
246
+ "acc_norm": 0.24954128440366974,
247
+ "acc_norm_stderr": 0.018553897629501614
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.373015873015873,
251
+ "acc_stderr": 0.04325506042017086,
252
+ "acc_norm": 0.373015873015873,
253
+ "acc_norm_stderr": 0.04325506042017086
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.24183006535947713,
257
+ "acc_stderr": 0.024518195641879334,
258
+ "acc_norm": 0.24183006535947713,
259
+ "acc_norm_stderr": 0.024518195641879334
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.2,
263
+ "acc_stderr": 0.04020151261036846,
264
+ "acc_norm": 0.2,
265
+ "acc_norm_stderr": 0.04020151261036846
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.38016528925619836,
269
+ "acc_stderr": 0.04431324501968432,
270
+ "acc_norm": 0.38016528925619836,
271
+ "acc_norm_stderr": 0.04431324501968432
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.21710526315789475,
275
+ "acc_stderr": 0.03355045304882924,
276
+ "acc_norm": 0.21710526315789475,
277
+ "acc_norm_stderr": 0.03355045304882924
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.2549019607843137,
281
+ "acc_stderr": 0.017630827375148383,
282
+ "acc_norm": 0.2549019607843137,
283
+ "acc_norm_stderr": 0.017630827375148383
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.26595744680851063,
287
+ "acc_stderr": 0.026358065698880592,
288
+ "acc_norm": 0.26595744680851063,
289
+ "acc_norm_stderr": 0.026358065698880592
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25892857142857145,
293
+ "acc_stderr": 0.041577515398656284,
294
+ "acc_norm": 0.25892857142857145,
295
+ "acc_norm_stderr": 0.041577515398656284
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.4722222222222222,
299
+ "acc_stderr": 0.0340470532865388,
300
+ "acc_norm": 0.4722222222222222,
301
+ "acc_norm_stderr": 0.0340470532865388
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.27262569832402234,
305
+ "acc_stderr": 0.014893391735249608,
306
+ "acc_norm": 0.27262569832402234,
307
+ "acc_norm_stderr": 0.014893391735249608
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.31,
311
+ "acc_stderr": 0.04648231987117316,
312
+ "acc_norm": 0.31,
313
+ "acc_norm_stderr": 0.04648231987117316
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.23,
317
+ "acc_stderr": 0.04229525846816505,
318
+ "acc_norm": 0.23,
319
+ "acc_norm_stderr": 0.04229525846816505
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.4522058823529412,
323
+ "acc_stderr": 0.030233758551596452,
324
+ "acc_norm": 0.4522058823529412,
325
+ "acc_norm_stderr": 0.030233758551596452
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.27346938775510204,
329
+ "acc_stderr": 0.02853556033712845,
330
+ "acc_norm": 0.27346938775510204,
331
+ "acc_norm_stderr": 0.02853556033712845
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.26582278481012656,
335
+ "acc_stderr": 0.02875679962965833,
336
+ "acc_norm": 0.26582278481012656,
337
+ "acc_norm_stderr": 0.02875679962965833
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.2561929595827901,
341
+ "acc_stderr": 0.011149173153110583,
342
+ "acc_norm": 0.2561929595827901,
343
+ "acc_norm_stderr": 0.011149173153110583
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.24019607843137256,
347
+ "acc_stderr": 0.02998373305591361,
348
+ "acc_norm": 0.24019607843137256,
349
+ "acc_norm_stderr": 0.02998373305591361
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.24242424242424243,
353
+ "acc_stderr": 0.03346409881055953,
354
+ "acc_norm": 0.24242424242424243,
355
+ "acc_norm_stderr": 0.03346409881055953
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.24479804161566707,
359
+ "mc1_stderr": 0.015051869486715006,
360
+ "mc2": 0.40367736123530334,
361
+ "mc2_stderr": 0.014824402657107816
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.2396694214876033,
365
+ "acc_stderr": 0.014676495332267253,
366
+ "acc_norm": 0.36835891381345925,
367
+ "acc_norm_stderr": 0.016583858982639074
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42dot/42dot_LLM-PLM-1.3B",
436
+ "model_sha": "a72bf57eb02cd4ea4388a344b4a5893aa95698da",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
42dot/42dot_LLM-SFT-1.3B/result_2023-10-18 01:47:03.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.28242320819112626,
5
+ "acc_stderr": 0.01315545688409722,
6
+ "acc_norm": 0.35494880546075086,
7
+ "acc_norm_stderr": 0.013983036904094094
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.36317466640111534,
11
+ "acc_stderr": 0.004799317209902023,
12
+ "acc_norm": 0.4613622784305915,
13
+ "acc_norm_stderr": 0.004974860878464429
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.32748538011695905,
17
+ "acc_stderr": 0.035993357714560276,
18
+ "acc_norm": 0.32748538011695905,
19
+ "acc_norm_stderr": 0.035993357714560276
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.13592233009708737,
23
+ "acc_stderr": 0.033932957297610124,
24
+ "acc_norm": 0.13592233009708737,
25
+ "acc_norm_stderr": 0.033932957297610124
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.23754789272030652,
29
+ "acc_stderr": 0.015218733046150193,
30
+ "acc_norm": 0.23754789272030652,
31
+ "acc_norm_stderr": 0.015218733046150193
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.23703703703703705,
35
+ "acc_stderr": 0.03673731683969506,
36
+ "acc_norm": 0.23703703703703705,
37
+ "acc_norm_stderr": 0.03673731683969506
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.28,
41
+ "acc_stderr": 0.04512608598542127,
42
+ "acc_norm": 0.28,
43
+ "acc_norm_stderr": 0.04512608598542127
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.2723404255319149,
47
+ "acc_stderr": 0.029101290698386698,
48
+ "acc_norm": 0.2723404255319149,
49
+ "acc_norm_stderr": 0.029101290698386698
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.2710843373493976,
53
+ "acc_stderr": 0.034605799075530276,
54
+ "acc_norm": 0.2710843373493976,
55
+ "acc_norm_stderr": 0.034605799075530276
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.2604501607717042,
59
+ "acc_stderr": 0.024926723224845543,
60
+ "acc_norm": 0.2604501607717042,
61
+ "acc_norm_stderr": 0.024926723224845543
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.242152466367713,
65
+ "acc_stderr": 0.028751392398694755,
66
+ "acc_norm": 0.242152466367713,
67
+ "acc_norm_stderr": 0.028751392398694755
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.2366412213740458,
71
+ "acc_stderr": 0.03727673575596918,
72
+ "acc_norm": 0.2366412213740458,
73
+ "acc_norm_stderr": 0.03727673575596918
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.27,
77
+ "acc_stderr": 0.0446196043338474,
78
+ "acc_norm": 0.27,
79
+ "acc_norm_stderr": 0.0446196043338474
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.22727272727272727,
83
+ "acc_stderr": 0.02985751567338641,
84
+ "acc_norm": 0.22727272727272727,
85
+ "acc_norm_stderr": 0.02985751567338641
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.2482758620689655,
89
+ "acc_stderr": 0.036001056927277716,
90
+ "acc_norm": 0.2482758620689655,
91
+ "acc_norm_stderr": 0.036001056927277716
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.1568627450980392,
95
+ "acc_stderr": 0.03618664819936246,
96
+ "acc_norm": 0.1568627450980392,
97
+ "acc_norm_stderr": 0.03618664819936246
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.24789915966386555,
101
+ "acc_stderr": 0.028047967224176892,
102
+ "acc_norm": 0.24789915966386555,
103
+ "acc_norm_stderr": 0.028047967224176892
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.23076923076923078,
107
+ "acc_stderr": 0.021362027725222728,
108
+ "acc_norm": 0.23076923076923078,
109
+ "acc_norm_stderr": 0.021362027725222728
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.32,
113
+ "acc_stderr": 0.04688261722621504,
114
+ "acc_norm": 0.32,
115
+ "acc_norm_stderr": 0.04688261722621504
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.19,
119
+ "acc_stderr": 0.03942772444036623,
120
+ "acc_norm": 0.19,
121
+ "acc_norm_stderr": 0.03942772444036623
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.25925925925925924,
125
+ "acc_stderr": 0.042365112580946336,
126
+ "acc_norm": 0.25925925925925924,
127
+ "acc_norm_stderr": 0.042365112580946336
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.18719211822660098,
131
+ "acc_stderr": 0.027444924966882618,
132
+ "acc_norm": 0.18719211822660098,
133
+ "acc_norm_stderr": 0.027444924966882618
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.2645161290322581,
137
+ "acc_stderr": 0.02509189237885928,
138
+ "acc_norm": 0.2645161290322581,
139
+ "acc_norm_stderr": 0.02509189237885928
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.3034188034188034,
143
+ "acc_stderr": 0.030118210106942652,
144
+ "acc_norm": 0.3034188034188034,
145
+ "acc_norm_stderr": 0.030118210106942652
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.2037735849056604,
149
+ "acc_stderr": 0.02479078450177541,
150
+ "acc_norm": 0.2037735849056604,
151
+ "acc_norm_stderr": 0.02479078450177541
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.22727272727272727,
155
+ "acc_stderr": 0.04013964554072775,
156
+ "acc_norm": 0.22727272727272727,
157
+ "acc_norm_stderr": 0.04013964554072775
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.24074074074074073,
161
+ "acc_stderr": 0.026067159222275794,
162
+ "acc_norm": 0.24074074074074073,
163
+ "acc_norm_stderr": 0.026067159222275794
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.24503311258278146,
167
+ "acc_stderr": 0.035118075718047245,
168
+ "acc_norm": 0.24503311258278146,
169
+ "acc_norm_stderr": 0.035118075718047245
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.24875621890547264,
173
+ "acc_stderr": 0.030567675938916707,
174
+ "acc_norm": 0.24875621890547264,
175
+ "acc_norm_stderr": 0.030567675938916707
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.18497109826589594,
179
+ "acc_stderr": 0.029605623981771204,
180
+ "acc_norm": 0.18497109826589594,
181
+ "acc_norm_stderr": 0.029605623981771204
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.25132275132275134,
185
+ "acc_stderr": 0.022340482339643898,
186
+ "acc_norm": 0.25132275132275134,
187
+ "acc_norm_stderr": 0.022340482339643898
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.22916666666666666,
191
+ "acc_stderr": 0.035146974678623884,
192
+ "acc_norm": 0.22916666666666666,
193
+ "acc_norm_stderr": 0.035146974678623884
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.24,
197
+ "acc_stderr": 0.04292346959909284,
198
+ "acc_norm": 0.24,
199
+ "acc_norm_stderr": 0.04292346959909284
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.28,
203
+ "acc_stderr": 0.04512608598542129,
204
+ "acc_norm": 0.28,
205
+ "acc_norm_stderr": 0.04512608598542129
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.23410404624277456,
209
+ "acc_stderr": 0.022797110278071128,
210
+ "acc_norm": 0.23410404624277456,
211
+ "acc_norm_stderr": 0.022797110278071128
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.26380368098159507,
215
+ "acc_stderr": 0.03462419931615624,
216
+ "acc_norm": 0.26380368098159507,
217
+ "acc_norm_stderr": 0.03462419931615624
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.28703703703703703,
221
+ "acc_stderr": 0.025171041915309684,
222
+ "acc_norm": 0.28703703703703703,
223
+ "acc_norm_stderr": 0.025171041915309684
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.3,
227
+ "acc_stderr": 0.046056618647183814,
228
+ "acc_norm": 0.3,
229
+ "acc_norm_stderr": 0.046056618647183814
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.20725388601036268,
233
+ "acc_stderr": 0.029252823291803644,
234
+ "acc_norm": 0.20725388601036268,
235
+ "acc_norm_stderr": 0.029252823291803644
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2631578947368421,
239
+ "acc_stderr": 0.04142439719489362,
240
+ "acc_norm": 0.2631578947368421,
241
+ "acc_norm_stderr": 0.04142439719489362
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.22201834862385322,
245
+ "acc_stderr": 0.017818849564796624,
246
+ "acc_norm": 0.22201834862385322,
247
+ "acc_norm_stderr": 0.017818849564796624
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.31746031746031744,
251
+ "acc_stderr": 0.04163453031302859,
252
+ "acc_norm": 0.31746031746031744,
253
+ "acc_norm_stderr": 0.04163453031302859
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.24836601307189543,
257
+ "acc_stderr": 0.024739981355113592,
258
+ "acc_norm": 0.24836601307189543,
259
+ "acc_norm_stderr": 0.024739981355113592
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.22,
263
+ "acc_stderr": 0.04163331998932269,
264
+ "acc_norm": 0.22,
265
+ "acc_norm_stderr": 0.04163331998932269
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.32231404958677684,
269
+ "acc_stderr": 0.042664163633521685,
270
+ "acc_norm": 0.32231404958677684,
271
+ "acc_norm_stderr": 0.042664163633521685
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.20394736842105263,
275
+ "acc_stderr": 0.0327900040631005,
276
+ "acc_norm": 0.20394736842105263,
277
+ "acc_norm_stderr": 0.0327900040631005
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.2581699346405229,
281
+ "acc_stderr": 0.017704531653250075,
282
+ "acc_norm": 0.2581699346405229,
283
+ "acc_norm_stderr": 0.017704531653250075
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.2375886524822695,
287
+ "acc_stderr": 0.025389512552729903,
288
+ "acc_norm": 0.2375886524822695,
289
+ "acc_norm_stderr": 0.025389512552729903
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25,
293
+ "acc_stderr": 0.04109974682633932,
294
+ "acc_norm": 0.25,
295
+ "acc_norm_stderr": 0.04109974682633932
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3333333333333333,
299
+ "acc_stderr": 0.03214952147802747,
300
+ "acc_norm": 0.3333333333333333,
301
+ "acc_norm_stderr": 0.03214952147802747
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.27262569832402234,
305
+ "acc_stderr": 0.014893391735249608,
306
+ "acc_norm": 0.27262569832402234,
307
+ "acc_norm_stderr": 0.014893391735249608
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.28,
311
+ "acc_stderr": 0.04512608598542127,
312
+ "acc_norm": 0.28,
313
+ "acc_norm_stderr": 0.04512608598542127
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.19,
317
+ "acc_stderr": 0.039427724440366234,
318
+ "acc_norm": 0.19,
319
+ "acc_norm_stderr": 0.039427724440366234
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.25,
323
+ "acc_stderr": 0.026303648393696036,
324
+ "acc_norm": 0.25,
325
+ "acc_norm_stderr": 0.026303648393696036
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.21224489795918366,
329
+ "acc_stderr": 0.026176967197866764,
330
+ "acc_norm": 0.21224489795918366,
331
+ "acc_norm_stderr": 0.026176967197866764
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.28270042194092826,
335
+ "acc_stderr": 0.029312814153955914,
336
+ "acc_norm": 0.28270042194092826,
337
+ "acc_norm_stderr": 0.029312814153955914
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.2503259452411995,
341
+ "acc_stderr": 0.011064151027165438,
342
+ "acc_norm": 0.2503259452411995,
343
+ "acc_norm_stderr": 0.011064151027165438
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.25,
347
+ "acc_stderr": 0.03039153369274154,
348
+ "acc_norm": 0.25,
349
+ "acc_norm_stderr": 0.03039153369274154
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.23030303030303031,
353
+ "acc_stderr": 0.03287666758603488,
354
+ "acc_norm": 0.23030303030303031,
355
+ "acc_norm_stderr": 0.03287666758603488
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.28518971848225216,
359
+ "mc1_stderr": 0.015805827874454895,
360
+ "mc2": 0.43765472485909873,
361
+ "mc2_stderr": 0.015405588178148114
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.2550177095631641,
365
+ "acc_stderr": 0.014985559533428578,
366
+ "acc_norm": 0.3754427390791027,
367
+ "acc_norm_stderr": 0.016648411589511095
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "42dot/42dot_LLM-SFT-1.3B",
436
+ "model_sha": "2dadd4492f0b27c302d8a5518003fa6045e32a8a",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIFT/PACK-13b-v1.0/result_2023-12-07 02:16:32.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3378839590443686,
5
+ "acc_stderr": 0.01382204792228351,
6
+ "acc_norm": 0.37542662116040953,
7
+ "acc_norm_stderr": 0.01415063143511173
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.37044413463453496,
11
+ "acc_stderr": 0.004819367172685971,
12
+ "acc_norm": 0.4788886675960964,
13
+ "acc_norm_stderr": 0.004985331652408348
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.45614035087719296,
17
+ "acc_stderr": 0.03820042586602967,
18
+ "acc_norm": 0.45614035087719296,
19
+ "acc_norm_stderr": 0.03820042586602967
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5048543689320388,
23
+ "acc_stderr": 0.04950504382128921,
24
+ "acc_norm": 0.5048543689320388,
25
+ "acc_norm_stderr": 0.04950504382128921
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.49169859514687103,
29
+ "acc_stderr": 0.017877498991072008,
30
+ "acc_norm": 0.49169859514687103,
31
+ "acc_norm_stderr": 0.017877498991072008
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.362962962962963,
35
+ "acc_stderr": 0.041539484047424,
36
+ "acc_norm": 0.362962962962963,
37
+ "acc_norm_stderr": 0.041539484047424
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.32,
41
+ "acc_stderr": 0.04688261722621504,
42
+ "acc_norm": 0.32,
43
+ "acc_norm_stderr": 0.04688261722621504
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.35319148936170214,
47
+ "acc_stderr": 0.031245325202761926,
48
+ "acc_norm": 0.35319148936170214,
49
+ "acc_norm_stderr": 0.031245325202761926
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39759036144578314,
53
+ "acc_stderr": 0.038099730845402184,
54
+ "acc_norm": 0.39759036144578314,
55
+ "acc_norm_stderr": 0.038099730845402184
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4340836012861736,
59
+ "acc_stderr": 0.0281502322445356,
60
+ "acc_norm": 0.4340836012861736,
61
+ "acc_norm_stderr": 0.0281502322445356
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.4125560538116592,
65
+ "acc_stderr": 0.03304062175449297,
66
+ "acc_norm": 0.4125560538116592,
67
+ "acc_norm_stderr": 0.03304062175449297
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.46564885496183206,
71
+ "acc_stderr": 0.04374928560599738,
72
+ "acc_norm": 0.46564885496183206,
73
+ "acc_norm_stderr": 0.04374928560599738
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.39,
77
+ "acc_stderr": 0.04902071300001975,
78
+ "acc_norm": 0.39,
79
+ "acc_norm_stderr": 0.04902071300001975
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5151515151515151,
83
+ "acc_stderr": 0.0356071651653106,
84
+ "acc_norm": 0.5151515151515151,
85
+ "acc_norm_stderr": 0.0356071651653106
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.3931034482758621,
89
+ "acc_stderr": 0.040703290137070705,
90
+ "acc_norm": 0.3931034482758621,
91
+ "acc_norm_stderr": 0.040703290137070705
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.18627450980392157,
95
+ "acc_stderr": 0.038739587141493524,
96
+ "acc_norm": 0.18627450980392157,
97
+ "acc_norm_stderr": 0.038739587141493524
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.37815126050420167,
101
+ "acc_stderr": 0.031499305777849054,
102
+ "acc_norm": 0.37815126050420167,
103
+ "acc_norm_stderr": 0.031499305777849054
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.3769230769230769,
107
+ "acc_stderr": 0.024570975364225995,
108
+ "acc_norm": 0.3769230769230769,
109
+ "acc_norm_stderr": 0.024570975364225995
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.55,
113
+ "acc_stderr": 0.049999999999999996,
114
+ "acc_norm": 0.55,
115
+ "acc_norm_stderr": 0.049999999999999996
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.32,
119
+ "acc_stderr": 0.04688261722621504,
120
+ "acc_norm": 0.32,
121
+ "acc_norm_stderr": 0.04688261722621504
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.49074074074074076,
125
+ "acc_stderr": 0.04832853553437055,
126
+ "acc_norm": 0.49074074074074076,
127
+ "acc_norm_stderr": 0.04832853553437055
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3793103448275862,
131
+ "acc_stderr": 0.03413963805906235,
132
+ "acc_norm": 0.3793103448275862,
133
+ "acc_norm_stderr": 0.03413963805906235
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.45161290322580644,
137
+ "acc_stderr": 0.02831050034856839,
138
+ "acc_norm": 0.45161290322580644,
139
+ "acc_norm_stderr": 0.02831050034856839
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6025641025641025,
143
+ "acc_stderr": 0.03205953453789293,
144
+ "acc_norm": 0.6025641025641025,
145
+ "acc_norm_stderr": 0.03205953453789293
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.39622641509433965,
149
+ "acc_stderr": 0.03010279378179119,
150
+ "acc_norm": 0.39622641509433965,
151
+ "acc_norm_stderr": 0.03010279378179119
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.42727272727272725,
155
+ "acc_stderr": 0.04738198703545483,
156
+ "acc_norm": 0.42727272727272725,
157
+ "acc_norm_stderr": 0.04738198703545483
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2740740740740741,
161
+ "acc_stderr": 0.027195934804085626,
162
+ "acc_norm": 0.2740740740740741,
163
+ "acc_norm_stderr": 0.027195934804085626
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.31788079470198677,
167
+ "acc_stderr": 0.03802039760107903,
168
+ "acc_norm": 0.31788079470198677,
169
+ "acc_norm_stderr": 0.03802039760107903
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5373134328358209,
173
+ "acc_stderr": 0.035256751674679745,
174
+ "acc_norm": 0.5373134328358209,
175
+ "acc_norm_stderr": 0.035256751674679745
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3352601156069364,
179
+ "acc_stderr": 0.03599586301247078,
180
+ "acc_norm": 0.3352601156069364,
181
+ "acc_norm_stderr": 0.03599586301247078
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.30687830687830686,
185
+ "acc_stderr": 0.02375292871211213,
186
+ "acc_norm": 0.30687830687830686,
187
+ "acc_norm_stderr": 0.02375292871211213
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3472222222222222,
191
+ "acc_stderr": 0.039812405437178615,
192
+ "acc_norm": 0.3472222222222222,
193
+ "acc_norm_stderr": 0.039812405437178615
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.21,
197
+ "acc_stderr": 0.040936018074033256,
198
+ "acc_norm": 0.21,
199
+ "acc_norm_stderr": 0.040936018074033256
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.55,
203
+ "acc_stderr": 0.049999999999999996,
204
+ "acc_norm": 0.55,
205
+ "acc_norm_stderr": 0.049999999999999996
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.47398843930635837,
209
+ "acc_stderr": 0.02688264343402289,
210
+ "acc_norm": 0.47398843930635837,
211
+ "acc_norm_stderr": 0.02688264343402289
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.37423312883435583,
215
+ "acc_stderr": 0.03802068102899615,
216
+ "acc_norm": 0.37423312883435583,
217
+ "acc_norm_stderr": 0.03802068102899615
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.3950617283950617,
221
+ "acc_stderr": 0.02720111766692566,
222
+ "acc_norm": 0.3950617283950617,
223
+ "acc_norm_stderr": 0.02720111766692566
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.35,
227
+ "acc_stderr": 0.047937248544110196,
228
+ "acc_norm": 0.35,
229
+ "acc_norm_stderr": 0.047937248544110196
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.41968911917098445,
233
+ "acc_stderr": 0.035615873276858834,
234
+ "acc_norm": 0.41968911917098445,
235
+ "acc_norm_stderr": 0.035615873276858834
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.24561403508771928,
239
+ "acc_stderr": 0.04049339297748142,
240
+ "acc_norm": 0.24561403508771928,
241
+ "acc_norm_stderr": 0.04049339297748142
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.3963302752293578,
245
+ "acc_stderr": 0.020971469947900525,
246
+ "acc_norm": 0.3963302752293578,
247
+ "acc_norm_stderr": 0.020971469947900525
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.30952380952380953,
251
+ "acc_stderr": 0.04134913018303316,
252
+ "acc_norm": 0.30952380952380953,
253
+ "acc_norm_stderr": 0.04134913018303316
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4084967320261438,
257
+ "acc_stderr": 0.028146405993096358,
258
+ "acc_norm": 0.4084967320261438,
259
+ "acc_norm_stderr": 0.028146405993096358
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.34,
263
+ "acc_stderr": 0.04760952285695235,
264
+ "acc_norm": 0.34,
265
+ "acc_norm_stderr": 0.04760952285695235
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6446280991735537,
269
+ "acc_stderr": 0.0436923632657398,
270
+ "acc_norm": 0.6446280991735537,
271
+ "acc_norm_stderr": 0.0436923632657398
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4473684210526316,
275
+ "acc_stderr": 0.04046336883978251,
276
+ "acc_norm": 0.4473684210526316,
277
+ "acc_norm_stderr": 0.04046336883978251
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.31699346405228757,
281
+ "acc_stderr": 0.018824219512706207,
282
+ "acc_norm": 0.31699346405228757,
283
+ "acc_norm_stderr": 0.018824219512706207
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.32269503546099293,
287
+ "acc_stderr": 0.027889139300534778,
288
+ "acc_norm": 0.32269503546099293,
289
+ "acc_norm_stderr": 0.027889139300534778
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.2857142857142857,
293
+ "acc_stderr": 0.04287858751340456,
294
+ "acc_norm": 0.2857142857142857,
295
+ "acc_norm_stderr": 0.04287858751340456
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.2638888888888889,
299
+ "acc_stderr": 0.030058202704309846,
300
+ "acc_norm": 0.2638888888888889,
301
+ "acc_norm_stderr": 0.030058202704309846
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2424581005586592,
305
+ "acc_stderr": 0.01433352205921789,
306
+ "acc_norm": 0.2424581005586592,
307
+ "acc_norm_stderr": 0.01433352205921789
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.39,
311
+ "acc_stderr": 0.04902071300001974,
312
+ "acc_norm": 0.39,
313
+ "acc_norm_stderr": 0.04902071300001974
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.47,
317
+ "acc_stderr": 0.05016135580465919,
318
+ "acc_norm": 0.47,
319
+ "acc_norm_stderr": 0.05016135580465919
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.29044117647058826,
323
+ "acc_stderr": 0.027576468622740505,
324
+ "acc_norm": 0.29044117647058826,
325
+ "acc_norm_stderr": 0.027576468622740505
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4163265306122449,
329
+ "acc_stderr": 0.031557828165561644,
330
+ "acc_norm": 0.4163265306122449,
331
+ "acc_norm_stderr": 0.031557828165561644
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5358649789029536,
335
+ "acc_stderr": 0.03246338898055659,
336
+ "acc_norm": 0.5358649789029536,
337
+ "acc_norm_stderr": 0.03246338898055659
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3200782268578879,
341
+ "acc_stderr": 0.011914791947638522,
342
+ "acc_norm": 0.3200782268578879,
343
+ "acc_norm_stderr": 0.011914791947638522
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.38235294117647056,
347
+ "acc_stderr": 0.034107853389047184,
348
+ "acc_norm": 0.38235294117647056,
349
+ "acc_norm_stderr": 0.034107853389047184
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.4909090909090909,
353
+ "acc_stderr": 0.03903698647748441,
354
+ "acc_norm": 0.4909090909090909,
355
+ "acc_norm_stderr": 0.03903698647748441
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2692778457772338,
359
+ "mc1_stderr": 0.015528566637087312,
360
+ "mc2": 0.4274629100267272,
361
+ "mc2_stderr": 0.015462888327553083
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3482880755608028,
365
+ "acc_stderr": 0.016379926739148044,
366
+ "acc_norm": 0.4132231404958678,
367
+ "acc_norm_stderr": 0.016929480234495232
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIFT/PACK-13b-v1.0",
436
+ "model_sha": "27f7b1eb3d926034aa90feb9ebc31788182046dd",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIFT/PACK-13b-v1.1/result_2023-12-11 08:43:39.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3361774744027304,
5
+ "acc_stderr": 0.013804855026205761,
6
+ "acc_norm": 0.38139931740614336,
7
+ "acc_norm_stderr": 0.014194389086685261
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3724357697669787,
11
+ "acc_stderr": 0.004824655406075561,
12
+ "acc_norm": 0.48078072097191793,
13
+ "acc_norm_stderr": 0.004986093791041656
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.45614035087719296,
17
+ "acc_stderr": 0.03820042586602966,
18
+ "acc_norm": 0.45614035087719296,
19
+ "acc_norm_stderr": 0.03820042586602966
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.49514563106796117,
23
+ "acc_stderr": 0.049505043821289195,
24
+ "acc_norm": 0.49514563106796117,
25
+ "acc_norm_stderr": 0.049505043821289195
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4891443167305236,
29
+ "acc_stderr": 0.017875748840242414,
30
+ "acc_norm": 0.4891443167305236,
31
+ "acc_norm_stderr": 0.017875748840242414
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.3851851851851852,
35
+ "acc_stderr": 0.042039210401562783,
36
+ "acc_norm": 0.3851851851851852,
37
+ "acc_norm_stderr": 0.042039210401562783
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.32,
41
+ "acc_stderr": 0.04688261722621504,
42
+ "acc_norm": 0.32,
43
+ "acc_norm_stderr": 0.04688261722621504
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.3574468085106383,
47
+ "acc_stderr": 0.03132941789476425,
48
+ "acc_norm": 0.3574468085106383,
49
+ "acc_norm_stderr": 0.03132941789476425
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.35542168674698793,
53
+ "acc_stderr": 0.037262143543224144,
54
+ "acc_norm": 0.35542168674698793,
55
+ "acc_norm_stderr": 0.037262143543224144
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4340836012861736,
59
+ "acc_stderr": 0.0281502322445356,
60
+ "acc_norm": 0.4340836012861736,
61
+ "acc_norm_stderr": 0.0281502322445356
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.3811659192825112,
65
+ "acc_stderr": 0.03259625118416827,
66
+ "acc_norm": 0.3811659192825112,
67
+ "acc_norm_stderr": 0.03259625118416827
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4732824427480916,
71
+ "acc_stderr": 0.04379024936553894,
72
+ "acc_norm": 0.4732824427480916,
73
+ "acc_norm_stderr": 0.04379024936553894
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.35,
77
+ "acc_stderr": 0.04793724854411022,
78
+ "acc_norm": 0.35,
79
+ "acc_norm_stderr": 0.04793724854411022
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5050505050505051,
83
+ "acc_stderr": 0.035621707606254015,
84
+ "acc_norm": 0.5050505050505051,
85
+ "acc_norm_stderr": 0.035621707606254015
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.41379310344827586,
89
+ "acc_stderr": 0.041042692118062316,
90
+ "acc_norm": 0.41379310344827586,
91
+ "acc_norm_stderr": 0.041042692118062316
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.17647058823529413,
95
+ "acc_stderr": 0.0379328118530781,
96
+ "acc_norm": 0.17647058823529413,
97
+ "acc_norm_stderr": 0.0379328118530781
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.3697478991596639,
101
+ "acc_stderr": 0.031357095996135904,
102
+ "acc_norm": 0.3697478991596639,
103
+ "acc_norm_stderr": 0.031357095996135904
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.36153846153846153,
107
+ "acc_stderr": 0.02435958146539696,
108
+ "acc_norm": 0.36153846153846153,
109
+ "acc_norm_stderr": 0.02435958146539696
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.57,
113
+ "acc_stderr": 0.049756985195624284,
114
+ "acc_norm": 0.57,
115
+ "acc_norm_stderr": 0.049756985195624284
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.31,
119
+ "acc_stderr": 0.04648231987117316,
120
+ "acc_norm": 0.31,
121
+ "acc_norm_stderr": 0.04648231987117316
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.4722222222222222,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.4722222222222222,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.37438423645320196,
131
+ "acc_stderr": 0.03405155380561952,
132
+ "acc_norm": 0.37438423645320196,
133
+ "acc_norm_stderr": 0.03405155380561952
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.44516129032258067,
137
+ "acc_stderr": 0.02827241018621491,
138
+ "acc_norm": 0.44516129032258067,
139
+ "acc_norm_stderr": 0.02827241018621491
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6196581196581197,
143
+ "acc_stderr": 0.03180425204384099,
144
+ "acc_norm": 0.6196581196581197,
145
+ "acc_norm_stderr": 0.03180425204384099
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.3622641509433962,
149
+ "acc_stderr": 0.029582245128384296,
150
+ "acc_norm": 0.3622641509433962,
151
+ "acc_norm_stderr": 0.029582245128384296
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.41818181818181815,
155
+ "acc_stderr": 0.04724577405731571,
156
+ "acc_norm": 0.41818181818181815,
157
+ "acc_norm_stderr": 0.04724577405731571
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.29259259259259257,
161
+ "acc_stderr": 0.02773896963217609,
162
+ "acc_norm": 0.29259259259259257,
163
+ "acc_norm_stderr": 0.02773896963217609
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.304635761589404,
167
+ "acc_stderr": 0.037579499229433426,
168
+ "acc_norm": 0.304635761589404,
169
+ "acc_norm_stderr": 0.037579499229433426
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5174129353233831,
173
+ "acc_stderr": 0.03533389234739245,
174
+ "acc_norm": 0.5174129353233831,
175
+ "acc_norm_stderr": 0.03533389234739245
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.32947976878612717,
179
+ "acc_stderr": 0.03583901754736412,
180
+ "acc_norm": 0.32947976878612717,
181
+ "acc_norm_stderr": 0.03583901754736412
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.30423280423280424,
185
+ "acc_stderr": 0.023695415009463087,
186
+ "acc_norm": 0.30423280423280424,
187
+ "acc_norm_stderr": 0.023695415009463087
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3472222222222222,
191
+ "acc_stderr": 0.039812405437178615,
192
+ "acc_norm": 0.3472222222222222,
193
+ "acc_norm_stderr": 0.039812405437178615
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.22,
197
+ "acc_stderr": 0.04163331998932269,
198
+ "acc_norm": 0.22,
199
+ "acc_norm_stderr": 0.04163331998932269
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.55,
203
+ "acc_stderr": 0.05,
204
+ "acc_norm": 0.55,
205
+ "acc_norm_stderr": 0.05
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.47398843930635837,
209
+ "acc_stderr": 0.02688264343402289,
210
+ "acc_norm": 0.47398843930635837,
211
+ "acc_norm_stderr": 0.02688264343402289
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.3803680981595092,
215
+ "acc_stderr": 0.038142698932618374,
216
+ "acc_norm": 0.3803680981595092,
217
+ "acc_norm_stderr": 0.038142698932618374
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.39814814814814814,
221
+ "acc_stderr": 0.027237415094592474,
222
+ "acc_norm": 0.39814814814814814,
223
+ "acc_norm_stderr": 0.027237415094592474
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.31,
227
+ "acc_stderr": 0.04648231987117316,
228
+ "acc_norm": 0.31,
229
+ "acc_norm_stderr": 0.04648231987117316
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.43523316062176165,
233
+ "acc_stderr": 0.03578038165008586,
234
+ "acc_norm": 0.43523316062176165,
235
+ "acc_norm_stderr": 0.03578038165008586
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2719298245614035,
239
+ "acc_stderr": 0.04185774424022057,
240
+ "acc_norm": 0.2719298245614035,
241
+ "acc_norm_stderr": 0.04185774424022057
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.3963302752293578,
245
+ "acc_stderr": 0.02097146994790053,
246
+ "acc_norm": 0.3963302752293578,
247
+ "acc_norm_stderr": 0.02097146994790053
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3253968253968254,
251
+ "acc_stderr": 0.041905964388711366,
252
+ "acc_norm": 0.3253968253968254,
253
+ "acc_norm_stderr": 0.041905964388711366
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.3888888888888889,
257
+ "acc_stderr": 0.02791405551046802,
258
+ "acc_norm": 0.3888888888888889,
259
+ "acc_norm_stderr": 0.02791405551046802
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.34,
263
+ "acc_stderr": 0.04760952285695235,
264
+ "acc_norm": 0.34,
265
+ "acc_norm_stderr": 0.04760952285695235
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6115702479338843,
269
+ "acc_stderr": 0.04449270350068383,
270
+ "acc_norm": 0.6115702479338843,
271
+ "acc_norm_stderr": 0.04449270350068383
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4342105263157895,
275
+ "acc_stderr": 0.04033565667848319,
276
+ "acc_norm": 0.4342105263157895,
277
+ "acc_norm_stderr": 0.04033565667848319
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.31862745098039214,
281
+ "acc_stderr": 0.01885008469646872,
282
+ "acc_norm": 0.31862745098039214,
283
+ "acc_norm_stderr": 0.01885008469646872
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3120567375886525,
287
+ "acc_stderr": 0.027640120545169945,
288
+ "acc_norm": 0.3120567375886525,
289
+ "acc_norm_stderr": 0.027640120545169945
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25892857142857145,
293
+ "acc_stderr": 0.041577515398656284,
294
+ "acc_norm": 0.25892857142857145,
295
+ "acc_norm_stderr": 0.041577515398656284
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.2824074074074074,
299
+ "acc_stderr": 0.03070137211151092,
300
+ "acc_norm": 0.2824074074074074,
301
+ "acc_norm_stderr": 0.03070137211151092
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2435754189944134,
305
+ "acc_stderr": 0.01435591196476786,
306
+ "acc_norm": 0.2435754189944134,
307
+ "acc_norm_stderr": 0.01435591196476786
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.38,
311
+ "acc_stderr": 0.04878317312145632,
312
+ "acc_norm": 0.38,
313
+ "acc_norm_stderr": 0.04878317312145632
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.48,
317
+ "acc_stderr": 0.050211673156867795,
318
+ "acc_norm": 0.48,
319
+ "acc_norm_stderr": 0.050211673156867795
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.29411764705882354,
323
+ "acc_stderr": 0.027678468642144686,
324
+ "acc_norm": 0.29411764705882354,
325
+ "acc_norm_stderr": 0.027678468642144686
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4163265306122449,
329
+ "acc_stderr": 0.031557828165561644,
330
+ "acc_norm": 0.4163265306122449,
331
+ "acc_norm_stderr": 0.031557828165561644
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.4936708860759494,
335
+ "acc_stderr": 0.032544620107678585,
336
+ "acc_norm": 0.4936708860759494,
337
+ "acc_norm_stderr": 0.032544620107678585
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.30182529335071706,
341
+ "acc_stderr": 0.01172435051810589,
342
+ "acc_norm": 0.30182529335071706,
343
+ "acc_norm_stderr": 0.01172435051810589
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.37745098039215685,
347
+ "acc_stderr": 0.03402272044340704,
348
+ "acc_norm": 0.37745098039215685,
349
+ "acc_norm_stderr": 0.03402272044340704
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.503030303030303,
353
+ "acc_stderr": 0.03904272341431856,
354
+ "acc_norm": 0.503030303030303,
355
+ "acc_norm_stderr": 0.03904272341431856
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2692778457772338,
359
+ "mc1_stderr": 0.015528566637087307,
360
+ "mc2": 0.42195295057052135,
361
+ "mc2_stderr": 0.015423294021851608
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3317591499409681,
365
+ "acc_stderr": 0.016187984642157312,
366
+ "acc_norm": 0.3955135773317591,
367
+ "acc_norm_stderr": 0.01681081590220604
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIFT/PACK-13b-v1.1",
436
+ "model_sha": "a547563032d1b762d80a80959f9b00aefab44eb5",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIFT/aift-llama2-koen-instruct-dpo-v1.01/result_2023-12-14 06:44:33.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.4069965870307167,
5
+ "acc_stderr": 0.014356399418009124,
6
+ "acc_norm": 0.46245733788395904,
7
+ "acc_norm_stderr": 0.01457014449507558
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.4193387771360287,
11
+ "acc_stderr": 0.00492442401807367,
12
+ "acc_norm": 0.5768771161123282,
13
+ "acc_norm_stderr": 0.004930448527146668
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.49122807017543857,
17
+ "acc_stderr": 0.038342347441649924,
18
+ "acc_norm": 0.49122807017543857,
19
+ "acc_norm_stderr": 0.038342347441649924
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.4563106796116505,
23
+ "acc_stderr": 0.049318019942204146,
24
+ "acc_norm": 0.4563106796116505,
25
+ "acc_norm_stderr": 0.049318019942204146
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5555555555555556,
29
+ "acc_stderr": 0.01776925058353325,
30
+ "acc_norm": 0.5555555555555556,
31
+ "acc_norm_stderr": 0.01776925058353325
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.5037037037037037,
35
+ "acc_stderr": 0.04319223625811331,
36
+ "acc_norm": 0.5037037037037037,
37
+ "acc_norm_stderr": 0.04319223625811331
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.3,
41
+ "acc_stderr": 0.046056618647183814,
42
+ "acc_norm": 0.3,
43
+ "acc_norm_stderr": 0.046056618647183814
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.4340425531914894,
47
+ "acc_stderr": 0.032400380867927465,
48
+ "acc_norm": 0.4340425531914894,
49
+ "acc_norm_stderr": 0.032400380867927465
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.37349397590361444,
53
+ "acc_stderr": 0.03765845117168862,
54
+ "acc_norm": 0.37349397590361444,
55
+ "acc_norm_stderr": 0.03765845117168862
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4887459807073955,
59
+ "acc_stderr": 0.028390897396863533,
60
+ "acc_norm": 0.4887459807073955,
61
+ "acc_norm_stderr": 0.028390897396863533
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.5022421524663677,
65
+ "acc_stderr": 0.033557465352232634,
66
+ "acc_norm": 0.5022421524663677,
67
+ "acc_norm_stderr": 0.033557465352232634
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.48091603053435117,
71
+ "acc_stderr": 0.04382094705550988,
72
+ "acc_norm": 0.48091603053435117,
73
+ "acc_norm_stderr": 0.04382094705550988
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.34,
77
+ "acc_stderr": 0.04760952285695236,
78
+ "acc_norm": 0.34,
79
+ "acc_norm_stderr": 0.04760952285695236
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5808080808080808,
83
+ "acc_stderr": 0.035155207286704175,
84
+ "acc_norm": 0.5808080808080808,
85
+ "acc_norm_stderr": 0.035155207286704175
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.3931034482758621,
89
+ "acc_stderr": 0.0407032901370707,
90
+ "acc_norm": 0.3931034482758621,
91
+ "acc_norm_stderr": 0.0407032901370707
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.21568627450980393,
95
+ "acc_stderr": 0.04092563958237657,
96
+ "acc_norm": 0.21568627450980393,
97
+ "acc_norm_stderr": 0.04092563958237657
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.44537815126050423,
101
+ "acc_stderr": 0.032284106267163895,
102
+ "acc_norm": 0.44537815126050423,
103
+ "acc_norm_stderr": 0.032284106267163895
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.41025641025641024,
107
+ "acc_stderr": 0.024939313906940777,
108
+ "acc_norm": 0.41025641025641024,
109
+ "acc_norm_stderr": 0.024939313906940777
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.55,
113
+ "acc_stderr": 0.04999999999999999,
114
+ "acc_norm": 0.55,
115
+ "acc_norm_stderr": 0.04999999999999999
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.34,
119
+ "acc_stderr": 0.04760952285695236,
120
+ "acc_norm": 0.34,
121
+ "acc_norm_stderr": 0.04760952285695236
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.48148148148148145,
125
+ "acc_stderr": 0.04830366024635331,
126
+ "acc_norm": 0.48148148148148145,
127
+ "acc_norm_stderr": 0.04830366024635331
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3793103448275862,
131
+ "acc_stderr": 0.03413963805906235,
132
+ "acc_norm": 0.3793103448275862,
133
+ "acc_norm_stderr": 0.03413963805906235
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.45161290322580644,
137
+ "acc_stderr": 0.02831050034856839,
138
+ "acc_norm": 0.45161290322580644,
139
+ "acc_norm_stderr": 0.02831050034856839
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6666666666666666,
143
+ "acc_stderr": 0.03088273697413865,
144
+ "acc_norm": 0.6666666666666666,
145
+ "acc_norm_stderr": 0.03088273697413865
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.3886792452830189,
149
+ "acc_stderr": 0.030000485448675986,
150
+ "acc_norm": 0.3886792452830189,
151
+ "acc_norm_stderr": 0.030000485448675986
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5363636363636364,
155
+ "acc_stderr": 0.04776449162396197,
156
+ "acc_norm": 0.5363636363636364,
157
+ "acc_norm_stderr": 0.04776449162396197
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2740740740740741,
161
+ "acc_stderr": 0.027195934804085622,
162
+ "acc_norm": 0.2740740740740741,
163
+ "acc_norm_stderr": 0.027195934804085622
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2847682119205298,
167
+ "acc_stderr": 0.03684881521389023,
168
+ "acc_norm": 0.2847682119205298,
169
+ "acc_norm_stderr": 0.03684881521389023
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5472636815920398,
173
+ "acc_stderr": 0.03519702717576915,
174
+ "acc_norm": 0.5472636815920398,
175
+ "acc_norm_stderr": 0.03519702717576915
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3815028901734104,
179
+ "acc_stderr": 0.03703851193099521,
180
+ "acc_norm": 0.3815028901734104,
181
+ "acc_norm_stderr": 0.03703851193099521
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.30158730158730157,
185
+ "acc_stderr": 0.023636975996101806,
186
+ "acc_norm": 0.30158730158730157,
187
+ "acc_norm_stderr": 0.023636975996101806
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4444444444444444,
191
+ "acc_stderr": 0.041553199555931467,
192
+ "acc_norm": 0.4444444444444444,
193
+ "acc_norm_stderr": 0.041553199555931467
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.36,
197
+ "acc_stderr": 0.04824181513244218,
198
+ "acc_norm": 0.36,
199
+ "acc_norm_stderr": 0.04824181513244218
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.58,
203
+ "acc_stderr": 0.04960449637488584,
204
+ "acc_norm": 0.58,
205
+ "acc_norm_stderr": 0.04960449637488584
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5028901734104047,
209
+ "acc_stderr": 0.026918645383239015,
210
+ "acc_norm": 0.5028901734104047,
211
+ "acc_norm_stderr": 0.026918645383239015
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.48466257668711654,
215
+ "acc_stderr": 0.03926522378708843,
216
+ "acc_norm": 0.48466257668711654,
217
+ "acc_norm_stderr": 0.03926522378708843
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4691358024691358,
221
+ "acc_stderr": 0.027767689606833935,
222
+ "acc_norm": 0.4691358024691358,
223
+ "acc_norm_stderr": 0.027767689606833935
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.31,
227
+ "acc_stderr": 0.046482319871173156,
228
+ "acc_norm": 0.31,
229
+ "acc_norm_stderr": 0.046482319871173156
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5077720207253886,
233
+ "acc_stderr": 0.03608003225569654,
234
+ "acc_norm": 0.5077720207253886,
235
+ "acc_norm_stderr": 0.03608003225569654
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2631578947368421,
239
+ "acc_stderr": 0.0414243971948936,
240
+ "acc_norm": 0.2631578947368421,
241
+ "acc_norm_stderr": 0.0414243971948936
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5541284403669725,
245
+ "acc_stderr": 0.02131133500970858,
246
+ "acc_norm": 0.5541284403669725,
247
+ "acc_norm_stderr": 0.02131133500970858
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.29365079365079366,
251
+ "acc_stderr": 0.04073524322147124,
252
+ "acc_norm": 0.29365079365079366,
253
+ "acc_norm_stderr": 0.04073524322147124
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.43137254901960786,
257
+ "acc_stderr": 0.028358956313423552,
258
+ "acc_norm": 0.43137254901960786,
259
+ "acc_norm_stderr": 0.028358956313423552
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.42,
263
+ "acc_stderr": 0.049604496374885836,
264
+ "acc_norm": 0.42,
265
+ "acc_norm_stderr": 0.049604496374885836
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.628099173553719,
269
+ "acc_stderr": 0.044120158066245044,
270
+ "acc_norm": 0.628099173553719,
271
+ "acc_norm_stderr": 0.044120158066245044
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4407894736842105,
275
+ "acc_stderr": 0.04040311062490437,
276
+ "acc_norm": 0.4407894736842105,
277
+ "acc_norm_stderr": 0.04040311062490437
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3839869281045752,
281
+ "acc_stderr": 0.019675808135281515,
282
+ "acc_norm": 0.3839869281045752,
283
+ "acc_norm_stderr": 0.019675808135281515
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.30141843971631205,
287
+ "acc_stderr": 0.02737412888263115,
288
+ "acc_norm": 0.30141843971631205,
289
+ "acc_norm_stderr": 0.02737412888263115
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25,
293
+ "acc_stderr": 0.04109974682633932,
294
+ "acc_norm": 0.25,
295
+ "acc_norm_stderr": 0.04109974682633932
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.2962962962962963,
299
+ "acc_stderr": 0.03114144782353604,
300
+ "acc_norm": 0.2962962962962963,
301
+ "acc_norm_stderr": 0.03114144782353604
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2569832402234637,
305
+ "acc_stderr": 0.014614465821966346,
306
+ "acc_norm": 0.2569832402234637,
307
+ "acc_norm_stderr": 0.014614465821966346
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.34,
311
+ "acc_stderr": 0.04760952285695236,
312
+ "acc_norm": 0.34,
313
+ "acc_norm_stderr": 0.04760952285695236
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.46,
317
+ "acc_stderr": 0.05009082659620333,
318
+ "acc_norm": 0.46,
319
+ "acc_norm_stderr": 0.05009082659620333
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3639705882352941,
323
+ "acc_stderr": 0.02922719246003203,
324
+ "acc_norm": 0.3639705882352941,
325
+ "acc_norm_stderr": 0.02922719246003203
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.37142857142857144,
329
+ "acc_stderr": 0.030932858792789848,
330
+ "acc_norm": 0.37142857142857144,
331
+ "acc_norm_stderr": 0.030932858792789848
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6244725738396625,
335
+ "acc_stderr": 0.03152256243091156,
336
+ "acc_norm": 0.6244725738396625,
337
+ "acc_norm_stderr": 0.03152256243091156
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.34028683181225555,
341
+ "acc_stderr": 0.012101217610223782,
342
+ "acc_norm": 0.34028683181225555,
343
+ "acc_norm_stderr": 0.012101217610223782
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.5,
347
+ "acc_stderr": 0.03509312031717982,
348
+ "acc_norm": 0.5,
349
+ "acc_norm_stderr": 0.03509312031717982
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5575757575757576,
353
+ "acc_stderr": 0.03878372113711275,
354
+ "acc_norm": 0.5575757575757576,
355
+ "acc_norm_stderr": 0.03878372113711275
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.24724602203182375,
359
+ "mc1_stderr": 0.01510240479735965,
360
+ "mc2": 0.40156672902861484,
361
+ "mc2_stderr": 0.015311585666350696
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3754427390791027,
365
+ "acc_stderr": 0.01664841158951109,
366
+ "acc_norm": 0.38134592680047225,
367
+ "acc_norm_stderr": 0.016699301768828084
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIFT/aift-llama2-koen-instruct-dpo-v1.01",
436
+ "model_sha": "1c1e407910427042580cb11ad8569567c769cf49",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIFT/aift-llama2-koen-instruct-dpo-v1.02/result_2023-12-15 03:06:44.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.4052901023890785,
5
+ "acc_stderr": 0.014346869060229327,
6
+ "acc_norm": 0.4658703071672355,
7
+ "acc_norm_stderr": 0.014577311315231097
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.4192391953794065,
11
+ "acc_stderr": 0.004924261467934422,
12
+ "acc_norm": 0.5763792073292173,
13
+ "acc_norm_stderr": 0.004931219148182245
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5087719298245614,
17
+ "acc_stderr": 0.038342347441649924,
18
+ "acc_norm": 0.5087719298245614,
19
+ "acc_norm_stderr": 0.038342347441649924
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5242718446601942,
23
+ "acc_stderr": 0.049449010929737795,
24
+ "acc_norm": 0.5242718446601942,
25
+ "acc_norm_stderr": 0.049449010929737795
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.565772669220945,
29
+ "acc_stderr": 0.017724589389677785,
30
+ "acc_norm": 0.565772669220945,
31
+ "acc_norm_stderr": 0.017724589389677785
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.5185185185185185,
35
+ "acc_stderr": 0.043163785995113245,
36
+ "acc_norm": 0.5185185185185185,
37
+ "acc_norm_stderr": 0.043163785995113245
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.35,
41
+ "acc_stderr": 0.0479372485441102,
42
+ "acc_norm": 0.35,
43
+ "acc_norm_stderr": 0.0479372485441102
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.4,
47
+ "acc_stderr": 0.032025630761017346,
48
+ "acc_norm": 0.4,
49
+ "acc_norm_stderr": 0.032025630761017346
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3855421686746988,
53
+ "acc_stderr": 0.037891344246115476,
54
+ "acc_norm": 0.3855421686746988,
55
+ "acc_norm_stderr": 0.037891344246115476
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5112540192926045,
59
+ "acc_stderr": 0.028390897396863526,
60
+ "acc_norm": 0.5112540192926045,
61
+ "acc_norm_stderr": 0.028390897396863526
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.48878923766816146,
65
+ "acc_stderr": 0.033549366530984746,
66
+ "acc_norm": 0.48878923766816146,
67
+ "acc_norm_stderr": 0.033549366530984746
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4961832061068702,
71
+ "acc_stderr": 0.04385162325601553,
72
+ "acc_norm": 0.4961832061068702,
73
+ "acc_norm_stderr": 0.04385162325601553
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.32,
77
+ "acc_stderr": 0.04688261722621504,
78
+ "acc_norm": 0.32,
79
+ "acc_norm_stderr": 0.04688261722621504
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5959595959595959,
83
+ "acc_stderr": 0.03496130972056127,
84
+ "acc_norm": 0.5959595959595959,
85
+ "acc_norm_stderr": 0.03496130972056127
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4206896551724138,
89
+ "acc_stderr": 0.0411391498118926,
90
+ "acc_norm": 0.4206896551724138,
91
+ "acc_norm_stderr": 0.0411391498118926
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.17647058823529413,
95
+ "acc_stderr": 0.0379328118530781,
96
+ "acc_norm": 0.17647058823529413,
97
+ "acc_norm_stderr": 0.0379328118530781
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.48739495798319327,
101
+ "acc_stderr": 0.03246816765752174,
102
+ "acc_norm": 0.48739495798319327,
103
+ "acc_norm_stderr": 0.03246816765752174
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4358974358974359,
107
+ "acc_stderr": 0.025141801511177498,
108
+ "acc_norm": 0.4358974358974359,
109
+ "acc_norm_stderr": 0.025141801511177498
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.56,
113
+ "acc_stderr": 0.04988876515698589,
114
+ "acc_norm": 0.56,
115
+ "acc_norm_stderr": 0.04988876515698589
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.34,
119
+ "acc_stderr": 0.04760952285695235,
120
+ "acc_norm": 0.34,
121
+ "acc_norm_stderr": 0.04760952285695235
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.4351851851851852,
125
+ "acc_stderr": 0.04792898170907061,
126
+ "acc_norm": 0.4351851851851852,
127
+ "acc_norm_stderr": 0.04792898170907061
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.35960591133004927,
131
+ "acc_stderr": 0.03376458246509567,
132
+ "acc_norm": 0.35960591133004927,
133
+ "acc_norm_stderr": 0.03376458246509567
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4774193548387097,
137
+ "acc_stderr": 0.028414985019707868,
138
+ "acc_norm": 0.4774193548387097,
139
+ "acc_norm_stderr": 0.028414985019707868
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6837606837606838,
143
+ "acc_stderr": 0.03046365674734026,
144
+ "acc_norm": 0.6837606837606838,
145
+ "acc_norm_stderr": 0.03046365674734026
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.44528301886792454,
149
+ "acc_stderr": 0.030588052974270655,
150
+ "acc_norm": 0.44528301886792454,
151
+ "acc_norm_stderr": 0.030588052974270655
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5727272727272728,
155
+ "acc_stderr": 0.047381987035454834,
156
+ "acc_norm": 0.5727272727272728,
157
+ "acc_norm_stderr": 0.047381987035454834
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.25555555555555554,
161
+ "acc_stderr": 0.026593939101844086,
162
+ "acc_norm": 0.25555555555555554,
163
+ "acc_norm_stderr": 0.026593939101844086
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2913907284768212,
167
+ "acc_stderr": 0.037101857261199946,
168
+ "acc_norm": 0.2913907284768212,
169
+ "acc_norm_stderr": 0.037101857261199946
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5522388059701493,
173
+ "acc_stderr": 0.03516184772952167,
174
+ "acc_norm": 0.5522388059701493,
175
+ "acc_norm_stderr": 0.03516184772952167
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.4046242774566474,
179
+ "acc_stderr": 0.03742461193887248,
180
+ "acc_norm": 0.4046242774566474,
181
+ "acc_norm_stderr": 0.03742461193887248
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.30952380952380953,
185
+ "acc_stderr": 0.023809523809523857,
186
+ "acc_norm": 0.30952380952380953,
187
+ "acc_norm_stderr": 0.023809523809523857
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4097222222222222,
191
+ "acc_stderr": 0.04112490974670787,
192
+ "acc_norm": 0.4097222222222222,
193
+ "acc_norm_stderr": 0.04112490974670787
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.36,
197
+ "acc_stderr": 0.04824181513244218,
198
+ "acc_norm": 0.36,
199
+ "acc_norm_stderr": 0.04824181513244218
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.63,
203
+ "acc_stderr": 0.04852365870939099,
204
+ "acc_norm": 0.63,
205
+ "acc_norm_stderr": 0.04852365870939099
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5289017341040463,
209
+ "acc_stderr": 0.02687408588351835,
210
+ "acc_norm": 0.5289017341040463,
211
+ "acc_norm_stderr": 0.02687408588351835
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.48466257668711654,
215
+ "acc_stderr": 0.03926522378708843,
216
+ "acc_norm": 0.48466257668711654,
217
+ "acc_norm_stderr": 0.03926522378708843
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4845679012345679,
221
+ "acc_stderr": 0.0278074900442762,
222
+ "acc_norm": 0.4845679012345679,
223
+ "acc_norm_stderr": 0.0278074900442762
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.34,
227
+ "acc_stderr": 0.04760952285695235,
228
+ "acc_norm": 0.34,
229
+ "acc_norm_stderr": 0.04760952285695235
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5233160621761658,
233
+ "acc_stderr": 0.03604513672442202,
234
+ "acc_norm": 0.5233160621761658,
235
+ "acc_norm_stderr": 0.03604513672442202
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2543859649122807,
239
+ "acc_stderr": 0.04096985139843671,
240
+ "acc_norm": 0.2543859649122807,
241
+ "acc_norm_stderr": 0.04096985139843671
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5798165137614679,
245
+ "acc_stderr": 0.021162420048273515,
246
+ "acc_norm": 0.5798165137614679,
247
+ "acc_norm_stderr": 0.021162420048273515
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.2777777777777778,
251
+ "acc_stderr": 0.04006168083848878,
252
+ "acc_norm": 0.2777777777777778,
253
+ "acc_norm_stderr": 0.04006168083848878
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.45751633986928103,
257
+ "acc_stderr": 0.02852638345214263,
258
+ "acc_norm": 0.45751633986928103,
259
+ "acc_norm_stderr": 0.02852638345214263
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.45,
263
+ "acc_stderr": 0.05,
264
+ "acc_norm": 0.45,
265
+ "acc_norm_stderr": 0.05
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6694214876033058,
269
+ "acc_stderr": 0.04294340845212094,
270
+ "acc_norm": 0.6694214876033058,
271
+ "acc_norm_stderr": 0.04294340845212094
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4473684210526316,
275
+ "acc_stderr": 0.0404633688397825,
276
+ "acc_norm": 0.4473684210526316,
277
+ "acc_norm_stderr": 0.0404633688397825
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.39705882352941174,
281
+ "acc_stderr": 0.019794488900024113,
282
+ "acc_norm": 0.39705882352941174,
283
+ "acc_norm_stderr": 0.019794488900024113
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.30141843971631205,
287
+ "acc_stderr": 0.02737412888263115,
288
+ "acc_norm": 0.30141843971631205,
289
+ "acc_norm_stderr": 0.02737412888263115
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.23214285714285715,
293
+ "acc_stderr": 0.04007341809755806,
294
+ "acc_norm": 0.23214285714285715,
295
+ "acc_norm_stderr": 0.04007341809755806
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.33796296296296297,
299
+ "acc_stderr": 0.03225941352631295,
300
+ "acc_norm": 0.33796296296296297,
301
+ "acc_norm_stderr": 0.03225941352631295
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2670391061452514,
305
+ "acc_stderr": 0.014796502622562546,
306
+ "acc_norm": 0.2670391061452514,
307
+ "acc_norm_stderr": 0.014796502622562546
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.32,
311
+ "acc_stderr": 0.046882617226215034,
312
+ "acc_norm": 0.32,
313
+ "acc_norm_stderr": 0.046882617226215034
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.46,
317
+ "acc_stderr": 0.05009082659620333,
318
+ "acc_norm": 0.46,
319
+ "acc_norm_stderr": 0.05009082659620333
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.39705882352941174,
323
+ "acc_stderr": 0.029722152099280055,
324
+ "acc_norm": 0.39705882352941174,
325
+ "acc_norm_stderr": 0.029722152099280055
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4326530612244898,
329
+ "acc_stderr": 0.03171752824062664,
330
+ "acc_norm": 0.4326530612244898,
331
+ "acc_norm_stderr": 0.03171752824062664
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6286919831223629,
335
+ "acc_stderr": 0.03145068600744859,
336
+ "acc_norm": 0.6286919831223629,
337
+ "acc_norm_stderr": 0.03145068600744859
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.34028683181225555,
341
+ "acc_stderr": 0.01210121761022378,
342
+ "acc_norm": 0.34028683181225555,
343
+ "acc_norm_stderr": 0.01210121761022378
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.5,
347
+ "acc_stderr": 0.03509312031717982,
348
+ "acc_norm": 0.5,
349
+ "acc_norm_stderr": 0.03509312031717982
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5757575757575758,
353
+ "acc_stderr": 0.038592681420702636,
354
+ "acc_norm": 0.5757575757575758,
355
+ "acc_norm_stderr": 0.038592681420702636
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.23990208078335373,
359
+ "mc1_stderr": 0.01494881267906214,
360
+ "mc2": 0.39165785820787247,
361
+ "mc2_stderr": 0.015096702357183963
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3577331759149941,
365
+ "acc_stderr": 0.016479808935749976,
366
+ "acc_norm": 0.36481700118063753,
367
+ "acc_norm_stderr": 0.01655014433704659
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIFT/aift-llama2-koen-instruct-dpo-v1.02",
436
+ "model_sha": "9e9887d8579e1d19943d9d10f0d340620328c852",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIFT/aift-llama2-koen-instruct-v1.0/result_2023-12-14 00:45:21.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3771331058020478,
5
+ "acc_stderr": 0.0141633668961926,
6
+ "acc_norm": 0.4351535836177474,
7
+ "acc_norm_stderr": 0.01448798619718605
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.4151563433578968,
11
+ "acc_stderr": 0.004917419367766031,
12
+ "acc_norm": 0.5669189404501095,
13
+ "acc_norm_stderr": 0.004944889545497955
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.49122807017543857,
17
+ "acc_stderr": 0.038342347441649924,
18
+ "acc_norm": 0.49122807017543857,
19
+ "acc_norm_stderr": 0.038342347441649924
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5242718446601942,
23
+ "acc_stderr": 0.049449010929737795,
24
+ "acc_norm": 0.5242718446601942,
25
+ "acc_norm_stderr": 0.049449010929737795
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5491698595146871,
29
+ "acc_stderr": 0.017793297572699034,
30
+ "acc_norm": 0.5491698595146871,
31
+ "acc_norm_stderr": 0.017793297572699034
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.5111111111111111,
35
+ "acc_stderr": 0.04318275491977976,
36
+ "acc_norm": 0.5111111111111111,
37
+ "acc_norm_stderr": 0.04318275491977976
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.35,
41
+ "acc_stderr": 0.0479372485441102,
42
+ "acc_norm": 0.35,
43
+ "acc_norm_stderr": 0.0479372485441102
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.39148936170212767,
47
+ "acc_stderr": 0.03190701242326812,
48
+ "acc_norm": 0.39148936170212767,
49
+ "acc_norm_stderr": 0.03190701242326812
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3855421686746988,
53
+ "acc_stderr": 0.037891344246115476,
54
+ "acc_norm": 0.3855421686746988,
55
+ "acc_norm_stderr": 0.037891344246115476
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5337620578778135,
59
+ "acc_stderr": 0.028333277109562804,
60
+ "acc_norm": 0.5337620578778135,
61
+ "acc_norm_stderr": 0.028333277109562804
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.5067264573991032,
65
+ "acc_stderr": 0.03355476596234354,
66
+ "acc_norm": 0.5067264573991032,
67
+ "acc_norm_stderr": 0.03355476596234354
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.5038167938931297,
71
+ "acc_stderr": 0.04385162325601553,
72
+ "acc_norm": 0.5038167938931297,
73
+ "acc_norm_stderr": 0.04385162325601553
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.42,
77
+ "acc_stderr": 0.049604496374885836,
78
+ "acc_norm": 0.42,
79
+ "acc_norm_stderr": 0.049604496374885836
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5858585858585859,
83
+ "acc_stderr": 0.03509438348879629,
84
+ "acc_norm": 0.5858585858585859,
85
+ "acc_norm_stderr": 0.03509438348879629
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.41379310344827586,
89
+ "acc_stderr": 0.04104269211806232,
90
+ "acc_norm": 0.41379310344827586,
91
+ "acc_norm_stderr": 0.04104269211806232
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.21568627450980393,
95
+ "acc_stderr": 0.04092563958237655,
96
+ "acc_norm": 0.21568627450980393,
97
+ "acc_norm_stderr": 0.04092563958237655
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.4579831932773109,
101
+ "acc_stderr": 0.03236361111951941,
102
+ "acc_norm": 0.4579831932773109,
103
+ "acc_norm_stderr": 0.03236361111951941
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4358974358974359,
107
+ "acc_stderr": 0.025141801511177498,
108
+ "acc_norm": 0.4358974358974359,
109
+ "acc_norm_stderr": 0.025141801511177498
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.53,
113
+ "acc_stderr": 0.05016135580465919,
114
+ "acc_norm": 0.53,
115
+ "acc_norm_stderr": 0.05016135580465919
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.37,
119
+ "acc_stderr": 0.048523658709391,
120
+ "acc_norm": 0.37,
121
+ "acc_norm_stderr": 0.048523658709391
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.4351851851851852,
125
+ "acc_stderr": 0.04792898170907061,
126
+ "acc_norm": 0.4351851851851852,
127
+ "acc_norm_stderr": 0.04792898170907061
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.39408866995073893,
131
+ "acc_stderr": 0.034381579670365446,
132
+ "acc_norm": 0.39408866995073893,
133
+ "acc_norm_stderr": 0.034381579670365446
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.47419354838709676,
137
+ "acc_stderr": 0.02840609505765332,
138
+ "acc_norm": 0.47419354838709676,
139
+ "acc_norm_stderr": 0.02840609505765332
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6581196581196581,
143
+ "acc_stderr": 0.03107502852650775,
144
+ "acc_norm": 0.6581196581196581,
145
+ "acc_norm_stderr": 0.03107502852650775
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4188679245283019,
149
+ "acc_stderr": 0.030365050829115208,
150
+ "acc_norm": 0.4188679245283019,
151
+ "acc_norm_stderr": 0.030365050829115208
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4909090909090909,
155
+ "acc_stderr": 0.04788339768702861,
156
+ "acc_norm": 0.4909090909090909,
157
+ "acc_norm_stderr": 0.04788339768702861
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2222222222222222,
161
+ "acc_stderr": 0.025348097468097856,
162
+ "acc_norm": 0.2222222222222222,
163
+ "acc_norm_stderr": 0.025348097468097856
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.32450331125827814,
167
+ "acc_stderr": 0.038227469376587525,
168
+ "acc_norm": 0.32450331125827814,
169
+ "acc_norm_stderr": 0.038227469376587525
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5472636815920398,
173
+ "acc_stderr": 0.03519702717576915,
174
+ "acc_norm": 0.5472636815920398,
175
+ "acc_norm_stderr": 0.03519702717576915
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.4161849710982659,
179
+ "acc_stderr": 0.03758517775404947,
180
+ "acc_norm": 0.4161849710982659,
181
+ "acc_norm_stderr": 0.03758517775404947
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.29365079365079366,
185
+ "acc_stderr": 0.02345603738398202,
186
+ "acc_norm": 0.29365079365079366,
187
+ "acc_norm_stderr": 0.02345603738398202
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4027777777777778,
191
+ "acc_stderr": 0.04101405519842426,
192
+ "acc_norm": 0.4027777777777778,
193
+ "acc_norm_stderr": 0.04101405519842426
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.35,
197
+ "acc_stderr": 0.0479372485441102,
198
+ "acc_norm": 0.35,
199
+ "acc_norm_stderr": 0.0479372485441102
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.61,
203
+ "acc_stderr": 0.04902071300001975,
204
+ "acc_norm": 0.61,
205
+ "acc_norm_stderr": 0.04902071300001975
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5144508670520231,
209
+ "acc_stderr": 0.026907849856282532,
210
+ "acc_norm": 0.5144508670520231,
211
+ "acc_norm_stderr": 0.026907849856282532
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4723926380368098,
215
+ "acc_stderr": 0.039223782906109894,
216
+ "acc_norm": 0.4723926380368098,
217
+ "acc_norm_stderr": 0.039223782906109894
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4876543209876543,
221
+ "acc_stderr": 0.027812262269327235,
222
+ "acc_norm": 0.4876543209876543,
223
+ "acc_norm_stderr": 0.027812262269327235
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.3,
227
+ "acc_stderr": 0.04605661864718381,
228
+ "acc_norm": 0.3,
229
+ "acc_norm_stderr": 0.04605661864718381
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5284974093264249,
233
+ "acc_stderr": 0.036025735712884414,
234
+ "acc_norm": 0.5284974093264249,
235
+ "acc_norm_stderr": 0.036025735712884414
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.22807017543859648,
239
+ "acc_stderr": 0.03947152782669415,
240
+ "acc_norm": 0.22807017543859648,
241
+ "acc_norm_stderr": 0.03947152782669415
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5596330275229358,
245
+ "acc_stderr": 0.021284310623761536,
246
+ "acc_norm": 0.5596330275229358,
247
+ "acc_norm_stderr": 0.021284310623761536
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.30952380952380953,
251
+ "acc_stderr": 0.04134913018303316,
252
+ "acc_norm": 0.30952380952380953,
253
+ "acc_norm_stderr": 0.04134913018303316
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.46078431372549017,
257
+ "acc_stderr": 0.028541722692618874,
258
+ "acc_norm": 0.46078431372549017,
259
+ "acc_norm_stderr": 0.028541722692618874
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.41,
263
+ "acc_stderr": 0.049431107042371025,
264
+ "acc_norm": 0.41,
265
+ "acc_norm_stderr": 0.049431107042371025
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6363636363636364,
269
+ "acc_stderr": 0.043913262867240704,
270
+ "acc_norm": 0.6363636363636364,
271
+ "acc_norm_stderr": 0.043913262867240704
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40131578947368424,
275
+ "acc_stderr": 0.03988903703336284,
276
+ "acc_norm": 0.40131578947368424,
277
+ "acc_norm_stderr": 0.03988903703336284
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.4035947712418301,
281
+ "acc_stderr": 0.019848280168401147,
282
+ "acc_norm": 0.4035947712418301,
283
+ "acc_norm_stderr": 0.019848280168401147
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.30851063829787234,
287
+ "acc_stderr": 0.027553366165101376,
288
+ "acc_norm": 0.30851063829787234,
289
+ "acc_norm_stderr": 0.027553366165101376
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.23214285714285715,
293
+ "acc_stderr": 0.04007341809755806,
294
+ "acc_norm": 0.23214285714285715,
295
+ "acc_norm_stderr": 0.04007341809755806
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3611111111111111,
299
+ "acc_stderr": 0.032757734861009996,
300
+ "acc_norm": 0.3611111111111111,
301
+ "acc_norm_stderr": 0.032757734861009996
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2435754189944134,
305
+ "acc_stderr": 0.014355911964767857,
306
+ "acc_norm": 0.2435754189944134,
307
+ "acc_norm_stderr": 0.014355911964767857
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.4,
311
+ "acc_stderr": 0.049236596391733084,
312
+ "acc_norm": 0.4,
313
+ "acc_norm_stderr": 0.049236596391733084
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.45,
317
+ "acc_stderr": 0.049999999999999996,
318
+ "acc_norm": 0.45,
319
+ "acc_norm_stderr": 0.049999999999999996
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.4338235294117647,
323
+ "acc_stderr": 0.030105636570016643,
324
+ "acc_norm": 0.4338235294117647,
325
+ "acc_norm_stderr": 0.030105636570016643
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.46530612244897956,
329
+ "acc_stderr": 0.03193207024425314,
330
+ "acc_norm": 0.46530612244897956,
331
+ "acc_norm_stderr": 0.03193207024425314
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5822784810126582,
335
+ "acc_stderr": 0.032103530322412685,
336
+ "acc_norm": 0.5822784810126582,
337
+ "acc_norm_stderr": 0.032103530322412685
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.34615384615384615,
341
+ "acc_stderr": 0.012150699768228563,
342
+ "acc_norm": 0.34615384615384615,
343
+ "acc_norm_stderr": 0.012150699768228563
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.46568627450980393,
347
+ "acc_stderr": 0.03501038327635897,
348
+ "acc_norm": 0.46568627450980393,
349
+ "acc_norm_stderr": 0.03501038327635897
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5515151515151515,
353
+ "acc_stderr": 0.038835659779569286,
354
+ "acc_norm": 0.5515151515151515,
355
+ "acc_norm_stderr": 0.038835659779569286
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.24724602203182375,
359
+ "mc1_stderr": 0.015102404797359649,
360
+ "mc2": 0.39725650408832863,
361
+ "mc2_stderr": 0.01469261681765968
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.4179456906729634,
365
+ "acc_stderr": 0.016957292005279713,
366
+ "acc_norm": 0.4817001180637544,
367
+ "acc_norm_stderr": 0.01717883663917776
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIFT/aift-llama2-koen-instruct-v1.0",
436
+ "model_sha": "54a5a30188cba6af653f20df22ff393472f0e161",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIFT/aift-llama2-koen-instruct-v1.1-dpo-test1/result_2023-12-18 03:42:48.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3822525597269625,
5
+ "acc_stderr": 0.014200454049979279,
6
+ "acc_norm": 0.44112627986348124,
7
+ "acc_norm_stderr": 0.014509747749064664
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.41943835889265085,
11
+ "acc_stderr": 0.004924586362301652,
12
+ "acc_norm": 0.5719976100378411,
13
+ "acc_norm_stderr": 0.004937779821908573
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4678362573099415,
17
+ "acc_stderr": 0.038268824176603704,
18
+ "acc_norm": 0.4678362573099415,
19
+ "acc_norm_stderr": 0.038268824176603704
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5145631067961165,
23
+ "acc_stderr": 0.04948637324026637,
24
+ "acc_norm": 0.5145631067961165,
25
+ "acc_norm_stderr": 0.04948637324026637
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5555555555555556,
29
+ "acc_stderr": 0.01776925058353325,
30
+ "acc_norm": 0.5555555555555556,
31
+ "acc_norm_stderr": 0.01776925058353325
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.48148148148148145,
35
+ "acc_stderr": 0.043163785995113245,
36
+ "acc_norm": 0.48148148148148145,
37
+ "acc_norm_stderr": 0.043163785995113245
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.35,
41
+ "acc_stderr": 0.047937248544110196,
42
+ "acc_norm": 0.35,
43
+ "acc_norm_stderr": 0.047937248544110196
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.425531914893617,
47
+ "acc_stderr": 0.03232146916224468,
48
+ "acc_norm": 0.425531914893617,
49
+ "acc_norm_stderr": 0.03232146916224468
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.40963855421686746,
53
+ "acc_stderr": 0.038284011150790206,
54
+ "acc_norm": 0.40963855421686746,
55
+ "acc_norm_stderr": 0.038284011150790206
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5176848874598071,
59
+ "acc_stderr": 0.028380322849077138,
60
+ "acc_norm": 0.5176848874598071,
61
+ "acc_norm_stderr": 0.028380322849077138
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.515695067264574,
65
+ "acc_stderr": 0.0335412657542081,
66
+ "acc_norm": 0.515695067264574,
67
+ "acc_norm_stderr": 0.0335412657542081
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.5190839694656488,
71
+ "acc_stderr": 0.043820947055509867,
72
+ "acc_norm": 0.5190839694656488,
73
+ "acc_norm_stderr": 0.043820947055509867
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.39,
77
+ "acc_stderr": 0.04902071300001974,
78
+ "acc_norm": 0.39,
79
+ "acc_norm_stderr": 0.04902071300001974
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5707070707070707,
83
+ "acc_stderr": 0.03526552724601198,
84
+ "acc_norm": 0.5707070707070707,
85
+ "acc_norm_stderr": 0.03526552724601198
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.43448275862068964,
89
+ "acc_stderr": 0.041307408795554966,
90
+ "acc_norm": 0.43448275862068964,
91
+ "acc_norm_stderr": 0.041307408795554966
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.19607843137254902,
95
+ "acc_stderr": 0.039505818611799616,
96
+ "acc_norm": 0.19607843137254902,
97
+ "acc_norm_stderr": 0.039505818611799616
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.47058823529411764,
101
+ "acc_stderr": 0.032422250271150074,
102
+ "acc_norm": 0.47058823529411764,
103
+ "acc_norm_stderr": 0.032422250271150074
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.40512820512820513,
107
+ "acc_stderr": 0.02489047176993815,
108
+ "acc_norm": 0.40512820512820513,
109
+ "acc_norm_stderr": 0.02489047176993815
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.63,
113
+ "acc_stderr": 0.04852365870939099,
114
+ "acc_norm": 0.63,
115
+ "acc_norm_stderr": 0.04852365870939099
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.32,
119
+ "acc_stderr": 0.04688261722621504,
120
+ "acc_norm": 0.32,
121
+ "acc_norm_stderr": 0.04688261722621504
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.46296296296296297,
125
+ "acc_stderr": 0.04820403072760627,
126
+ "acc_norm": 0.46296296296296297,
127
+ "acc_norm_stderr": 0.04820403072760627
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3694581280788177,
131
+ "acc_stderr": 0.03395970381998574,
132
+ "acc_norm": 0.3694581280788177,
133
+ "acc_norm_stderr": 0.03395970381998574
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.5129032258064516,
137
+ "acc_stderr": 0.02843453315268187,
138
+ "acc_norm": 0.5129032258064516,
139
+ "acc_norm_stderr": 0.02843453315268187
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6752136752136753,
143
+ "acc_stderr": 0.03067902276549883,
144
+ "acc_norm": 0.6752136752136753,
145
+ "acc_norm_stderr": 0.03067902276549883
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.44150943396226416,
149
+ "acc_stderr": 0.030561590426731833,
150
+ "acc_norm": 0.44150943396226416,
151
+ "acc_norm_stderr": 0.030561590426731833
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5363636363636364,
155
+ "acc_stderr": 0.04776449162396197,
156
+ "acc_norm": 0.5363636363636364,
157
+ "acc_norm_stderr": 0.04776449162396197
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.23703703703703705,
161
+ "acc_stderr": 0.025928876132766107,
162
+ "acc_norm": 0.23703703703703705,
163
+ "acc_norm_stderr": 0.025928876132766107
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.31788079470198677,
167
+ "acc_stderr": 0.03802039760107903,
168
+ "acc_norm": 0.31788079470198677,
169
+ "acc_norm_stderr": 0.03802039760107903
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5124378109452736,
173
+ "acc_stderr": 0.0353443984853958,
174
+ "acc_norm": 0.5124378109452736,
175
+ "acc_norm_stderr": 0.0353443984853958
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3815028901734104,
179
+ "acc_stderr": 0.03703851193099521,
180
+ "acc_norm": 0.3815028901734104,
181
+ "acc_norm_stderr": 0.03703851193099521
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.28835978835978837,
185
+ "acc_stderr": 0.0233306540545359,
186
+ "acc_norm": 0.28835978835978837,
187
+ "acc_norm_stderr": 0.0233306540545359
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4166666666666667,
191
+ "acc_stderr": 0.041227287076512825,
192
+ "acc_norm": 0.4166666666666667,
193
+ "acc_norm_stderr": 0.041227287076512825
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.36,
197
+ "acc_stderr": 0.04824181513244218,
198
+ "acc_norm": 0.36,
199
+ "acc_norm_stderr": 0.04824181513244218
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.63,
203
+ "acc_stderr": 0.048523658709391,
204
+ "acc_norm": 0.63,
205
+ "acc_norm_stderr": 0.048523658709391
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.4884393063583815,
209
+ "acc_stderr": 0.026911898686377927,
210
+ "acc_norm": 0.4884393063583815,
211
+ "acc_norm_stderr": 0.026911898686377927
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4539877300613497,
215
+ "acc_stderr": 0.0391170190467718,
216
+ "acc_norm": 0.4539877300613497,
217
+ "acc_norm_stderr": 0.0391170190467718
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.49382716049382713,
221
+ "acc_stderr": 0.027818623962583295,
222
+ "acc_norm": 0.49382716049382713,
223
+ "acc_norm_stderr": 0.027818623962583295
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.34,
227
+ "acc_stderr": 0.04760952285695235,
228
+ "acc_norm": 0.34,
229
+ "acc_norm_stderr": 0.04760952285695235
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5233160621761658,
233
+ "acc_stderr": 0.036045136724422014,
234
+ "acc_norm": 0.5233160621761658,
235
+ "acc_norm_stderr": 0.036045136724422014
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2543859649122807,
239
+ "acc_stderr": 0.04096985139843671,
240
+ "acc_norm": 0.2543859649122807,
241
+ "acc_norm_stderr": 0.04096985139843671
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5651376146788991,
245
+ "acc_stderr": 0.021254631465609273,
246
+ "acc_norm": 0.5651376146788991,
247
+ "acc_norm_stderr": 0.021254631465609273
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.2619047619047619,
251
+ "acc_stderr": 0.03932537680392869,
252
+ "acc_norm": 0.2619047619047619,
253
+ "acc_norm_stderr": 0.03932537680392869
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.42810457516339867,
257
+ "acc_stderr": 0.028332397483664274,
258
+ "acc_norm": 0.42810457516339867,
259
+ "acc_norm_stderr": 0.028332397483664274
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.4,
263
+ "acc_stderr": 0.049236596391733084,
264
+ "acc_norm": 0.4,
265
+ "acc_norm_stderr": 0.049236596391733084
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6446280991735537,
269
+ "acc_stderr": 0.0436923632657398,
270
+ "acc_norm": 0.6446280991735537,
271
+ "acc_norm_stderr": 0.0436923632657398
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40131578947368424,
275
+ "acc_stderr": 0.03988903703336284,
276
+ "acc_norm": 0.40131578947368424,
277
+ "acc_norm_stderr": 0.03988903703336284
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3954248366013072,
281
+ "acc_stderr": 0.019780465954777515,
282
+ "acc_norm": 0.3954248366013072,
283
+ "acc_norm_stderr": 0.019780465954777515
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.30851063829787234,
287
+ "acc_stderr": 0.027553366165101376,
288
+ "acc_norm": 0.30851063829787234,
289
+ "acc_norm_stderr": 0.027553366165101376
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25892857142857145,
293
+ "acc_stderr": 0.04157751539865629,
294
+ "acc_norm": 0.25892857142857145,
295
+ "acc_norm_stderr": 0.04157751539865629
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.33796296296296297,
299
+ "acc_stderr": 0.03225941352631295,
300
+ "acc_norm": 0.33796296296296297,
301
+ "acc_norm_stderr": 0.03225941352631295
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2558659217877095,
305
+ "acc_stderr": 0.01459362092321074,
306
+ "acc_norm": 0.2558659217877095,
307
+ "acc_norm_stderr": 0.01459362092321074
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.42,
311
+ "acc_stderr": 0.049604496374885836,
312
+ "acc_norm": 0.42,
313
+ "acc_norm_stderr": 0.049604496374885836
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.46,
317
+ "acc_stderr": 0.05009082659620332,
318
+ "acc_norm": 0.46,
319
+ "acc_norm_stderr": 0.05009082659620332
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.34191176470588236,
323
+ "acc_stderr": 0.028814722422254167,
324
+ "acc_norm": 0.34191176470588236,
325
+ "acc_norm_stderr": 0.028814722422254167
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.43673469387755104,
329
+ "acc_stderr": 0.03175195237583322,
330
+ "acc_norm": 0.43673469387755104,
331
+ "acc_norm_stderr": 0.03175195237583322
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6286919831223629,
335
+ "acc_stderr": 0.03145068600744858,
336
+ "acc_norm": 0.6286919831223629,
337
+ "acc_norm_stderr": 0.03145068600744858
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.34615384615384615,
341
+ "acc_stderr": 0.012150699768228563,
342
+ "acc_norm": 0.34615384615384615,
343
+ "acc_norm_stderr": 0.012150699768228563
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.5294117647058824,
347
+ "acc_stderr": 0.03503235296367994,
348
+ "acc_norm": 0.5294117647058824,
349
+ "acc_norm_stderr": 0.03503235296367994
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5696969696969697,
353
+ "acc_stderr": 0.03866225962879077,
354
+ "acc_norm": 0.5696969696969697,
355
+ "acc_norm_stderr": 0.03866225962879077
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2460220318237454,
359
+ "mc1_stderr": 0.015077219200662574,
360
+ "mc2": 0.40225459810500935,
361
+ "mc2_stderr": 0.01478258523910622
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.38488783943329397,
365
+ "acc_stderr": 0.016728579701498665,
366
+ "acc_norm": 0.4427390791027155,
367
+ "acc_norm_stderr": 0.017077254131556217
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIFT/aift-llama2-koen-instruct-v1.1-dpo-test1",
436
+ "model_sha": "a416328b862669edfe25be6c305bc9f5ccc4d727",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIFT/aift-llama2-koen-instruct-v1.1/result_2023-12-15 09:57:09.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3822525597269625,
5
+ "acc_stderr": 0.014200454049979279,
6
+ "acc_norm": 0.4402730375426621,
7
+ "acc_norm_stderr": 0.014506769524804243
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.41894045010953995,
11
+ "acc_stderr": 0.004923772581848503,
12
+ "acc_norm": 0.5714997012547302,
13
+ "acc_norm_stderr": 0.004938500303990289
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4678362573099415,
17
+ "acc_stderr": 0.038268824176603704,
18
+ "acc_norm": 0.4678362573099415,
19
+ "acc_norm_stderr": 0.038268824176603704
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5242718446601942,
23
+ "acc_stderr": 0.049449010929737795,
24
+ "acc_norm": 0.5242718446601942,
25
+ "acc_norm_stderr": 0.049449010929737795
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5555555555555556,
29
+ "acc_stderr": 0.01776925058353325,
30
+ "acc_norm": 0.5555555555555556,
31
+ "acc_norm_stderr": 0.01776925058353325
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4740740740740741,
35
+ "acc_stderr": 0.04313531696750574,
36
+ "acc_norm": 0.4740740740740741,
37
+ "acc_norm_stderr": 0.04313531696750574
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.35,
41
+ "acc_stderr": 0.047937248544110196,
42
+ "acc_norm": 0.35,
43
+ "acc_norm_stderr": 0.047937248544110196
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.4127659574468085,
47
+ "acc_stderr": 0.03218471141400351,
48
+ "acc_norm": 0.4127659574468085,
49
+ "acc_norm_stderr": 0.03218471141400351
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.41566265060240964,
53
+ "acc_stderr": 0.03836722176598052,
54
+ "acc_norm": 0.41566265060240964,
55
+ "acc_norm_stderr": 0.03836722176598052
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5241157556270096,
59
+ "acc_stderr": 0.028365041542564584,
60
+ "acc_norm": 0.5241157556270096,
61
+ "acc_norm_stderr": 0.028365041542564584
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.5246636771300448,
65
+ "acc_stderr": 0.03351695167652628,
66
+ "acc_norm": 0.5246636771300448,
67
+ "acc_norm_stderr": 0.03351695167652628
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.5190839694656488,
71
+ "acc_stderr": 0.043820947055509867,
72
+ "acc_norm": 0.5190839694656488,
73
+ "acc_norm_stderr": 0.043820947055509867
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.39,
77
+ "acc_stderr": 0.04902071300001974,
78
+ "acc_norm": 0.39,
79
+ "acc_norm_stderr": 0.04902071300001974
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5707070707070707,
83
+ "acc_stderr": 0.03526552724601198,
84
+ "acc_norm": 0.5707070707070707,
85
+ "acc_norm_stderr": 0.03526552724601198
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.43448275862068964,
89
+ "acc_stderr": 0.041307408795554966,
90
+ "acc_norm": 0.43448275862068964,
91
+ "acc_norm_stderr": 0.041307408795554966
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.20588235294117646,
95
+ "acc_stderr": 0.040233822736177476,
96
+ "acc_norm": 0.20588235294117646,
97
+ "acc_norm_stderr": 0.040233822736177476
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.47058823529411764,
101
+ "acc_stderr": 0.032422250271150074,
102
+ "acc_norm": 0.47058823529411764,
103
+ "acc_norm_stderr": 0.032422250271150074
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4025641025641026,
107
+ "acc_stderr": 0.024864995159767762,
108
+ "acc_norm": 0.4025641025641026,
109
+ "acc_norm_stderr": 0.024864995159767762
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.66,
113
+ "acc_stderr": 0.04760952285695237,
114
+ "acc_norm": 0.66,
115
+ "acc_norm_stderr": 0.04760952285695237
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.31,
119
+ "acc_stderr": 0.04648231987117316,
120
+ "acc_norm": 0.31,
121
+ "acc_norm_stderr": 0.04648231987117316
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.46296296296296297,
125
+ "acc_stderr": 0.04820403072760627,
126
+ "acc_norm": 0.46296296296296297,
127
+ "acc_norm_stderr": 0.04820403072760627
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3694581280788177,
131
+ "acc_stderr": 0.03395970381998574,
132
+ "acc_norm": 0.3694581280788177,
133
+ "acc_norm_stderr": 0.03395970381998574
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.5129032258064516,
137
+ "acc_stderr": 0.02843453315268187,
138
+ "acc_norm": 0.5129032258064516,
139
+ "acc_norm_stderr": 0.02843453315268187
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6709401709401709,
143
+ "acc_stderr": 0.03078232157768817,
144
+ "acc_norm": 0.6709401709401709,
145
+ "acc_norm_stderr": 0.03078232157768817
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4377358490566038,
149
+ "acc_stderr": 0.030533338430467516,
150
+ "acc_norm": 0.4377358490566038,
151
+ "acc_norm_stderr": 0.030533338430467516
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5181818181818182,
155
+ "acc_stderr": 0.04785964010794916,
156
+ "acc_norm": 0.5181818181818182,
157
+ "acc_norm_stderr": 0.04785964010794916
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.24074074074074073,
161
+ "acc_stderr": 0.026067159222275798,
162
+ "acc_norm": 0.24074074074074073,
163
+ "acc_norm_stderr": 0.026067159222275798
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.33112582781456956,
167
+ "acc_stderr": 0.038425817186598696,
168
+ "acc_norm": 0.33112582781456956,
169
+ "acc_norm_stderr": 0.038425817186598696
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5223880597014925,
173
+ "acc_stderr": 0.03531987930208731,
174
+ "acc_norm": 0.5223880597014925,
175
+ "acc_norm_stderr": 0.03531987930208731
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3930635838150289,
179
+ "acc_stderr": 0.037242495958177295,
180
+ "acc_norm": 0.3930635838150289,
181
+ "acc_norm_stderr": 0.037242495958177295
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2962962962962963,
185
+ "acc_stderr": 0.023517294335963286,
186
+ "acc_norm": 0.2962962962962963,
187
+ "acc_norm_stderr": 0.023517294335963286
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4236111111111111,
191
+ "acc_stderr": 0.041321250197233685,
192
+ "acc_norm": 0.4236111111111111,
193
+ "acc_norm_stderr": 0.041321250197233685
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.34,
197
+ "acc_stderr": 0.04760952285695235,
198
+ "acc_norm": 0.34,
199
+ "acc_norm_stderr": 0.04760952285695235
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.63,
203
+ "acc_stderr": 0.048523658709391,
204
+ "acc_norm": 0.63,
205
+ "acc_norm_stderr": 0.048523658709391
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.48554913294797686,
209
+ "acc_stderr": 0.02690784985628254,
210
+ "acc_norm": 0.48554913294797686,
211
+ "acc_norm_stderr": 0.02690784985628254
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4539877300613497,
215
+ "acc_stderr": 0.0391170190467718,
216
+ "acc_norm": 0.4539877300613497,
217
+ "acc_norm_stderr": 0.0391170190467718
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4845679012345679,
221
+ "acc_stderr": 0.0278074900442762,
222
+ "acc_norm": 0.4845679012345679,
223
+ "acc_norm_stderr": 0.0278074900442762
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.34,
227
+ "acc_stderr": 0.04760952285695235,
228
+ "acc_norm": 0.34,
229
+ "acc_norm_stderr": 0.04760952285695235
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5181347150259067,
233
+ "acc_stderr": 0.036060650018329185,
234
+ "acc_norm": 0.5181347150259067,
235
+ "acc_norm_stderr": 0.036060650018329185
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2719298245614035,
239
+ "acc_stderr": 0.041857744240220575,
240
+ "acc_norm": 0.2719298245614035,
241
+ "acc_norm_stderr": 0.041857744240220575
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5614678899082569,
245
+ "acc_stderr": 0.021274713073954562,
246
+ "acc_norm": 0.5614678899082569,
247
+ "acc_norm_stderr": 0.021274713073954562
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.2619047619047619,
251
+ "acc_stderr": 0.0393253768039287,
252
+ "acc_norm": 0.2619047619047619,
253
+ "acc_norm_stderr": 0.0393253768039287
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4215686274509804,
257
+ "acc_stderr": 0.02827549015679143,
258
+ "acc_norm": 0.4215686274509804,
259
+ "acc_norm_stderr": 0.02827549015679143
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.4,
263
+ "acc_stderr": 0.049236596391733084,
264
+ "acc_norm": 0.4,
265
+ "acc_norm_stderr": 0.049236596391733084
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6528925619834711,
269
+ "acc_stderr": 0.04345724570292535,
270
+ "acc_norm": 0.6528925619834711,
271
+ "acc_norm_stderr": 0.04345724570292535
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40131578947368424,
275
+ "acc_stderr": 0.03988903703336284,
276
+ "acc_norm": 0.40131578947368424,
277
+ "acc_norm_stderr": 0.03988903703336284
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.4035947712418301,
281
+ "acc_stderr": 0.01984828016840115,
282
+ "acc_norm": 0.4035947712418301,
283
+ "acc_norm_stderr": 0.01984828016840115
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.32269503546099293,
287
+ "acc_stderr": 0.027889139300534778,
288
+ "acc_norm": 0.32269503546099293,
289
+ "acc_norm_stderr": 0.027889139300534778
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25,
293
+ "acc_stderr": 0.04109974682633932,
294
+ "acc_norm": 0.25,
295
+ "acc_norm_stderr": 0.04109974682633932
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3333333333333333,
299
+ "acc_stderr": 0.03214952147802749,
300
+ "acc_norm": 0.3333333333333333,
301
+ "acc_norm_stderr": 0.03214952147802749
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.25921787709497207,
305
+ "acc_stderr": 0.01465578083749773,
306
+ "acc_norm": 0.25921787709497207,
307
+ "acc_norm_stderr": 0.01465578083749773
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.41,
311
+ "acc_stderr": 0.049431107042371025,
312
+ "acc_norm": 0.41,
313
+ "acc_norm_stderr": 0.049431107042371025
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.46,
317
+ "acc_stderr": 0.05009082659620332,
318
+ "acc_norm": 0.46,
319
+ "acc_norm_stderr": 0.05009082659620332
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.33455882352941174,
323
+ "acc_stderr": 0.028661996202335314,
324
+ "acc_norm": 0.33455882352941174,
325
+ "acc_norm_stderr": 0.028661996202335314
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.43673469387755104,
329
+ "acc_stderr": 0.03175195237583322,
330
+ "acc_norm": 0.43673469387755104,
331
+ "acc_norm_stderr": 0.03175195237583322
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6286919831223629,
335
+ "acc_stderr": 0.03145068600744858,
336
+ "acc_norm": 0.6286919831223629,
337
+ "acc_norm_stderr": 0.03145068600744858
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.34810951760104303,
341
+ "acc_stderr": 0.0121667389936982,
342
+ "acc_norm": 0.34810951760104303,
343
+ "acc_norm_stderr": 0.0121667389936982
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.5294117647058824,
347
+ "acc_stderr": 0.03503235296367994,
348
+ "acc_norm": 0.5294117647058824,
349
+ "acc_norm_stderr": 0.03503235296367994
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5696969696969697,
353
+ "acc_stderr": 0.038662259628790774,
354
+ "acc_norm": 0.5696969696969697,
355
+ "acc_norm_stderr": 0.038662259628790774
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.24724602203182375,
359
+ "mc1_stderr": 0.015102404797359649,
360
+ "mc2": 0.4028203603366851,
361
+ "mc2_stderr": 0.01477432836961688
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.38961038961038963,
365
+ "acc_stderr": 0.0167661616718935,
366
+ "acc_norm": 0.45218417945690675,
367
+ "acc_norm_stderr": 0.017111567130916782
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIFT/aift-llama2-koen-instruct-v1.1",
436
+ "model_sha": "202c3e3df0c4a321503df8d4c78da213f1ae5475",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIFT/aift-llama2-koen-instruct-v1.2/result_2023-12-16 03:29:50.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.38310580204778155,
5
+ "acc_stderr": 0.01420647266167288,
6
+ "acc_norm": 0.4453924914675768,
7
+ "acc_norm_stderr": 0.014523987638344074
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.42113124875522806,
11
+ "acc_stderr": 0.004927314729433555,
12
+ "acc_norm": 0.578370842461661,
13
+ "acc_norm_stderr": 0.004928105880776078
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4619883040935672,
17
+ "acc_stderr": 0.038237270928823064,
18
+ "acc_norm": 0.4619883040935672,
19
+ "acc_norm_stderr": 0.038237270928823064
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.49514563106796117,
23
+ "acc_stderr": 0.04950504382128921,
24
+ "acc_norm": 0.49514563106796117,
25
+ "acc_norm_stderr": 0.04950504382128921
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5644955300127714,
29
+ "acc_stderr": 0.017730589927926588,
30
+ "acc_norm": 0.5644955300127714,
31
+ "acc_norm_stderr": 0.017730589927926588
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4666666666666667,
35
+ "acc_stderr": 0.043097329010363554,
36
+ "acc_norm": 0.4666666666666667,
37
+ "acc_norm_stderr": 0.043097329010363554
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.3,
41
+ "acc_stderr": 0.046056618647183814,
42
+ "acc_norm": 0.3,
43
+ "acc_norm_stderr": 0.046056618647183814
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.4127659574468085,
47
+ "acc_stderr": 0.03218471141400352,
48
+ "acc_norm": 0.4127659574468085,
49
+ "acc_norm_stderr": 0.03218471141400352
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.41566265060240964,
53
+ "acc_stderr": 0.03836722176598052,
54
+ "acc_norm": 0.41566265060240964,
55
+ "acc_norm_stderr": 0.03836722176598052
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.5144694533762058,
59
+ "acc_stderr": 0.028386198084177687,
60
+ "acc_norm": 0.5144694533762058,
61
+ "acc_norm_stderr": 0.028386198084177687
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.5112107623318386,
65
+ "acc_stderr": 0.033549366530984746,
66
+ "acc_norm": 0.5112107623318386,
67
+ "acc_norm_stderr": 0.033549366530984746
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.48854961832061067,
71
+ "acc_stderr": 0.04384140024078016,
72
+ "acc_norm": 0.48854961832061067,
73
+ "acc_norm_stderr": 0.04384140024078016
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.35,
77
+ "acc_stderr": 0.0479372485441102,
78
+ "acc_norm": 0.35,
79
+ "acc_norm_stderr": 0.0479372485441102
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5858585858585859,
83
+ "acc_stderr": 0.03509438348879629,
84
+ "acc_norm": 0.5858585858585859,
85
+ "acc_norm_stderr": 0.03509438348879629
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4206896551724138,
89
+ "acc_stderr": 0.0411391498118926,
90
+ "acc_norm": 0.4206896551724138,
91
+ "acc_norm_stderr": 0.0411391498118926
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.21568627450980393,
95
+ "acc_stderr": 0.04092563958237655,
96
+ "acc_norm": 0.21568627450980393,
97
+ "acc_norm_stderr": 0.04092563958237655
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.46638655462184875,
101
+ "acc_stderr": 0.03240501447690071,
102
+ "acc_norm": 0.46638655462184875,
103
+ "acc_norm_stderr": 0.03240501447690071
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.3974358974358974,
107
+ "acc_stderr": 0.024811920017903836,
108
+ "acc_norm": 0.3974358974358974,
109
+ "acc_norm_stderr": 0.024811920017903836
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.62,
113
+ "acc_stderr": 0.04878317312145632,
114
+ "acc_norm": 0.62,
115
+ "acc_norm_stderr": 0.04878317312145632
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.32,
119
+ "acc_stderr": 0.046882617226215034,
120
+ "acc_norm": 0.32,
121
+ "acc_norm_stderr": 0.046882617226215034
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.46296296296296297,
125
+ "acc_stderr": 0.04820403072760627,
126
+ "acc_norm": 0.46296296296296297,
127
+ "acc_norm_stderr": 0.04820403072760627
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3645320197044335,
131
+ "acc_stderr": 0.0338640574606209,
132
+ "acc_norm": 0.3645320197044335,
133
+ "acc_norm_stderr": 0.0338640574606209
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4935483870967742,
137
+ "acc_stderr": 0.02844163823354051,
138
+ "acc_norm": 0.4935483870967742,
139
+ "acc_norm_stderr": 0.02844163823354051
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6538461538461539,
143
+ "acc_stderr": 0.0311669573672359,
144
+ "acc_norm": 0.6538461538461539,
145
+ "acc_norm_stderr": 0.0311669573672359
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4226415094339623,
149
+ "acc_stderr": 0.03040233144576954,
150
+ "acc_norm": 0.4226415094339623,
151
+ "acc_norm_stderr": 0.03040233144576954
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5545454545454546,
155
+ "acc_stderr": 0.047605488214603246,
156
+ "acc_norm": 0.5545454545454546,
157
+ "acc_norm_stderr": 0.047605488214603246
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.2074074074074074,
161
+ "acc_stderr": 0.024720713193952148,
162
+ "acc_norm": 0.2074074074074074,
163
+ "acc_norm_stderr": 0.024720713193952148
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.33774834437086093,
167
+ "acc_stderr": 0.03861557546255169,
168
+ "acc_norm": 0.33774834437086093,
169
+ "acc_norm_stderr": 0.03861557546255169
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5124378109452736,
173
+ "acc_stderr": 0.0353443984853958,
174
+ "acc_norm": 0.5124378109452736,
175
+ "acc_norm_stderr": 0.0353443984853958
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3872832369942196,
179
+ "acc_stderr": 0.03714325906302065,
180
+ "acc_norm": 0.3872832369942196,
181
+ "acc_norm_stderr": 0.03714325906302065
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2962962962962963,
185
+ "acc_stderr": 0.023517294335963286,
186
+ "acc_norm": 0.2962962962962963,
187
+ "acc_norm_stderr": 0.023517294335963286
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4166666666666667,
191
+ "acc_stderr": 0.041227287076512825,
192
+ "acc_norm": 0.4166666666666667,
193
+ "acc_norm_stderr": 0.041227287076512825
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.36,
197
+ "acc_stderr": 0.04824181513244218,
198
+ "acc_norm": 0.36,
199
+ "acc_norm_stderr": 0.04824181513244218
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.6,
203
+ "acc_stderr": 0.04923659639173309,
204
+ "acc_norm": 0.6,
205
+ "acc_norm_stderr": 0.04923659639173309
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.49710982658959535,
209
+ "acc_stderr": 0.02691864538323901,
210
+ "acc_norm": 0.49710982658959535,
211
+ "acc_norm_stderr": 0.02691864538323901
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4233128834355828,
215
+ "acc_stderr": 0.038818912133343826,
216
+ "acc_norm": 0.4233128834355828,
217
+ "acc_norm_stderr": 0.038818912133343826
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.49691358024691357,
221
+ "acc_stderr": 0.027820214158594377,
222
+ "acc_norm": 0.49691358024691357,
223
+ "acc_norm_stderr": 0.027820214158594377
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.34,
227
+ "acc_stderr": 0.04760952285695236,
228
+ "acc_norm": 0.34,
229
+ "acc_norm_stderr": 0.04760952285695236
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5181347150259067,
233
+ "acc_stderr": 0.036060650018329185,
234
+ "acc_norm": 0.5181347150259067,
235
+ "acc_norm_stderr": 0.036060650018329185
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2631578947368421,
239
+ "acc_stderr": 0.0414243971948936,
240
+ "acc_norm": 0.2631578947368421,
241
+ "acc_norm_stderr": 0.0414243971948936
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5688073394495413,
245
+ "acc_stderr": 0.02123336503031956,
246
+ "acc_norm": 0.5688073394495413,
247
+ "acc_norm_stderr": 0.02123336503031956
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.2857142857142857,
251
+ "acc_stderr": 0.0404061017820884,
252
+ "acc_norm": 0.2857142857142857,
253
+ "acc_norm_stderr": 0.0404061017820884
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.434640522875817,
257
+ "acc_stderr": 0.028384256704883037,
258
+ "acc_norm": 0.434640522875817,
259
+ "acc_norm_stderr": 0.028384256704883037
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.4,
263
+ "acc_stderr": 0.049236596391733084,
264
+ "acc_norm": 0.4,
265
+ "acc_norm_stderr": 0.049236596391733084
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6528925619834711,
269
+ "acc_stderr": 0.04345724570292535,
270
+ "acc_norm": 0.6528925619834711,
271
+ "acc_norm_stderr": 0.04345724570292535
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.39473684210526316,
275
+ "acc_stderr": 0.039777499346220734,
276
+ "acc_norm": 0.39473684210526316,
277
+ "acc_norm_stderr": 0.039777499346220734
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.380718954248366,
281
+ "acc_stderr": 0.019643801557924806,
282
+ "acc_norm": 0.380718954248366,
283
+ "acc_norm_stderr": 0.019643801557924806
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.30141843971631205,
287
+ "acc_stderr": 0.02737412888263115,
288
+ "acc_norm": 0.30141843971631205,
289
+ "acc_norm_stderr": 0.02737412888263115
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.26785714285714285,
293
+ "acc_stderr": 0.04203277291467762,
294
+ "acc_norm": 0.26785714285714285,
295
+ "acc_norm_stderr": 0.04203277291467762
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3425925925925926,
299
+ "acc_stderr": 0.032365852526021574,
300
+ "acc_norm": 0.3425925925925926,
301
+ "acc_norm_stderr": 0.032365852526021574
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2681564245810056,
305
+ "acc_stderr": 0.014816119635317,
306
+ "acc_norm": 0.2681564245810056,
307
+ "acc_norm_stderr": 0.014816119635317
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.39,
311
+ "acc_stderr": 0.04902071300001974,
312
+ "acc_norm": 0.39,
313
+ "acc_norm_stderr": 0.04902071300001974
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.46,
317
+ "acc_stderr": 0.05009082659620332,
318
+ "acc_norm": 0.46,
319
+ "acc_norm_stderr": 0.05009082659620332
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3272058823529412,
323
+ "acc_stderr": 0.028501452860396587,
324
+ "acc_norm": 0.3272058823529412,
325
+ "acc_norm_stderr": 0.028501452860396587
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4326530612244898,
329
+ "acc_stderr": 0.03171752824062664,
330
+ "acc_norm": 0.4326530612244898,
331
+ "acc_norm_stderr": 0.03171752824062664
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6286919831223629,
335
+ "acc_stderr": 0.03145068600744859,
336
+ "acc_norm": 0.6286919831223629,
337
+ "acc_norm_stderr": 0.03145068600744859
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.35658409387222945,
341
+ "acc_stderr": 0.012233642989273886,
342
+ "acc_norm": 0.35658409387222945,
343
+ "acc_norm_stderr": 0.012233642989273886
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.553921568627451,
347
+ "acc_stderr": 0.034888454513049734,
348
+ "acc_norm": 0.553921568627451,
349
+ "acc_norm_stderr": 0.034888454513049734
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5757575757575758,
353
+ "acc_stderr": 0.038592681420702636,
354
+ "acc_norm": 0.5757575757575758,
355
+ "acc_norm_stderr": 0.038592681420702636
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2521419828641371,
359
+ "mc1_stderr": 0.015201522246299946,
360
+ "mc2": 0.4052899642454083,
361
+ "mc2_stderr": 0.014924042516908636
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.38488783943329397,
365
+ "acc_stderr": 0.016728579701498672,
366
+ "acc_norm": 0.4510035419126328,
367
+ "acc_norm_stderr": 0.017107618859549357
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIFT/aift-llama2-koen-instruct-v1.2",
436
+ "model_sha": "95f3e7cce5bebe90ac4ff8f07597be444e7e1a9e",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIdenU/LLAMA-2-13b-ko-Y24-DPO_v0.1/result_2023-12-18 01:34:17.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3191126279863481,
5
+ "acc_stderr": 0.013621696119173297,
6
+ "acc_norm": 0.3779863481228669,
7
+ "acc_norm_stderr": 0.014169664520303103
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.36446922923720376,
11
+ "acc_stderr": 0.004802974070507201,
12
+ "acc_norm": 0.46883091017725553,
13
+ "acc_norm_stderr": 0.004980076707392429
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.42105263157894735,
17
+ "acc_stderr": 0.03786720706234215,
18
+ "acc_norm": 0.42105263157894735,
19
+ "acc_norm_stderr": 0.03786720706234215
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.4854368932038835,
23
+ "acc_stderr": 0.04948637324026637,
24
+ "acc_norm": 0.4854368932038835,
25
+ "acc_norm_stderr": 0.04948637324026637
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4891443167305236,
29
+ "acc_stderr": 0.01787574884024242,
30
+ "acc_norm": 0.4891443167305236,
31
+ "acc_norm_stderr": 0.01787574884024242
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.37037037037037035,
35
+ "acc_stderr": 0.04171654161354543,
36
+ "acc_norm": 0.37037037037037035,
37
+ "acc_norm_stderr": 0.04171654161354543
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.37,
41
+ "acc_stderr": 0.048523658709391,
42
+ "acc_norm": 0.37,
43
+ "acc_norm_stderr": 0.048523658709391
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.39148936170212767,
47
+ "acc_stderr": 0.03190701242326812,
48
+ "acc_norm": 0.39148936170212767,
49
+ "acc_norm_stderr": 0.03190701242326812
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.35542168674698793,
53
+ "acc_stderr": 0.03726214354322415,
54
+ "acc_norm": 0.35542168674698793,
55
+ "acc_norm_stderr": 0.03726214354322415
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4662379421221865,
59
+ "acc_stderr": 0.02833327710956278,
60
+ "acc_norm": 0.4662379421221865,
61
+ "acc_norm_stderr": 0.02833327710956278
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.3811659192825112,
65
+ "acc_stderr": 0.03259625118416828,
66
+ "acc_norm": 0.3811659192825112,
67
+ "acc_norm_stderr": 0.03259625118416828
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.5419847328244275,
71
+ "acc_stderr": 0.04369802690578756,
72
+ "acc_norm": 0.5419847328244275,
73
+ "acc_norm_stderr": 0.04369802690578756
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.31,
77
+ "acc_stderr": 0.046482319871173156,
78
+ "acc_norm": 0.31,
79
+ "acc_norm_stderr": 0.046482319871173156
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5151515151515151,
83
+ "acc_stderr": 0.03560716516531061,
84
+ "acc_norm": 0.5151515151515151,
85
+ "acc_norm_stderr": 0.03560716516531061
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4206896551724138,
89
+ "acc_stderr": 0.0411391498118926,
90
+ "acc_norm": 0.4206896551724138,
91
+ "acc_norm_stderr": 0.0411391498118926
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.22549019607843138,
95
+ "acc_stderr": 0.041583075330832865,
96
+ "acc_norm": 0.22549019607843138,
97
+ "acc_norm_stderr": 0.041583075330832865
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.4327731092436975,
101
+ "acc_stderr": 0.03218358107742613,
102
+ "acc_norm": 0.4327731092436975,
103
+ "acc_norm_stderr": 0.03218358107742613
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.3974358974358974,
107
+ "acc_stderr": 0.024811920017903836,
108
+ "acc_norm": 0.3974358974358974,
109
+ "acc_norm_stderr": 0.024811920017903836
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.47,
113
+ "acc_stderr": 0.050161355804659205,
114
+ "acc_norm": 0.47,
115
+ "acc_norm_stderr": 0.050161355804659205
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.32,
119
+ "acc_stderr": 0.04688261722621504,
120
+ "acc_norm": 0.32,
121
+ "acc_norm_stderr": 0.04688261722621504
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.49074074074074076,
125
+ "acc_stderr": 0.04832853553437055,
126
+ "acc_norm": 0.49074074074074076,
127
+ "acc_norm_stderr": 0.04832853553437055
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3694581280788177,
131
+ "acc_stderr": 0.033959703819985754,
132
+ "acc_norm": 0.3694581280788177,
133
+ "acc_norm_stderr": 0.033959703819985754
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4645161290322581,
137
+ "acc_stderr": 0.02837228779796295,
138
+ "acc_norm": 0.4645161290322581,
139
+ "acc_norm_stderr": 0.02837228779796295
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6111111111111112,
143
+ "acc_stderr": 0.031937057262002924,
144
+ "acc_norm": 0.6111111111111112,
145
+ "acc_norm_stderr": 0.031937057262002924
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4075471698113208,
149
+ "acc_stderr": 0.030242233800854498,
150
+ "acc_norm": 0.4075471698113208,
151
+ "acc_norm_stderr": 0.030242233800854498
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.509090909090909,
155
+ "acc_stderr": 0.04788339768702861,
156
+ "acc_norm": 0.509090909090909,
157
+ "acc_norm_stderr": 0.04788339768702861
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.23703703703703705,
161
+ "acc_stderr": 0.02592887613276612,
162
+ "acc_norm": 0.23703703703703705,
163
+ "acc_norm_stderr": 0.02592887613276612
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.31788079470198677,
167
+ "acc_stderr": 0.03802039760107903,
168
+ "acc_norm": 0.31788079470198677,
169
+ "acc_norm_stderr": 0.03802039760107903
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5771144278606966,
173
+ "acc_stderr": 0.034932317774212816,
174
+ "acc_norm": 0.5771144278606966,
175
+ "acc_norm_stderr": 0.034932317774212816
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3583815028901734,
179
+ "acc_stderr": 0.036563436533531585,
180
+ "acc_norm": 0.3583815028901734,
181
+ "acc_norm_stderr": 0.036563436533531585
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.25925925925925924,
185
+ "acc_stderr": 0.02256989707491841,
186
+ "acc_norm": 0.25925925925925924,
187
+ "acc_norm_stderr": 0.02256989707491841
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3402777777777778,
191
+ "acc_stderr": 0.03962135573486219,
192
+ "acc_norm": 0.3402777777777778,
193
+ "acc_norm_stderr": 0.03962135573486219
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.35,
197
+ "acc_stderr": 0.0479372485441102,
198
+ "acc_norm": 0.35,
199
+ "acc_norm_stderr": 0.0479372485441102
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.54,
203
+ "acc_stderr": 0.05009082659620333,
204
+ "acc_norm": 0.54,
205
+ "acc_norm_stderr": 0.05009082659620333
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.47109826589595377,
209
+ "acc_stderr": 0.026874085883518348,
210
+ "acc_norm": 0.47109826589595377,
211
+ "acc_norm_stderr": 0.026874085883518348
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.3619631901840491,
215
+ "acc_stderr": 0.037757007291414416,
216
+ "acc_norm": 0.3619631901840491,
217
+ "acc_norm_stderr": 0.037757007291414416
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4012345679012346,
221
+ "acc_stderr": 0.0272725828498398,
222
+ "acc_norm": 0.4012345679012346,
223
+ "acc_norm_stderr": 0.0272725828498398
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.27,
227
+ "acc_stderr": 0.044619604333847394,
228
+ "acc_norm": 0.27,
229
+ "acc_norm_stderr": 0.044619604333847394
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.48186528497409326,
233
+ "acc_stderr": 0.036060650018329185,
234
+ "acc_norm": 0.48186528497409326,
235
+ "acc_norm_stderr": 0.036060650018329185
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.3684210526315789,
239
+ "acc_stderr": 0.0453781535493939,
240
+ "acc_norm": 0.3684210526315789,
241
+ "acc_norm_stderr": 0.0453781535493939
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.43302752293577984,
245
+ "acc_stderr": 0.021244146569074345,
246
+ "acc_norm": 0.43302752293577984,
247
+ "acc_norm_stderr": 0.021244146569074345
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.31746031746031744,
251
+ "acc_stderr": 0.04163453031302859,
252
+ "acc_norm": 0.31746031746031744,
253
+ "acc_norm_stderr": 0.04163453031302859
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.43137254901960786,
257
+ "acc_stderr": 0.028358956313423556,
258
+ "acc_norm": 0.43137254901960786,
259
+ "acc_norm_stderr": 0.028358956313423556
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.38,
263
+ "acc_stderr": 0.04878317312145633,
264
+ "acc_norm": 0.38,
265
+ "acc_norm_stderr": 0.04878317312145633
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6033057851239669,
269
+ "acc_stderr": 0.04465869780531009,
270
+ "acc_norm": 0.6033057851239669,
271
+ "acc_norm_stderr": 0.04465869780531009
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4868421052631579,
275
+ "acc_stderr": 0.04067533136309173,
276
+ "acc_norm": 0.4868421052631579,
277
+ "acc_norm_stderr": 0.04067533136309173
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3202614379084967,
281
+ "acc_stderr": 0.018875682938069436,
282
+ "acc_norm": 0.3202614379084967,
283
+ "acc_norm_stderr": 0.018875682938069436
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3049645390070922,
287
+ "acc_stderr": 0.02746470844202212,
288
+ "acc_norm": 0.3049645390070922,
289
+ "acc_norm_stderr": 0.02746470844202212
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.23214285714285715,
293
+ "acc_stderr": 0.04007341809755806,
294
+ "acc_norm": 0.23214285714285715,
295
+ "acc_norm_stderr": 0.04007341809755806
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3101851851851852,
299
+ "acc_stderr": 0.03154696285656628,
300
+ "acc_norm": 0.3101851851851852,
301
+ "acc_norm_stderr": 0.03154696285656628
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2424581005586592,
305
+ "acc_stderr": 0.01433352205921789,
306
+ "acc_norm": 0.2424581005586592,
307
+ "acc_norm_stderr": 0.01433352205921789
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.37,
311
+ "acc_stderr": 0.04852365870939099,
312
+ "acc_norm": 0.37,
313
+ "acc_norm_stderr": 0.04852365870939099
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.39,
317
+ "acc_stderr": 0.04902071300001975,
318
+ "acc_norm": 0.39,
319
+ "acc_norm_stderr": 0.04902071300001975
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.33455882352941174,
323
+ "acc_stderr": 0.028661996202335317,
324
+ "acc_norm": 0.33455882352941174,
325
+ "acc_norm_stderr": 0.028661996202335317
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.43673469387755104,
329
+ "acc_stderr": 0.031751952375833226,
330
+ "acc_norm": 0.43673469387755104,
331
+ "acc_norm_stderr": 0.031751952375833226
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.47257383966244726,
335
+ "acc_stderr": 0.032498227183013026,
336
+ "acc_norm": 0.47257383966244726,
337
+ "acc_norm_stderr": 0.032498227183013026
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.30638852672750977,
341
+ "acc_stderr": 0.011773980329380731,
342
+ "acc_norm": 0.30638852672750977,
343
+ "acc_norm_stderr": 0.011773980329380731
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.35784313725490197,
347
+ "acc_stderr": 0.033644872860882996,
348
+ "acc_norm": 0.35784313725490197,
349
+ "acc_norm_stderr": 0.033644872860882996
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.44242424242424244,
353
+ "acc_stderr": 0.03878372113711274,
354
+ "acc_norm": 0.44242424242424244,
355
+ "acc_norm_stderr": 0.03878372113711274
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2460220318237454,
359
+ "mc1_stderr": 0.015077219200662568,
360
+ "mc2": 0.4153514851890886,
361
+ "mc2_stderr": 0.01500188114852866
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.42266824085005905,
365
+ "acc_stderr": 0.0169835060795776,
366
+ "acc_norm": 0.5053128689492326,
367
+ "acc_norm_stderr": 0.01718938362722971
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIdenU/LLAMA-2-13b-ko-Y24-DPO_v0.1",
436
+ "model_sha": "10c4f59aa0a45a331f9a3288f05daa29d9dc79df",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIdenU/Mistral-7b-ko-Y24-DPO_v0.1/result_2023-12-21 04:18:43.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.35665529010238906,
5
+ "acc_stderr": 0.013998056902620196,
6
+ "acc_norm": 0.4052901023890785,
7
+ "acc_norm_stderr": 0.014346869060229328
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.377912766381199,
11
+ "acc_stderr": 0.004838747305783345,
12
+ "acc_norm": 0.49153555068711413,
13
+ "acc_norm_stderr": 0.004989066355449555
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.43859649122807015,
17
+ "acc_stderr": 0.038057975055904594,
18
+ "acc_norm": 0.43859649122807015,
19
+ "acc_norm_stderr": 0.038057975055904594
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5631067961165048,
23
+ "acc_stderr": 0.04911147107365777,
24
+ "acc_norm": 0.5631067961165048,
25
+ "acc_norm_stderr": 0.04911147107365777
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5006385696040868,
29
+ "acc_stderr": 0.017879948914431697,
30
+ "acc_norm": 0.5006385696040868,
31
+ "acc_norm_stderr": 0.017879948914431697
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.31851851851851853,
35
+ "acc_stderr": 0.0402477840197711,
36
+ "acc_norm": 0.31851851851851853,
37
+ "acc_norm_stderr": 0.0402477840197711
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.26,
41
+ "acc_stderr": 0.04408440022768077,
42
+ "acc_norm": 0.26,
43
+ "acc_norm_stderr": 0.04408440022768077
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.43829787234042555,
47
+ "acc_stderr": 0.03243618636108102,
48
+ "acc_norm": 0.43829787234042555,
49
+ "acc_norm_stderr": 0.03243618636108102
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.43373493975903615,
53
+ "acc_stderr": 0.03858158940685515,
54
+ "acc_norm": 0.43373493975903615,
55
+ "acc_norm_stderr": 0.03858158940685515
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4758842443729904,
59
+ "acc_stderr": 0.02836504154256457,
60
+ "acc_norm": 0.4758842443729904,
61
+ "acc_norm_stderr": 0.02836504154256457
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.5112107623318386,
65
+ "acc_stderr": 0.033549366530984746,
66
+ "acc_norm": 0.5112107623318386,
67
+ "acc_norm_stderr": 0.033549366530984746
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.5038167938931297,
71
+ "acc_stderr": 0.043851623256015534,
72
+ "acc_norm": 0.5038167938931297,
73
+ "acc_norm_stderr": 0.043851623256015534
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.33,
77
+ "acc_stderr": 0.047258156262526045,
78
+ "acc_norm": 0.33,
79
+ "acc_norm_stderr": 0.047258156262526045
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5606060606060606,
83
+ "acc_stderr": 0.035360859475294805,
84
+ "acc_norm": 0.5606060606060606,
85
+ "acc_norm_stderr": 0.035360859475294805
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4482758620689655,
89
+ "acc_stderr": 0.04144311810878151,
90
+ "acc_norm": 0.4482758620689655,
91
+ "acc_norm_stderr": 0.04144311810878151
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.22549019607843138,
95
+ "acc_stderr": 0.041583075330832865,
96
+ "acc_norm": 0.22549019607843138,
97
+ "acc_norm_stderr": 0.041583075330832865
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.47478991596638653,
101
+ "acc_stderr": 0.0324371805513741,
102
+ "acc_norm": 0.47478991596638653,
103
+ "acc_norm_stderr": 0.0324371805513741
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4307692307692308,
107
+ "acc_stderr": 0.025106820660539743,
108
+ "acc_norm": 0.4307692307692308,
109
+ "acc_norm_stderr": 0.025106820660539743
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.58,
113
+ "acc_stderr": 0.04960449637488583,
114
+ "acc_norm": 0.58,
115
+ "acc_norm_stderr": 0.04960449637488583
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.35,
119
+ "acc_stderr": 0.04793724854411019,
120
+ "acc_norm": 0.35,
121
+ "acc_norm_stderr": 0.04793724854411019
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5277777777777778,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.5277777777777778,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.41379310344827586,
131
+ "acc_stderr": 0.03465304488406796,
132
+ "acc_norm": 0.41379310344827586,
133
+ "acc_norm_stderr": 0.03465304488406796
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.47096774193548385,
137
+ "acc_stderr": 0.028396016402761008,
138
+ "acc_norm": 0.47096774193548385,
139
+ "acc_norm_stderr": 0.028396016402761008
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.7478632478632479,
143
+ "acc_stderr": 0.028447965476231022,
144
+ "acc_norm": 0.7478632478632479,
145
+ "acc_norm_stderr": 0.028447965476231022
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.43018867924528303,
149
+ "acc_stderr": 0.030471445867183235,
150
+ "acc_norm": 0.43018867924528303,
151
+ "acc_norm_stderr": 0.030471445867183235
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5272727272727272,
155
+ "acc_stderr": 0.04782001791380061,
156
+ "acc_norm": 0.5272727272727272,
157
+ "acc_norm_stderr": 0.04782001791380061
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.3,
161
+ "acc_stderr": 0.02794045713622841,
162
+ "acc_norm": 0.3,
163
+ "acc_norm_stderr": 0.02794045713622841
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2781456953642384,
167
+ "acc_stderr": 0.03658603262763743,
168
+ "acc_norm": 0.2781456953642384,
169
+ "acc_norm_stderr": 0.03658603262763743
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.6069651741293532,
173
+ "acc_stderr": 0.0345368246603156,
174
+ "acc_norm": 0.6069651741293532,
175
+ "acc_norm_stderr": 0.0345368246603156
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.32947976878612717,
179
+ "acc_stderr": 0.03583901754736411,
180
+ "acc_norm": 0.32947976878612717,
181
+ "acc_norm_stderr": 0.03583901754736411
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.37037037037037035,
185
+ "acc_stderr": 0.024870815251057093,
186
+ "acc_norm": 0.37037037037037035,
187
+ "acc_norm_stderr": 0.024870815251057093
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3263888888888889,
191
+ "acc_stderr": 0.03921067198982266,
192
+ "acc_norm": 0.3263888888888889,
193
+ "acc_norm_stderr": 0.03921067198982266
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.31,
197
+ "acc_stderr": 0.04648231987117316,
198
+ "acc_norm": 0.31,
199
+ "acc_norm_stderr": 0.04648231987117316
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.63,
203
+ "acc_stderr": 0.048523658709391,
204
+ "acc_norm": 0.63,
205
+ "acc_norm_stderr": 0.048523658709391
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.4682080924855491,
209
+ "acc_stderr": 0.02686462436675664,
210
+ "acc_norm": 0.4682080924855491,
211
+ "acc_norm_stderr": 0.02686462436675664
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4662576687116564,
215
+ "acc_stderr": 0.03919415545048409,
216
+ "acc_norm": 0.4662576687116564,
217
+ "acc_norm_stderr": 0.03919415545048409
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4567901234567901,
221
+ "acc_stderr": 0.027716661650194038,
222
+ "acc_norm": 0.4567901234567901,
223
+ "acc_norm_stderr": 0.027716661650194038
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.32,
227
+ "acc_stderr": 0.04688261722621504,
228
+ "acc_norm": 0.32,
229
+ "acc_norm_stderr": 0.04688261722621504
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.47668393782383417,
233
+ "acc_stderr": 0.03604513672442205,
234
+ "acc_norm": 0.47668393782383417,
235
+ "acc_norm_stderr": 0.03604513672442205
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.3157894736842105,
239
+ "acc_stderr": 0.043727482902780085,
240
+ "acc_norm": 0.3157894736842105,
241
+ "acc_norm_stderr": 0.043727482902780085
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5137614678899083,
245
+ "acc_stderr": 0.021429202089874075,
246
+ "acc_norm": 0.5137614678899083,
247
+ "acc_norm_stderr": 0.021429202089874075
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3333333333333333,
251
+ "acc_stderr": 0.042163702135578345,
252
+ "acc_norm": 0.3333333333333333,
253
+ "acc_norm_stderr": 0.042163702135578345
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.434640522875817,
257
+ "acc_stderr": 0.028384256704883037,
258
+ "acc_norm": 0.434640522875817,
259
+ "acc_norm_stderr": 0.028384256704883037
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.44,
263
+ "acc_stderr": 0.04988876515698589,
264
+ "acc_norm": 0.44,
265
+ "acc_norm_stderr": 0.04988876515698589
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6776859504132231,
269
+ "acc_stderr": 0.04266416363352167,
270
+ "acc_norm": 0.6776859504132231,
271
+ "acc_norm_stderr": 0.04266416363352167
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4407894736842105,
275
+ "acc_stderr": 0.040403110624904356,
276
+ "acc_norm": 0.4407894736842105,
277
+ "acc_norm_stderr": 0.040403110624904356
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3872549019607843,
281
+ "acc_stderr": 0.01970687580408563,
282
+ "acc_norm": 0.3872549019607843,
283
+ "acc_norm_stderr": 0.01970687580408563
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.35106382978723405,
287
+ "acc_stderr": 0.028473501272963764,
288
+ "acc_norm": 0.35106382978723405,
289
+ "acc_norm_stderr": 0.028473501272963764
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.375,
293
+ "acc_stderr": 0.04595091388086298,
294
+ "acc_norm": 0.375,
295
+ "acc_norm_stderr": 0.04595091388086298
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.38425925925925924,
299
+ "acc_stderr": 0.03317354514310742,
300
+ "acc_norm": 0.38425925925925924,
301
+ "acc_norm_stderr": 0.03317354514310742
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2581005586592179,
305
+ "acc_stderr": 0.01463518561652784,
306
+ "acc_norm": 0.2581005586592179,
307
+ "acc_norm_stderr": 0.01463518561652784
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.35,
311
+ "acc_stderr": 0.0479372485441102,
312
+ "acc_norm": 0.35,
313
+ "acc_norm_stderr": 0.0479372485441102
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.56,
317
+ "acc_stderr": 0.04988876515698589,
318
+ "acc_norm": 0.56,
319
+ "acc_norm_stderr": 0.04988876515698589
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3235294117647059,
323
+ "acc_stderr": 0.028418208619406787,
324
+ "acc_norm": 0.3235294117647059,
325
+ "acc_norm_stderr": 0.028418208619406787
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4897959183673469,
329
+ "acc_stderr": 0.03200255347893782,
330
+ "acc_norm": 0.4897959183673469,
331
+ "acc_norm_stderr": 0.03200255347893782
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.569620253164557,
335
+ "acc_stderr": 0.032230171959376,
336
+ "acc_norm": 0.569620253164557,
337
+ "acc_norm_stderr": 0.032230171959376
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.318122555410691,
341
+ "acc_stderr": 0.011895407281104104,
342
+ "acc_norm": 0.318122555410691,
343
+ "acc_norm_stderr": 0.011895407281104104
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4019607843137255,
347
+ "acc_stderr": 0.03441190023482465,
348
+ "acc_norm": 0.4019607843137255,
349
+ "acc_norm_stderr": 0.03441190023482465
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.45454545454545453,
353
+ "acc_stderr": 0.03888176921674098,
354
+ "acc_norm": 0.45454545454545453,
355
+ "acc_norm_stderr": 0.03888176921674098
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.26193390452876375,
359
+ "mc1_stderr": 0.015392118805015021,
360
+ "mc2": 0.4379686054133816,
361
+ "mc2_stderr": 0.015396278996687385
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.4923258559622196,
365
+ "acc_stderr": 0.017188329219654276,
366
+ "acc_norm": 0.5419126328217237,
367
+ "acc_norm_stderr": 0.017129852117911147
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIdenU/Mistral-7b-ko-Y24-DPO_v0.1",
436
+ "model_sha": "78813fb52898d37d6c0637b7fb93eb2c5bc23f55",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AIdenU/Mistral-7b-ko-Y24_v0.1/result_2023-12-21 04:19:12.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3506825938566553,
5
+ "acc_stderr": 0.01394463593072609,
6
+ "acc_norm": 0.39590443686006827,
7
+ "acc_norm_stderr": 0.014291228393536588
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.37870942043417644,
11
+ "acc_stderr": 0.004840742206718092,
12
+ "acc_norm": 0.4885480979884485,
13
+ "acc_norm_stderr": 0.0049884724594180295
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.45614035087719296,
17
+ "acc_stderr": 0.03820042586602967,
18
+ "acc_norm": 0.45614035087719296,
19
+ "acc_norm_stderr": 0.03820042586602967
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5631067961165048,
23
+ "acc_stderr": 0.04911147107365777,
24
+ "acc_norm": 0.5631067961165048,
25
+ "acc_norm_stderr": 0.04911147107365777
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5044699872286079,
29
+ "acc_stderr": 0.017879248970584353,
30
+ "acc_norm": 0.5044699872286079,
31
+ "acc_norm_stderr": 0.017879248970584353
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.35555555555555557,
35
+ "acc_stderr": 0.04135176749720386,
36
+ "acc_norm": 0.35555555555555557,
37
+ "acc_norm_stderr": 0.04135176749720386
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.26,
41
+ "acc_stderr": 0.04408440022768077,
42
+ "acc_norm": 0.26,
43
+ "acc_norm_stderr": 0.04408440022768077
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.4425531914893617,
47
+ "acc_stderr": 0.03246956919789958,
48
+ "acc_norm": 0.4425531914893617,
49
+ "acc_norm_stderr": 0.03246956919789958
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.43373493975903615,
53
+ "acc_stderr": 0.03858158940685515,
54
+ "acc_norm": 0.43373493975903615,
55
+ "acc_norm_stderr": 0.03858158940685515
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4790996784565916,
59
+ "acc_stderr": 0.028373270961069414,
60
+ "acc_norm": 0.4790996784565916,
61
+ "acc_norm_stderr": 0.028373270961069414
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.4798206278026906,
65
+ "acc_stderr": 0.033530461674123,
66
+ "acc_norm": 0.4798206278026906,
67
+ "acc_norm_stderr": 0.033530461674123
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.5190839694656488,
71
+ "acc_stderr": 0.04382094705550988,
72
+ "acc_norm": 0.5190839694656488,
73
+ "acc_norm_stderr": 0.04382094705550988
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.33,
77
+ "acc_stderr": 0.047258156262526045,
78
+ "acc_norm": 0.33,
79
+ "acc_norm_stderr": 0.047258156262526045
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5454545454545454,
83
+ "acc_stderr": 0.03547601494006937,
84
+ "acc_norm": 0.5454545454545454,
85
+ "acc_norm_stderr": 0.03547601494006937
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.46206896551724136,
89
+ "acc_stderr": 0.041546596717075474,
90
+ "acc_norm": 0.46206896551724136,
91
+ "acc_norm_stderr": 0.041546596717075474
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.2549019607843137,
95
+ "acc_stderr": 0.043364327079931785,
96
+ "acc_norm": 0.2549019607843137,
97
+ "acc_norm_stderr": 0.043364327079931785
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.49159663865546216,
101
+ "acc_stderr": 0.03247390276569669,
102
+ "acc_norm": 0.49159663865546216,
103
+ "acc_norm_stderr": 0.03247390276569669
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4205128205128205,
107
+ "acc_stderr": 0.025028610276710855,
108
+ "acc_norm": 0.4205128205128205,
109
+ "acc_norm_stderr": 0.025028610276710855
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.53,
113
+ "acc_stderr": 0.05016135580465919,
114
+ "acc_norm": 0.53,
115
+ "acc_norm_stderr": 0.05016135580465919
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.33,
119
+ "acc_stderr": 0.04725815626252605,
120
+ "acc_norm": 0.33,
121
+ "acc_norm_stderr": 0.04725815626252605
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5370370370370371,
125
+ "acc_stderr": 0.04820403072760627,
126
+ "acc_norm": 0.5370370370370371,
127
+ "acc_norm_stderr": 0.04820403072760627
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4088669950738916,
131
+ "acc_stderr": 0.034590588158832314,
132
+ "acc_norm": 0.4088669950738916,
133
+ "acc_norm_stderr": 0.034590588158832314
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4774193548387097,
137
+ "acc_stderr": 0.028414985019707868,
138
+ "acc_norm": 0.4774193548387097,
139
+ "acc_norm_stderr": 0.028414985019707868
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.7307692307692307,
143
+ "acc_stderr": 0.029058588303748845,
144
+ "acc_norm": 0.7307692307692307,
145
+ "acc_norm_stderr": 0.029058588303748845
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.41509433962264153,
149
+ "acc_stderr": 0.030325945789286105,
150
+ "acc_norm": 0.41509433962264153,
151
+ "acc_norm_stderr": 0.030325945789286105
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5181818181818182,
155
+ "acc_stderr": 0.04785964010794915,
156
+ "acc_norm": 0.5181818181818182,
157
+ "acc_norm_stderr": 0.04785964010794915
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.29259259259259257,
161
+ "acc_stderr": 0.02773896963217609,
162
+ "acc_norm": 0.29259259259259257,
163
+ "acc_norm_stderr": 0.02773896963217609
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.26490066225165565,
167
+ "acc_stderr": 0.03603038545360384,
168
+ "acc_norm": 0.26490066225165565,
169
+ "acc_norm_stderr": 0.03603038545360384
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5920398009950248,
173
+ "acc_stderr": 0.03475116365194092,
174
+ "acc_norm": 0.5920398009950248,
175
+ "acc_norm_stderr": 0.03475116365194092
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3236994219653179,
179
+ "acc_stderr": 0.0356760379963917,
180
+ "acc_norm": 0.3236994219653179,
181
+ "acc_norm_stderr": 0.0356760379963917
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.36243386243386244,
185
+ "acc_stderr": 0.024757473902752052,
186
+ "acc_norm": 0.36243386243386244,
187
+ "acc_norm_stderr": 0.024757473902752052
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3125,
191
+ "acc_stderr": 0.038760854559127644,
192
+ "acc_norm": 0.3125,
193
+ "acc_norm_stderr": 0.038760854559127644
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.33,
197
+ "acc_stderr": 0.04725815626252606,
198
+ "acc_norm": 0.33,
199
+ "acc_norm_stderr": 0.04725815626252606
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.65,
203
+ "acc_stderr": 0.047937248544110175,
204
+ "acc_norm": 0.65,
205
+ "acc_norm_stderr": 0.047937248544110175
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.4884393063583815,
209
+ "acc_stderr": 0.026911898686377913,
210
+ "acc_norm": 0.4884393063583815,
211
+ "acc_norm_stderr": 0.026911898686377913
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.44785276073619634,
215
+ "acc_stderr": 0.03906947479456601,
216
+ "acc_norm": 0.44785276073619634,
217
+ "acc_norm_stderr": 0.03906947479456601
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4660493827160494,
221
+ "acc_stderr": 0.027756535257347666,
222
+ "acc_norm": 0.4660493827160494,
223
+ "acc_norm_stderr": 0.027756535257347666
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.32,
227
+ "acc_stderr": 0.04688261722621505,
228
+ "acc_norm": 0.32,
229
+ "acc_norm_stderr": 0.04688261722621505
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5233160621761658,
233
+ "acc_stderr": 0.03604513672442202,
234
+ "acc_norm": 0.5233160621761658,
235
+ "acc_norm_stderr": 0.03604513672442202
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.3508771929824561,
239
+ "acc_stderr": 0.04489539350270698,
240
+ "acc_norm": 0.3508771929824561,
241
+ "acc_norm_stderr": 0.04489539350270698
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.4972477064220184,
245
+ "acc_stderr": 0.021436998359765317,
246
+ "acc_norm": 0.4972477064220184,
247
+ "acc_norm_stderr": 0.021436998359765317
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3253968253968254,
251
+ "acc_stderr": 0.04190596438871136,
252
+ "acc_norm": 0.3253968253968254,
253
+ "acc_norm_stderr": 0.04190596438871136
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4411764705882353,
257
+ "acc_stderr": 0.028431095444176643,
258
+ "acc_norm": 0.4411764705882353,
259
+ "acc_norm_stderr": 0.028431095444176643
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.46,
263
+ "acc_stderr": 0.05009082659620333,
264
+ "acc_norm": 0.46,
265
+ "acc_norm_stderr": 0.05009082659620333
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6694214876033058,
269
+ "acc_stderr": 0.04294340845212095,
270
+ "acc_norm": 0.6694214876033058,
271
+ "acc_norm_stderr": 0.04294340845212095
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4276315789473684,
275
+ "acc_stderr": 0.04026097083296558,
276
+ "acc_norm": 0.4276315789473684,
277
+ "acc_norm_stderr": 0.04026097083296558
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.40032679738562094,
281
+ "acc_stderr": 0.019821843688271768,
282
+ "acc_norm": 0.40032679738562094,
283
+ "acc_norm_stderr": 0.019821843688271768
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3617021276595745,
287
+ "acc_stderr": 0.028663820147199492,
288
+ "acc_norm": 0.3617021276595745,
289
+ "acc_norm_stderr": 0.028663820147199492
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.36607142857142855,
293
+ "acc_stderr": 0.0457237235873743,
294
+ "acc_norm": 0.36607142857142855,
295
+ "acc_norm_stderr": 0.0457237235873743
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.39351851851851855,
299
+ "acc_stderr": 0.03331747876370312,
300
+ "acc_norm": 0.39351851851851855,
301
+ "acc_norm_stderr": 0.03331747876370312
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2558659217877095,
305
+ "acc_stderr": 0.014593620923210746,
306
+ "acc_norm": 0.2558659217877095,
307
+ "acc_norm_stderr": 0.014593620923210746
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.36,
311
+ "acc_stderr": 0.04824181513244218,
312
+ "acc_norm": 0.36,
313
+ "acc_norm_stderr": 0.04824181513244218
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.57,
317
+ "acc_stderr": 0.049756985195624284,
318
+ "acc_norm": 0.57,
319
+ "acc_norm_stderr": 0.049756985195624284
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.33455882352941174,
323
+ "acc_stderr": 0.028661996202335317,
324
+ "acc_norm": 0.33455882352941174,
325
+ "acc_norm_stderr": 0.028661996202335317
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5224489795918368,
329
+ "acc_stderr": 0.03197694118713672,
330
+ "acc_norm": 0.5224489795918368,
331
+ "acc_norm_stderr": 0.03197694118713672
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5654008438818565,
335
+ "acc_stderr": 0.03226759995510145,
336
+ "acc_norm": 0.5654008438818565,
337
+ "acc_norm_stderr": 0.03226759995510145
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3213820078226858,
341
+ "acc_stderr": 0.011927581352265076,
342
+ "acc_norm": 0.3213820078226858,
343
+ "acc_norm_stderr": 0.011927581352265076
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4215686274509804,
347
+ "acc_stderr": 0.03465868196380757,
348
+ "acc_norm": 0.4215686274509804,
349
+ "acc_norm_stderr": 0.03465868196380757
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.4727272727272727,
353
+ "acc_stderr": 0.03898531605579419,
354
+ "acc_norm": 0.4727272727272727,
355
+ "acc_norm_stderr": 0.03898531605579419
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2668298653610771,
359
+ "mc1_stderr": 0.015483691939237272,
360
+ "mc2": 0.4338150951405425,
361
+ "mc2_stderr": 0.01536129905959147
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.4982290436835891,
365
+ "acc_stderr": 0.01719024627623186,
366
+ "acc_norm": 0.5489964580873672,
367
+ "acc_norm_stderr": 0.01710761885954935
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AIdenU/Mistral-7b-ko-Y24_v0.1",
436
+ "model_sha": "a0b72b81f985f7fb06695cae82877ca482947dbf",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
AtAndDev/ShortKingv0.1/result_2023-09-29 19:59:47.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.19112627986348124,
5
+ "acc_stderr": 0.011490055292778596,
6
+ "acc_norm": 0.24829351535836178,
7
+ "acc_norm_stderr": 0.012624912868089764
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.2756423023302131,
11
+ "acc_stderr": 0.0044592414745187915,
12
+ "acc_norm": 0.29884485162318264,
13
+ "acc_norm_stderr": 0.004568161710399566
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.27485380116959063,
17
+ "acc_stderr": 0.03424042924691582,
18
+ "acc_norm": 0.27485380116959063,
19
+ "acc_norm_stderr": 0.03424042924691582
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.17475728155339806,
23
+ "acc_stderr": 0.037601780060266196,
24
+ "acc_norm": 0.17475728155339806,
25
+ "acc_norm_stderr": 0.037601780060266196
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.2669220945083014,
29
+ "acc_stderr": 0.015818450894777555,
30
+ "acc_norm": 0.2669220945083014,
31
+ "acc_norm_stderr": 0.015818450894777555
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.3111111111111111,
35
+ "acc_stderr": 0.03999262876617722,
36
+ "acc_norm": 0.3111111111111111,
37
+ "acc_norm_stderr": 0.03999262876617722
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.23,
41
+ "acc_stderr": 0.04229525846816506,
42
+ "acc_norm": 0.23,
43
+ "acc_norm_stderr": 0.04229525846816506
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.2936170212765957,
47
+ "acc_stderr": 0.02977164271249123,
48
+ "acc_norm": 0.2936170212765957,
49
+ "acc_norm_stderr": 0.02977164271249123
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.27710843373493976,
53
+ "acc_stderr": 0.03484331592680588,
54
+ "acc_norm": 0.27710843373493976,
55
+ "acc_norm_stderr": 0.03484331592680588
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.20257234726688103,
59
+ "acc_stderr": 0.022827317491059686,
60
+ "acc_norm": 0.20257234726688103,
61
+ "acc_norm_stderr": 0.022827317491059686
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.3632286995515695,
65
+ "acc_stderr": 0.032277904428505,
66
+ "acc_norm": 0.3632286995515695,
67
+ "acc_norm_stderr": 0.032277904428505
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.2824427480916031,
71
+ "acc_stderr": 0.03948406125768361,
72
+ "acc_norm": 0.2824427480916031,
73
+ "acc_norm_stderr": 0.03948406125768361
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.3,
77
+ "acc_stderr": 0.046056618647183814,
78
+ "acc_norm": 0.3,
79
+ "acc_norm_stderr": 0.046056618647183814
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.21717171717171718,
83
+ "acc_stderr": 0.029376616484945644,
84
+ "acc_norm": 0.21717171717171718,
85
+ "acc_norm_stderr": 0.029376616484945644
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.2206896551724138,
89
+ "acc_stderr": 0.03455930201924812,
90
+ "acc_norm": 0.2206896551724138,
91
+ "acc_norm_stderr": 0.03455930201924812
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.14705882352941177,
95
+ "acc_stderr": 0.035240689515674474,
96
+ "acc_norm": 0.14705882352941177,
97
+ "acc_norm_stderr": 0.035240689515674474
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.226890756302521,
101
+ "acc_stderr": 0.027205371538279476,
102
+ "acc_norm": 0.226890756302521,
103
+ "acc_norm_stderr": 0.027205371538279476
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.2153846153846154,
107
+ "acc_stderr": 0.020843034557462878,
108
+ "acc_norm": 0.2153846153846154,
109
+ "acc_norm_stderr": 0.020843034557462878
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.21,
113
+ "acc_stderr": 0.040936018074033256,
114
+ "acc_norm": 0.21,
115
+ "acc_norm_stderr": 0.040936018074033256
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.26,
119
+ "acc_stderr": 0.0440844002276808,
120
+ "acc_norm": 0.26,
121
+ "acc_norm_stderr": 0.0440844002276808
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.25,
125
+ "acc_stderr": 0.04186091791394607,
126
+ "acc_norm": 0.25,
127
+ "acc_norm_stderr": 0.04186091791394607
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.13793103448275862,
131
+ "acc_stderr": 0.024261984301044565,
132
+ "acc_norm": 0.13793103448275862,
133
+ "acc_norm_stderr": 0.024261984301044565
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.20967741935483872,
137
+ "acc_stderr": 0.023157879349083536,
138
+ "acc_norm": 0.20967741935483872,
139
+ "acc_norm_stderr": 0.023157879349083536
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.2094017094017094,
143
+ "acc_stderr": 0.026655699653922737,
144
+ "acc_norm": 0.2094017094017094,
145
+ "acc_norm_stderr": 0.026655699653922737
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.2490566037735849,
149
+ "acc_stderr": 0.026616482980501715,
150
+ "acc_norm": 0.2490566037735849,
151
+ "acc_norm_stderr": 0.026616482980501715
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.21818181818181817,
155
+ "acc_stderr": 0.03955932861795833,
156
+ "acc_norm": 0.21818181818181817,
157
+ "acc_norm_stderr": 0.03955932861795833
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.26296296296296295,
161
+ "acc_stderr": 0.02684205787383371,
162
+ "acc_norm": 0.26296296296296295,
163
+ "acc_norm_stderr": 0.02684205787383371
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2847682119205298,
167
+ "acc_stderr": 0.03684881521389024,
168
+ "acc_norm": 0.2847682119205298,
169
+ "acc_norm_stderr": 0.03684881521389024
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.24378109452736318,
173
+ "acc_stderr": 0.03036049015401464,
174
+ "acc_norm": 0.24378109452736318,
175
+ "acc_norm_stderr": 0.03036049015401464
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.2138728323699422,
179
+ "acc_stderr": 0.03126511206173043,
180
+ "acc_norm": 0.2138728323699422,
181
+ "acc_norm_stderr": 0.03126511206173043
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.25396825396825395,
185
+ "acc_stderr": 0.022418042891113946,
186
+ "acc_norm": 0.25396825396825395,
187
+ "acc_norm_stderr": 0.022418042891113946
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.2222222222222222,
191
+ "acc_stderr": 0.03476590104304134,
192
+ "acc_norm": 0.2222222222222222,
193
+ "acc_norm_stderr": 0.03476590104304134
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.24,
197
+ "acc_stderr": 0.04292346959909283,
198
+ "acc_norm": 0.24,
199
+ "acc_norm_stderr": 0.04292346959909283
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.26,
203
+ "acc_stderr": 0.0440844002276808,
204
+ "acc_norm": 0.26,
205
+ "acc_norm_stderr": 0.0440844002276808
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.26011560693641617,
209
+ "acc_stderr": 0.023618678310069363,
210
+ "acc_norm": 0.26011560693641617,
211
+ "acc_norm_stderr": 0.023618678310069363
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.27607361963190186,
215
+ "acc_stderr": 0.0351238528370505,
216
+ "acc_norm": 0.27607361963190186,
217
+ "acc_norm_stderr": 0.0351238528370505
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.24074074074074073,
221
+ "acc_stderr": 0.02378858355165854,
222
+ "acc_norm": 0.24074074074074073,
223
+ "acc_norm_stderr": 0.02378858355165854
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.29,
227
+ "acc_stderr": 0.045604802157206845,
228
+ "acc_norm": 0.29,
229
+ "acc_norm_stderr": 0.045604802157206845
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.20725388601036268,
233
+ "acc_stderr": 0.02925282329180363,
234
+ "acc_norm": 0.20725388601036268,
235
+ "acc_norm_stderr": 0.02925282329180363
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2807017543859649,
239
+ "acc_stderr": 0.042270544512322004,
240
+ "acc_norm": 0.2807017543859649,
241
+ "acc_norm_stderr": 0.042270544512322004
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.21100917431192662,
245
+ "acc_stderr": 0.017493922404112648,
246
+ "acc_norm": 0.21100917431192662,
247
+ "acc_norm_stderr": 0.017493922404112648
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.23809523809523808,
251
+ "acc_stderr": 0.038095238095238106,
252
+ "acc_norm": 0.23809523809523808,
253
+ "acc_norm_stderr": 0.038095238095238106
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.23529411764705882,
257
+ "acc_stderr": 0.024288619466046116,
258
+ "acc_norm": 0.23529411764705882,
259
+ "acc_norm_stderr": 0.024288619466046116
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.3,
263
+ "acc_stderr": 0.046056618647183814,
264
+ "acc_norm": 0.3,
265
+ "acc_norm_stderr": 0.046056618647183814
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.19008264462809918,
269
+ "acc_stderr": 0.03581796951709282,
270
+ "acc_norm": 0.19008264462809918,
271
+ "acc_norm_stderr": 0.03581796951709282
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.17763157894736842,
275
+ "acc_stderr": 0.03110318238312338,
276
+ "acc_norm": 0.17763157894736842,
277
+ "acc_norm_stderr": 0.03110318238312338
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.25980392156862747,
281
+ "acc_stderr": 0.017740899509177795,
282
+ "acc_norm": 0.25980392156862747,
283
+ "acc_norm_stderr": 0.017740899509177795
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.2765957446808511,
287
+ "acc_stderr": 0.026684564340460997,
288
+ "acc_norm": 0.2765957446808511,
289
+ "acc_norm_stderr": 0.026684564340460997
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.2857142857142857,
293
+ "acc_stderr": 0.042878587513404544,
294
+ "acc_norm": 0.2857142857142857,
295
+ "acc_norm_stderr": 0.042878587513404544
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.35648148148148145,
299
+ "acc_stderr": 0.032664783315272714,
300
+ "acc_norm": 0.35648148148148145,
301
+ "acc_norm_stderr": 0.032664783315272714
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.23016759776536314,
305
+ "acc_stderr": 0.014078339253425809,
306
+ "acc_norm": 0.23016759776536314,
307
+ "acc_norm_stderr": 0.014078339253425809
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.31,
311
+ "acc_stderr": 0.04648231987117316,
312
+ "acc_norm": 0.31,
313
+ "acc_norm_stderr": 0.04648231987117316
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.28,
317
+ "acc_stderr": 0.04512608598542128,
318
+ "acc_norm": 0.28,
319
+ "acc_norm_stderr": 0.04512608598542128
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.29411764705882354,
323
+ "acc_stderr": 0.02767846864214471,
324
+ "acc_norm": 0.29411764705882354,
325
+ "acc_norm_stderr": 0.02767846864214471
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.3142857142857143,
329
+ "acc_stderr": 0.029719329422417468,
330
+ "acc_norm": 0.3142857142857143,
331
+ "acc_norm_stderr": 0.029719329422417468
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.2742616033755274,
335
+ "acc_stderr": 0.02904133351059804,
336
+ "acc_norm": 0.2742616033755274,
337
+ "acc_norm_stderr": 0.02904133351059804
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.24511082138200782,
341
+ "acc_stderr": 0.010986307870045526,
342
+ "acc_norm": 0.24511082138200782,
343
+ "acc_norm_stderr": 0.010986307870045526
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.27941176470588236,
347
+ "acc_stderr": 0.03149328104507955,
348
+ "acc_norm": 0.27941176470588236,
349
+ "acc_norm_stderr": 0.03149328104507955
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.21818181818181817,
353
+ "acc_stderr": 0.03225078108306289,
354
+ "acc_norm": 0.21818181818181817,
355
+ "acc_norm_stderr": 0.03225078108306289
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.29498164014687883,
359
+ "mc1_stderr": 0.015964400965589674,
360
+ "mc2": 0.49219803033147647,
361
+ "mc2_stderr": 0.015947492879186672
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.2361275088547816,
365
+ "acc_stderr": 0.014601536093324397,
366
+ "acc_norm": 0.27508854781582054,
367
+ "acc_norm_stderr": 0.015353010757952649
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "AtAndDev/ShortKingv0.1",
436
+ "model_sha": "6cd9b5bc13ee15b5e7e7cfb46477bc6a7c0b5d47",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/llama-2-ko-7b-it-v1.0.0/result_2023-11-15 11:34:52.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.34897610921501704,
5
+ "acc_stderr": 0.0139289334613825,
6
+ "acc_norm": 0.38310580204778155,
7
+ "acc_norm_stderr": 0.01420647266167288
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3973312089225254,
11
+ "acc_stderr": 0.00488345518890897,
12
+ "acc_norm": 0.518621788488349,
13
+ "acc_norm_stderr": 0.004986319587524962
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.3508771929824561,
17
+ "acc_stderr": 0.036602988340491624,
18
+ "acc_norm": 0.3508771929824561,
19
+ "acc_norm_stderr": 0.036602988340491624
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.2524271844660194,
23
+ "acc_stderr": 0.04301250399690879,
24
+ "acc_norm": 0.2524271844660194,
25
+ "acc_norm_stderr": 0.04301250399690879
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.35759897828863346,
29
+ "acc_stderr": 0.017139488998803288,
30
+ "acc_norm": 0.35759897828863346,
31
+ "acc_norm_stderr": 0.017139488998803288
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.362962962962963,
35
+ "acc_stderr": 0.041539484047424,
36
+ "acc_norm": 0.362962962962963,
37
+ "acc_norm_stderr": 0.041539484047424
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.29,
41
+ "acc_stderr": 0.04560480215720683,
42
+ "acc_norm": 0.29,
43
+ "acc_norm_stderr": 0.04560480215720683
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.28936170212765955,
47
+ "acc_stderr": 0.029644006577009618,
48
+ "acc_norm": 0.28936170212765955,
49
+ "acc_norm_stderr": 0.029644006577009618
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.25301204819277107,
53
+ "acc_stderr": 0.033844291552331346,
54
+ "acc_norm": 0.25301204819277107,
55
+ "acc_norm_stderr": 0.033844291552331346
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.3729903536977492,
59
+ "acc_stderr": 0.027466610213140105,
60
+ "acc_norm": 0.3729903536977492,
61
+ "acc_norm_stderr": 0.027466610213140105
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.26905829596412556,
65
+ "acc_stderr": 0.029763779406874975,
66
+ "acc_norm": 0.26905829596412556,
67
+ "acc_norm_stderr": 0.029763779406874975
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.33587786259541985,
71
+ "acc_stderr": 0.04142313771996664,
72
+ "acc_norm": 0.33587786259541985,
73
+ "acc_norm_stderr": 0.04142313771996664
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.33,
77
+ "acc_stderr": 0.047258156262526045,
78
+ "acc_norm": 0.33,
79
+ "acc_norm_stderr": 0.047258156262526045
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.3888888888888889,
83
+ "acc_stderr": 0.0347327959083696,
84
+ "acc_norm": 0.3888888888888889,
85
+ "acc_norm_stderr": 0.0347327959083696
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.38620689655172413,
89
+ "acc_stderr": 0.04057324734419036,
90
+ "acc_norm": 0.38620689655172413,
91
+ "acc_norm_stderr": 0.04057324734419036
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.18627450980392157,
95
+ "acc_stderr": 0.038739587141493524,
96
+ "acc_norm": 0.18627450980392157,
97
+ "acc_norm_stderr": 0.038739587141493524
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.25210084033613445,
101
+ "acc_stderr": 0.028205545033277723,
102
+ "acc_norm": 0.25210084033613445,
103
+ "acc_norm_stderr": 0.028205545033277723
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.23076923076923078,
107
+ "acc_stderr": 0.021362027725222738,
108
+ "acc_norm": 0.23076923076923078,
109
+ "acc_norm_stderr": 0.021362027725222738
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.33,
113
+ "acc_stderr": 0.04725815626252605,
114
+ "acc_norm": 0.33,
115
+ "acc_norm_stderr": 0.04725815626252605
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.19,
119
+ "acc_stderr": 0.03942772444036623,
120
+ "acc_norm": 0.19,
121
+ "acc_norm_stderr": 0.03942772444036623
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.35185185185185186,
125
+ "acc_stderr": 0.04616631111801713,
126
+ "acc_norm": 0.35185185185185186,
127
+ "acc_norm_stderr": 0.04616631111801713
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.270935960591133,
131
+ "acc_stderr": 0.031270907132976984,
132
+ "acc_norm": 0.270935960591133,
133
+ "acc_norm_stderr": 0.031270907132976984
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.2806451612903226,
137
+ "acc_stderr": 0.025560604721022877,
138
+ "acc_norm": 0.2806451612903226,
139
+ "acc_norm_stderr": 0.025560604721022877
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.3974358974358974,
143
+ "acc_stderr": 0.03205953453789293,
144
+ "acc_norm": 0.3974358974358974,
145
+ "acc_norm_stderr": 0.03205953453789293
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.32452830188679244,
149
+ "acc_stderr": 0.028815615713432118,
150
+ "acc_norm": 0.32452830188679244,
151
+ "acc_norm_stderr": 0.028815615713432118
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.2818181818181818,
155
+ "acc_stderr": 0.04309118709946458,
156
+ "acc_norm": 0.2818181818181818,
157
+ "acc_norm_stderr": 0.04309118709946458
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.26666666666666666,
161
+ "acc_stderr": 0.026962424325073838,
162
+ "acc_norm": 0.26666666666666666,
163
+ "acc_norm_stderr": 0.026962424325073838
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2781456953642384,
167
+ "acc_stderr": 0.03658603262763743,
168
+ "acc_norm": 0.2781456953642384,
169
+ "acc_norm_stderr": 0.03658603262763743
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.3880597014925373,
173
+ "acc_stderr": 0.034457899643627506,
174
+ "acc_norm": 0.3880597014925373,
175
+ "acc_norm_stderr": 0.034457899643627506
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.27167630057803466,
179
+ "acc_stderr": 0.03391750322321658,
180
+ "acc_norm": 0.27167630057803466,
181
+ "acc_norm_stderr": 0.03391750322321658
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.2671957671957672,
185
+ "acc_stderr": 0.022789673145776578,
186
+ "acc_norm": 0.2671957671957672,
187
+ "acc_norm_stderr": 0.022789673145776578
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.2777777777777778,
191
+ "acc_stderr": 0.037455547914624576,
192
+ "acc_norm": 0.2777777777777778,
193
+ "acc_norm_stderr": 0.037455547914624576
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.16,
197
+ "acc_stderr": 0.0368452949177471,
198
+ "acc_norm": 0.16,
199
+ "acc_norm_stderr": 0.0368452949177471
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.34,
203
+ "acc_stderr": 0.047609522856952365,
204
+ "acc_norm": 0.34,
205
+ "acc_norm_stderr": 0.047609522856952365
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.33815028901734107,
209
+ "acc_stderr": 0.02546977014940017,
210
+ "acc_norm": 0.33815028901734107,
211
+ "acc_norm_stderr": 0.02546977014940017
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.2392638036809816,
215
+ "acc_stderr": 0.03351953879521269,
216
+ "acc_norm": 0.2392638036809816,
217
+ "acc_norm_stderr": 0.03351953879521269
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.3549382716049383,
221
+ "acc_stderr": 0.026624152478845853,
222
+ "acc_norm": 0.3549382716049383,
223
+ "acc_norm_stderr": 0.026624152478845853
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.27,
227
+ "acc_stderr": 0.0446196043338474,
228
+ "acc_norm": 0.27,
229
+ "acc_norm_stderr": 0.0446196043338474
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.23316062176165803,
233
+ "acc_stderr": 0.03051611137147601,
234
+ "acc_norm": 0.23316062176165803,
235
+ "acc_norm_stderr": 0.03051611137147601
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.23684210526315788,
239
+ "acc_stderr": 0.039994238792813344,
240
+ "acc_norm": 0.23684210526315788,
241
+ "acc_norm_stderr": 0.039994238792813344
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.3211009174311927,
245
+ "acc_stderr": 0.020018149772733747,
246
+ "acc_norm": 0.3211009174311927,
247
+ "acc_norm_stderr": 0.020018149772733747
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.18253968253968253,
251
+ "acc_stderr": 0.03455071019102149,
252
+ "acc_norm": 0.18253968253968253,
253
+ "acc_norm_stderr": 0.03455071019102149
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.3235294117647059,
257
+ "acc_stderr": 0.026787453111906532,
258
+ "acc_norm": 0.3235294117647059,
259
+ "acc_norm_stderr": 0.026787453111906532
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.26,
263
+ "acc_stderr": 0.04408440022768079,
264
+ "acc_norm": 0.26,
265
+ "acc_norm_stderr": 0.04408440022768079
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.4628099173553719,
269
+ "acc_stderr": 0.045517111961042175,
270
+ "acc_norm": 0.4628099173553719,
271
+ "acc_norm_stderr": 0.045517111961042175
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3355263157894737,
275
+ "acc_stderr": 0.03842498559395268,
276
+ "acc_norm": 0.3355263157894737,
277
+ "acc_norm_stderr": 0.03842498559395268
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3006535947712418,
281
+ "acc_stderr": 0.018550634502952964,
282
+ "acc_norm": 0.3006535947712418,
283
+ "acc_norm_stderr": 0.018550634502952964
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.2730496453900709,
287
+ "acc_stderr": 0.026577860943307854,
288
+ "acc_norm": 0.2730496453900709,
289
+ "acc_norm_stderr": 0.026577860943307854
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.22321428571428573,
293
+ "acc_stderr": 0.03952301967702511,
294
+ "acc_norm": 0.22321428571428573,
295
+ "acc_norm_stderr": 0.03952301967702511
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.25462962962962965,
299
+ "acc_stderr": 0.02971127586000534,
300
+ "acc_norm": 0.25462962962962965,
301
+ "acc_norm_stderr": 0.02971127586000534
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.25251396648044694,
305
+ "acc_stderr": 0.014530330201468638,
306
+ "acc_norm": 0.25251396648044694,
307
+ "acc_norm_stderr": 0.014530330201468638
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.36,
311
+ "acc_stderr": 0.048241815132442176,
312
+ "acc_norm": 0.36,
313
+ "acc_norm_stderr": 0.048241815132442176
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.3,
317
+ "acc_stderr": 0.046056618647183814,
318
+ "acc_norm": 0.3,
319
+ "acc_norm_stderr": 0.046056618647183814
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.27941176470588236,
323
+ "acc_stderr": 0.027257202606114948,
324
+ "acc_norm": 0.27941176470588236,
325
+ "acc_norm_stderr": 0.027257202606114948
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.2816326530612245,
329
+ "acc_stderr": 0.0287951855742913,
330
+ "acc_norm": 0.2816326530612245,
331
+ "acc_norm_stderr": 0.0287951855742913
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.35864978902953587,
335
+ "acc_stderr": 0.031219569445301843,
336
+ "acc_norm": 0.35864978902953587,
337
+ "acc_norm_stderr": 0.031219569445301843
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.2907431551499348,
341
+ "acc_stderr": 0.011598062372851981,
342
+ "acc_norm": 0.2907431551499348,
343
+ "acc_norm_stderr": 0.011598062372851981
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.25980392156862747,
347
+ "acc_stderr": 0.030778554678693268,
348
+ "acc_norm": 0.25980392156862747,
349
+ "acc_norm_stderr": 0.030778554678693268
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.28484848484848485,
353
+ "acc_stderr": 0.03524390844511784,
354
+ "acc_norm": 0.28484848484848485,
355
+ "acc_norm_stderr": 0.03524390844511784
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.24112607099143207,
359
+ "mc1_stderr": 0.014974827279752329,
360
+ "mc2": 0.3762518297834469,
361
+ "mc2_stderr": 0.015197001689915996
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.31759149940968123,
365
+ "acc_stderr": 0.016005581876229306,
366
+ "acc_norm": 0.3990554899645809,
367
+ "acc_norm_stderr": 0.0168363772928493
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/llama-2-ko-7b-it-v1.0.0",
436
+ "model_sha": "d77fd44b31382f84fa4b8b9afd63a92ded7bde93",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.0.1/result_2023-11-07 22:19:25.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.35665529010238906,
5
+ "acc_stderr": 0.013998056902620196,
6
+ "acc_norm": 0.41467576791808874,
7
+ "acc_norm_stderr": 0.014397070564409174
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.38329018123879705,
11
+ "acc_stderr": 0.004851944170671259,
12
+ "acc_norm": 0.4987054371639116,
13
+ "acc_norm_stderr": 0.004989764686738831
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.52046783625731,
17
+ "acc_stderr": 0.0383161053282193,
18
+ "acc_norm": 0.52046783625731,
19
+ "acc_norm_stderr": 0.0383161053282193
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5339805825242718,
23
+ "acc_stderr": 0.04939291447273481,
24
+ "acc_norm": 0.5339805825242718,
25
+ "acc_norm_stderr": 0.04939291447273481
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5159642401021711,
29
+ "acc_stderr": 0.017870847506081738,
30
+ "acc_norm": 0.5159642401021711,
31
+ "acc_norm_stderr": 0.017870847506081738
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.3925925925925926,
35
+ "acc_stderr": 0.04218506215368879,
36
+ "acc_norm": 0.3925925925925926,
37
+ "acc_norm_stderr": 0.04218506215368879
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.35,
41
+ "acc_stderr": 0.047937248544110196,
42
+ "acc_norm": 0.35,
43
+ "acc_norm_stderr": 0.047937248544110196
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.37872340425531914,
47
+ "acc_stderr": 0.031709956060406545,
48
+ "acc_norm": 0.37872340425531914,
49
+ "acc_norm_stderr": 0.031709956060406545
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39156626506024095,
53
+ "acc_stderr": 0.03799857454479636,
54
+ "acc_norm": 0.39156626506024095,
55
+ "acc_norm_stderr": 0.03799857454479636
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4790996784565916,
59
+ "acc_stderr": 0.028373270961069414,
60
+ "acc_norm": 0.4790996784565916,
61
+ "acc_norm_stderr": 0.028373270961069414
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.49327354260089684,
65
+ "acc_stderr": 0.03355476596234353,
66
+ "acc_norm": 0.49327354260089684,
67
+ "acc_norm_stderr": 0.03355476596234353
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4732824427480916,
71
+ "acc_stderr": 0.04379024936553894,
72
+ "acc_norm": 0.4732824427480916,
73
+ "acc_norm_stderr": 0.04379024936553894
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.38,
77
+ "acc_stderr": 0.048783173121456316,
78
+ "acc_norm": 0.38,
79
+ "acc_norm_stderr": 0.048783173121456316
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5555555555555556,
83
+ "acc_stderr": 0.035402943770953675,
84
+ "acc_norm": 0.5555555555555556,
85
+ "acc_norm_stderr": 0.035402943770953675
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.46206896551724136,
89
+ "acc_stderr": 0.041546596717075474,
90
+ "acc_norm": 0.46206896551724136,
91
+ "acc_norm_stderr": 0.041546596717075474
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.13725490196078433,
95
+ "acc_stderr": 0.03424084669891524,
96
+ "acc_norm": 0.13725490196078433,
97
+ "acc_norm_stderr": 0.03424084669891524
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.41596638655462187,
101
+ "acc_stderr": 0.03201650100739615,
102
+ "acc_norm": 0.41596638655462187,
103
+ "acc_norm_stderr": 0.03201650100739615
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4230769230769231,
107
+ "acc_stderr": 0.025049197876042335,
108
+ "acc_norm": 0.4230769230769231,
109
+ "acc_norm_stderr": 0.025049197876042335
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.56,
113
+ "acc_stderr": 0.04988876515698589,
114
+ "acc_norm": 0.56,
115
+ "acc_norm_stderr": 0.04988876515698589
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.32,
119
+ "acc_stderr": 0.04688261722621504,
120
+ "acc_norm": 0.32,
121
+ "acc_norm_stderr": 0.04688261722621504
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5,
125
+ "acc_stderr": 0.04833682445228318,
126
+ "acc_norm": 0.5,
127
+ "acc_norm_stderr": 0.04833682445228318
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.39408866995073893,
131
+ "acc_stderr": 0.034381579670365446,
132
+ "acc_norm": 0.39408866995073893,
133
+ "acc_norm_stderr": 0.034381579670365446
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.432258064516129,
137
+ "acc_stderr": 0.028181739720019413,
138
+ "acc_norm": 0.432258064516129,
139
+ "acc_norm_stderr": 0.028181739720019413
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.7307692307692307,
143
+ "acc_stderr": 0.029058588303748842,
144
+ "acc_norm": 0.7307692307692307,
145
+ "acc_norm_stderr": 0.029058588303748842
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.41509433962264153,
149
+ "acc_stderr": 0.030325945789286105,
150
+ "acc_norm": 0.41509433962264153,
151
+ "acc_norm_stderr": 0.030325945789286105
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5181818181818182,
155
+ "acc_stderr": 0.04785964010794915,
156
+ "acc_norm": 0.5181818181818182,
157
+ "acc_norm_stderr": 0.04785964010794915
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.34074074074074073,
161
+ "acc_stderr": 0.02889774874113114,
162
+ "acc_norm": 0.34074074074074073,
163
+ "acc_norm_stderr": 0.02889774874113114
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.271523178807947,
167
+ "acc_stderr": 0.03631329803969653,
168
+ "acc_norm": 0.271523178807947,
169
+ "acc_norm_stderr": 0.03631329803969653
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.6019900497512438,
173
+ "acc_stderr": 0.03461199429040014,
174
+ "acc_norm": 0.6019900497512438,
175
+ "acc_norm_stderr": 0.03461199429040014
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.35260115606936415,
179
+ "acc_stderr": 0.036430371689585496,
180
+ "acc_norm": 0.35260115606936415,
181
+ "acc_norm_stderr": 0.036430371689585496
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.37037037037037035,
185
+ "acc_stderr": 0.0248708152510571,
186
+ "acc_norm": 0.37037037037037035,
187
+ "acc_norm_stderr": 0.0248708152510571
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3888888888888889,
191
+ "acc_stderr": 0.04076663253918567,
192
+ "acc_norm": 0.3888888888888889,
193
+ "acc_norm_stderr": 0.04076663253918567
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.32,
197
+ "acc_stderr": 0.04688261722621504,
198
+ "acc_norm": 0.32,
199
+ "acc_norm_stderr": 0.04688261722621504
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.58,
203
+ "acc_stderr": 0.049604496374885836,
204
+ "acc_norm": 0.58,
205
+ "acc_norm_stderr": 0.049604496374885836
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5202312138728323,
209
+ "acc_stderr": 0.026897049996382868,
210
+ "acc_norm": 0.5202312138728323,
211
+ "acc_norm_stderr": 0.026897049996382868
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.43558282208588955,
215
+ "acc_stderr": 0.03895632464138937,
216
+ "acc_norm": 0.43558282208588955,
217
+ "acc_norm_stderr": 0.03895632464138937
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.44135802469135804,
221
+ "acc_stderr": 0.027628737155668784,
222
+ "acc_norm": 0.44135802469135804,
223
+ "acc_norm_stderr": 0.027628737155668784
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.31,
227
+ "acc_stderr": 0.04648231987117316,
228
+ "acc_norm": 0.31,
229
+ "acc_norm_stderr": 0.04648231987117316
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5751295336787565,
233
+ "acc_stderr": 0.035674713352125395,
234
+ "acc_norm": 0.5751295336787565,
235
+ "acc_norm_stderr": 0.035674713352125395
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.32456140350877194,
239
+ "acc_stderr": 0.04404556157374768,
240
+ "acc_norm": 0.32456140350877194,
241
+ "acc_norm_stderr": 0.04404556157374768
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.46605504587155966,
245
+ "acc_stderr": 0.021387863350354,
246
+ "acc_norm": 0.46605504587155966,
247
+ "acc_norm_stderr": 0.021387863350354
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.30158730158730157,
251
+ "acc_stderr": 0.04104947269903394,
252
+ "acc_norm": 0.30158730158730157,
253
+ "acc_norm_stderr": 0.04104947269903394
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4738562091503268,
257
+ "acc_stderr": 0.028590752958852394,
258
+ "acc_norm": 0.4738562091503268,
259
+ "acc_norm_stderr": 0.028590752958852394
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.48,
263
+ "acc_stderr": 0.050211673156867795,
264
+ "acc_norm": 0.48,
265
+ "acc_norm_stderr": 0.050211673156867795
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.5950413223140496,
269
+ "acc_stderr": 0.04481137755942469,
270
+ "acc_norm": 0.5950413223140496,
271
+ "acc_norm_stderr": 0.04481137755942469
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.34868421052631576,
275
+ "acc_stderr": 0.038781398887976104,
276
+ "acc_norm": 0.34868421052631576,
277
+ "acc_norm_stderr": 0.038781398887976104
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.40032679738562094,
281
+ "acc_stderr": 0.01982184368827177,
282
+ "acc_norm": 0.40032679738562094,
283
+ "acc_norm_stderr": 0.01982184368827177
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.30141843971631205,
287
+ "acc_stderr": 0.027374128882631157,
288
+ "acc_norm": 0.30141843971631205,
289
+ "acc_norm_stderr": 0.027374128882631157
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.3125,
293
+ "acc_stderr": 0.043994650575715215,
294
+ "acc_norm": 0.3125,
295
+ "acc_norm_stderr": 0.043994650575715215
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.30092592592592593,
299
+ "acc_stderr": 0.031280390843298825,
300
+ "acc_norm": 0.30092592592592593,
301
+ "acc_norm_stderr": 0.031280390843298825
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2435754189944134,
305
+ "acc_stderr": 0.014355911964767864,
306
+ "acc_norm": 0.2435754189944134,
307
+ "acc_norm_stderr": 0.014355911964767864
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.37,
311
+ "acc_stderr": 0.04852365870939099,
312
+ "acc_norm": 0.37,
313
+ "acc_norm_stderr": 0.04852365870939099
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.58,
317
+ "acc_stderr": 0.049604496374885836,
318
+ "acc_norm": 0.58,
319
+ "acc_norm_stderr": 0.049604496374885836
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3272058823529412,
323
+ "acc_stderr": 0.02850145286039656,
324
+ "acc_norm": 0.3272058823529412,
325
+ "acc_norm_stderr": 0.02850145286039656
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.44081632653061226,
329
+ "acc_stderr": 0.03178419114175363,
330
+ "acc_norm": 0.44081632653061226,
331
+ "acc_norm_stderr": 0.03178419114175363
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5611814345991561,
335
+ "acc_stderr": 0.032302649315470375,
336
+ "acc_norm": 0.5611814345991561,
337
+ "acc_norm_stderr": 0.032302649315470375
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3344198174706649,
341
+ "acc_stderr": 0.012049668983214936,
342
+ "acc_norm": 0.3344198174706649,
343
+ "acc_norm_stderr": 0.012049668983214936
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.44607843137254904,
347
+ "acc_stderr": 0.03488845451304974,
348
+ "acc_norm": 0.44607843137254904,
349
+ "acc_norm_stderr": 0.03488845451304974
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.47878787878787876,
353
+ "acc_stderr": 0.03900828913737301,
354
+ "acc_norm": 0.47878787878787876,
355
+ "acc_norm_stderr": 0.03900828913737301
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2864137086903305,
359
+ "mc1_stderr": 0.01582614243950235,
360
+ "mc2": 0.453794908688158,
361
+ "mc2_stderr": 0.015317536289389658
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3600944510035419,
365
+ "acc_stderr": 0.016503686720440072,
366
+ "acc_norm": 0.48760330578512395,
367
+ "acc_norm_stderr": 0.017185069732676514
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.0.1",
436
+ "model_sha": "710fbce5dd54e5794f1bcdf4f53d3c0ceeafb405",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.0.2/result_2023-11-08 13:07:51.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3387372013651877,
5
+ "acc_stderr": 0.013830568927974332,
6
+ "acc_norm": 0.4035836177474403,
7
+ "acc_norm_stderr": 0.01433715891426845
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3811989643497311,
11
+ "acc_stderr": 0.00484688692976347,
12
+ "acc_norm": 0.4954192391953794,
13
+ "acc_norm_stderr": 0.004989572002196691
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.49707602339181284,
17
+ "acc_stderr": 0.03834759370936839,
18
+ "acc_norm": 0.49707602339181284,
19
+ "acc_norm_stderr": 0.03834759370936839
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5825242718446602,
23
+ "acc_stderr": 0.048828405482122375,
24
+ "acc_norm": 0.5825242718446602,
25
+ "acc_norm_stderr": 0.048828405482122375
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.508301404853129,
29
+ "acc_stderr": 0.017877498991072,
30
+ "acc_norm": 0.508301404853129,
31
+ "acc_norm_stderr": 0.017877498991072
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.31851851851851853,
35
+ "acc_stderr": 0.040247784019771124,
36
+ "acc_norm": 0.31851851851851853,
37
+ "acc_norm_stderr": 0.040247784019771124
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.32,
41
+ "acc_stderr": 0.046882617226215034,
42
+ "acc_norm": 0.32,
43
+ "acc_norm_stderr": 0.046882617226215034
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.37872340425531914,
47
+ "acc_stderr": 0.03170995606040655,
48
+ "acc_norm": 0.37872340425531914,
49
+ "acc_norm_stderr": 0.03170995606040655
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.43373493975903615,
53
+ "acc_stderr": 0.03858158940685516,
54
+ "acc_norm": 0.43373493975903615,
55
+ "acc_norm_stderr": 0.03858158940685516
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.45980707395498394,
59
+ "acc_stderr": 0.028306190403305696,
60
+ "acc_norm": 0.45980707395498394,
61
+ "acc_norm_stderr": 0.028306190403305696
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.48878923766816146,
65
+ "acc_stderr": 0.033549366530984746,
66
+ "acc_norm": 0.48878923766816146,
67
+ "acc_norm_stderr": 0.033549366530984746
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4351145038167939,
71
+ "acc_stderr": 0.04348208051644858,
72
+ "acc_norm": 0.4351145038167939,
73
+ "acc_norm_stderr": 0.04348208051644858
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.46,
77
+ "acc_stderr": 0.05009082659620332,
78
+ "acc_norm": 0.46,
79
+ "acc_norm_stderr": 0.05009082659620332
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5151515151515151,
83
+ "acc_stderr": 0.03560716516531061,
84
+ "acc_norm": 0.5151515151515151,
85
+ "acc_norm_stderr": 0.03560716516531061
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4689655172413793,
89
+ "acc_stderr": 0.04158632762097828,
90
+ "acc_norm": 0.4689655172413793,
91
+ "acc_norm_stderr": 0.04158632762097828
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.2549019607843137,
95
+ "acc_stderr": 0.043364327079931785,
96
+ "acc_norm": 0.2549019607843137,
97
+ "acc_norm_stderr": 0.043364327079931785
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.4327731092436975,
101
+ "acc_stderr": 0.03218358107742613,
102
+ "acc_norm": 0.4327731092436975,
103
+ "acc_norm_stderr": 0.03218358107742613
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4230769230769231,
107
+ "acc_stderr": 0.025049197876042335,
108
+ "acc_norm": 0.4230769230769231,
109
+ "acc_norm_stderr": 0.025049197876042335
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.55,
113
+ "acc_stderr": 0.049999999999999996,
114
+ "acc_norm": 0.55,
115
+ "acc_norm_stderr": 0.049999999999999996
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.35,
119
+ "acc_stderr": 0.0479372485441102,
120
+ "acc_norm": 0.35,
121
+ "acc_norm_stderr": 0.0479372485441102
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5185185185185185,
125
+ "acc_stderr": 0.04830366024635331,
126
+ "acc_norm": 0.5185185185185185,
127
+ "acc_norm_stderr": 0.04830366024635331
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4187192118226601,
131
+ "acc_stderr": 0.03471192860518468,
132
+ "acc_norm": 0.4187192118226601,
133
+ "acc_norm_stderr": 0.03471192860518468
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.43870967741935485,
137
+ "acc_stderr": 0.028229497320317216,
138
+ "acc_norm": 0.43870967741935485,
139
+ "acc_norm_stderr": 0.028229497320317216
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6923076923076923,
143
+ "acc_stderr": 0.030236389942173078,
144
+ "acc_norm": 0.6923076923076923,
145
+ "acc_norm_stderr": 0.030236389942173078
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.44528301886792454,
149
+ "acc_stderr": 0.030588052974270655,
150
+ "acc_norm": 0.44528301886792454,
151
+ "acc_norm_stderr": 0.030588052974270655
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5,
155
+ "acc_stderr": 0.04789131426105757,
156
+ "acc_norm": 0.5,
157
+ "acc_norm_stderr": 0.04789131426105757
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.3111111111111111,
161
+ "acc_stderr": 0.028226446749683522,
162
+ "acc_norm": 0.3111111111111111,
163
+ "acc_norm_stderr": 0.028226446749683522
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2847682119205298,
167
+ "acc_stderr": 0.03684881521389023,
168
+ "acc_norm": 0.2847682119205298,
169
+ "acc_norm_stderr": 0.03684881521389023
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.6069651741293532,
173
+ "acc_stderr": 0.0345368246603156,
174
+ "acc_norm": 0.6069651741293532,
175
+ "acc_norm_stderr": 0.0345368246603156
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3872832369942196,
179
+ "acc_stderr": 0.03714325906302064,
180
+ "acc_norm": 0.3872832369942196,
181
+ "acc_norm_stderr": 0.03714325906302064
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.373015873015873,
185
+ "acc_stderr": 0.02490699045899257,
186
+ "acc_norm": 0.373015873015873,
187
+ "acc_norm_stderr": 0.02490699045899257
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3888888888888889,
191
+ "acc_stderr": 0.04076663253918567,
192
+ "acc_norm": 0.3888888888888889,
193
+ "acc_norm_stderr": 0.04076663253918567
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.28,
197
+ "acc_stderr": 0.045126085985421276,
198
+ "acc_norm": 0.28,
199
+ "acc_norm_stderr": 0.045126085985421276
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.63,
203
+ "acc_stderr": 0.048523658709391,
204
+ "acc_norm": 0.63,
205
+ "acc_norm_stderr": 0.048523658709391
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5115606936416185,
209
+ "acc_stderr": 0.026911898686377913,
210
+ "acc_norm": 0.5115606936416185,
211
+ "acc_norm_stderr": 0.026911898686377913
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.43558282208588955,
215
+ "acc_stderr": 0.03895632464138937,
216
+ "acc_norm": 0.43558282208588955,
217
+ "acc_norm_stderr": 0.03895632464138937
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4660493827160494,
221
+ "acc_stderr": 0.027756535257347663,
222
+ "acc_norm": 0.4660493827160494,
223
+ "acc_norm_stderr": 0.027756535257347663
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.31,
227
+ "acc_stderr": 0.04648231987117316,
228
+ "acc_norm": 0.31,
229
+ "acc_norm_stderr": 0.04648231987117316
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5544041450777202,
233
+ "acc_stderr": 0.03587014986075659,
234
+ "acc_norm": 0.5544041450777202,
235
+ "acc_norm_stderr": 0.03587014986075659
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.32456140350877194,
239
+ "acc_stderr": 0.04404556157374768,
240
+ "acc_norm": 0.32456140350877194,
241
+ "acc_norm_stderr": 0.04404556157374768
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.5045871559633027,
245
+ "acc_stderr": 0.02143642095552942,
246
+ "acc_norm": 0.5045871559633027,
247
+ "acc_norm_stderr": 0.02143642095552942
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.30158730158730157,
251
+ "acc_stderr": 0.04104947269903394,
252
+ "acc_norm": 0.30158730158730157,
253
+ "acc_norm_stderr": 0.04104947269903394
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.45751633986928103,
257
+ "acc_stderr": 0.02852638345214264,
258
+ "acc_norm": 0.45751633986928103,
259
+ "acc_norm_stderr": 0.02852638345214264
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.45,
263
+ "acc_stderr": 0.05,
264
+ "acc_norm": 0.45,
265
+ "acc_norm_stderr": 0.05
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6694214876033058,
269
+ "acc_stderr": 0.04294340845212094,
270
+ "acc_norm": 0.6694214876033058,
271
+ "acc_norm_stderr": 0.04294340845212094
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40789473684210525,
275
+ "acc_stderr": 0.03999309712777473,
276
+ "acc_norm": 0.40789473684210525,
277
+ "acc_norm_stderr": 0.03999309712777473
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3790849673202614,
281
+ "acc_stderr": 0.019627444748412236,
282
+ "acc_norm": 0.3790849673202614,
283
+ "acc_norm_stderr": 0.019627444748412236
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.35106382978723405,
287
+ "acc_stderr": 0.02847350127296378,
288
+ "acc_norm": 0.35106382978723405,
289
+ "acc_norm_stderr": 0.02847350127296378
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.38392857142857145,
293
+ "acc_stderr": 0.04616143075028547,
294
+ "acc_norm": 0.38392857142857145,
295
+ "acc_norm_stderr": 0.04616143075028547
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.35648148148148145,
299
+ "acc_stderr": 0.032664783315272714,
300
+ "acc_norm": 0.35648148148148145,
301
+ "acc_norm_stderr": 0.032664783315272714
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.21899441340782122,
305
+ "acc_stderr": 0.013831676687303205,
306
+ "acc_norm": 0.21899441340782122,
307
+ "acc_norm_stderr": 0.013831676687303205
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.39,
311
+ "acc_stderr": 0.04902071300001975,
312
+ "acc_norm": 0.39,
313
+ "acc_norm_stderr": 0.04902071300001975
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.57,
317
+ "acc_stderr": 0.049756985195624284,
318
+ "acc_norm": 0.57,
319
+ "acc_norm_stderr": 0.049756985195624284
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3492647058823529,
323
+ "acc_stderr": 0.028959755196824873,
324
+ "acc_norm": 0.3492647058823529,
325
+ "acc_norm_stderr": 0.028959755196824873
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.39591836734693875,
329
+ "acc_stderr": 0.03130802899065686,
330
+ "acc_norm": 0.39591836734693875,
331
+ "acc_norm_stderr": 0.03130802899065686
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6075949367088608,
335
+ "acc_stderr": 0.0317847187456473,
336
+ "acc_norm": 0.6075949367088608,
337
+ "acc_norm_stderr": 0.0317847187456473
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.33833116036505867,
341
+ "acc_stderr": 0.012084265626344204,
342
+ "acc_norm": 0.33833116036505867,
343
+ "acc_norm_stderr": 0.012084265626344204
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.47058823529411764,
347
+ "acc_stderr": 0.03503235296367992,
348
+ "acc_norm": 0.47058823529411764,
349
+ "acc_norm_stderr": 0.03503235296367992
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.5151515151515151,
353
+ "acc_stderr": 0.03902551007374448,
354
+ "acc_norm": 0.5151515151515151,
355
+ "acc_norm_stderr": 0.03902551007374448
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2876376988984088,
359
+ "mc1_stderr": 0.015846315101394812,
360
+ "mc2": 0.4533712341088757,
361
+ "mc2_stderr": 0.015449105919584536
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3825265643447462,
365
+ "acc_stderr": 0.016709165387228806,
366
+ "acc_norm": 0.49586776859504134,
367
+ "acc_norm_stderr": 0.017189767032130824
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.0.2",
436
+ "model_sha": "06668a57b990007d15d178c94aabd162d6af9531",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.0.3/result_2023-11-09 00:04:55.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3370307167235495,
5
+ "acc_stderr": 0.013813476652902272,
6
+ "acc_norm": 0.3993174061433447,
7
+ "acc_norm_stderr": 0.014312094557946707
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3791077474606652,
11
+ "acc_stderr": 0.00484173445350666,
12
+ "acc_norm": 0.4907388966341366,
13
+ "acc_norm_stderr": 0.00498892541052277
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.47368421052631576,
17
+ "acc_stderr": 0.03829509868994727,
18
+ "acc_norm": 0.47368421052631576,
19
+ "acc_norm_stderr": 0.03829509868994727
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5242718446601942,
23
+ "acc_stderr": 0.04944901092973779,
24
+ "acc_norm": 0.5242718446601942,
25
+ "acc_norm_stderr": 0.04944901092973779
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.51213282247765,
29
+ "acc_stderr": 0.017874698667491334,
30
+ "acc_norm": 0.51213282247765,
31
+ "acc_norm_stderr": 0.017874698667491334
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.3333333333333333,
35
+ "acc_stderr": 0.04072314811876837,
36
+ "acc_norm": 0.3333333333333333,
37
+ "acc_norm_stderr": 0.04072314811876837
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.31,
41
+ "acc_stderr": 0.04648231987117316,
42
+ "acc_norm": 0.31,
43
+ "acc_norm_stderr": 0.04648231987117316
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.39148936170212767,
47
+ "acc_stderr": 0.03190701242326812,
48
+ "acc_norm": 0.39148936170212767,
49
+ "acc_norm_stderr": 0.03190701242326812
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.4397590361445783,
53
+ "acc_stderr": 0.03864139923699121,
54
+ "acc_norm": 0.4397590361445783,
55
+ "acc_norm_stderr": 0.03864139923699121
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4662379421221865,
59
+ "acc_stderr": 0.028333277109562786,
60
+ "acc_norm": 0.4662379421221865,
61
+ "acc_norm_stderr": 0.028333277109562786
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.43946188340807174,
65
+ "acc_stderr": 0.03331092511038179,
66
+ "acc_norm": 0.43946188340807174,
67
+ "acc_norm_stderr": 0.03331092511038179
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.40458015267175573,
71
+ "acc_stderr": 0.043046937953806645,
72
+ "acc_norm": 0.40458015267175573,
73
+ "acc_norm_stderr": 0.043046937953806645
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.41,
77
+ "acc_stderr": 0.049431107042371025,
78
+ "acc_norm": 0.41,
79
+ "acc_norm_stderr": 0.049431107042371025
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5252525252525253,
83
+ "acc_stderr": 0.03557806245087314,
84
+ "acc_norm": 0.5252525252525253,
85
+ "acc_norm_stderr": 0.03557806245087314
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4482758620689655,
89
+ "acc_stderr": 0.04144311810878152,
90
+ "acc_norm": 0.4482758620689655,
91
+ "acc_norm_stderr": 0.04144311810878152
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.18627450980392157,
95
+ "acc_stderr": 0.03873958714149352,
96
+ "acc_norm": 0.18627450980392157,
97
+ "acc_norm_stderr": 0.03873958714149352
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.3907563025210084,
101
+ "acc_stderr": 0.031693802357129965,
102
+ "acc_norm": 0.3907563025210084,
103
+ "acc_norm_stderr": 0.031693802357129965
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.3641025641025641,
107
+ "acc_stderr": 0.024396672985094785,
108
+ "acc_norm": 0.3641025641025641,
109
+ "acc_norm_stderr": 0.024396672985094785
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.55,
113
+ "acc_stderr": 0.049999999999999996,
114
+ "acc_norm": 0.55,
115
+ "acc_norm_stderr": 0.049999999999999996
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.3,
119
+ "acc_stderr": 0.046056618647183814,
120
+ "acc_norm": 0.3,
121
+ "acc_norm_stderr": 0.046056618647183814
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5370370370370371,
125
+ "acc_stderr": 0.04820403072760627,
126
+ "acc_norm": 0.5370370370370371,
127
+ "acc_norm_stderr": 0.04820403072760627
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4039408866995074,
131
+ "acc_stderr": 0.0345245390382204,
132
+ "acc_norm": 0.4039408866995074,
133
+ "acc_norm_stderr": 0.0345245390382204
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.44193548387096776,
137
+ "acc_stderr": 0.02825155790684973,
138
+ "acc_norm": 0.44193548387096776,
139
+ "acc_norm_stderr": 0.02825155790684973
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6623931623931624,
143
+ "acc_stderr": 0.030980296992618554,
144
+ "acc_norm": 0.6623931623931624,
145
+ "acc_norm_stderr": 0.030980296992618554
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.41509433962264153,
149
+ "acc_stderr": 0.030325945789286105,
150
+ "acc_norm": 0.41509433962264153,
151
+ "acc_norm_stderr": 0.030325945789286105
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.43636363636363634,
155
+ "acc_stderr": 0.04750185058907297,
156
+ "acc_norm": 0.43636363636363634,
157
+ "acc_norm_stderr": 0.04750185058907297
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.31851851851851853,
161
+ "acc_stderr": 0.028406533090608466,
162
+ "acc_norm": 0.31851851851851853,
163
+ "acc_norm_stderr": 0.028406533090608466
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2913907284768212,
167
+ "acc_stderr": 0.037101857261199946,
168
+ "acc_norm": 0.2913907284768212,
169
+ "acc_norm_stderr": 0.037101857261199946
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.6218905472636815,
173
+ "acc_stderr": 0.034288678487786564,
174
+ "acc_norm": 0.6218905472636815,
175
+ "acc_norm_stderr": 0.034288678487786564
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3930635838150289,
179
+ "acc_stderr": 0.03724249595817729,
180
+ "acc_norm": 0.3930635838150289,
181
+ "acc_norm_stderr": 0.03724249595817729
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.3544973544973545,
185
+ "acc_stderr": 0.024636830602842,
186
+ "acc_norm": 0.3544973544973545,
187
+ "acc_norm_stderr": 0.024636830602842
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3055555555555556,
191
+ "acc_stderr": 0.03852084696008534,
192
+ "acc_norm": 0.3055555555555556,
193
+ "acc_norm_stderr": 0.03852084696008534
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.26,
197
+ "acc_stderr": 0.0440844002276808,
198
+ "acc_norm": 0.26,
199
+ "acc_norm_stderr": 0.0440844002276808
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.63,
203
+ "acc_stderr": 0.048523658709391,
204
+ "acc_norm": 0.63,
205
+ "acc_norm_stderr": 0.048523658709391
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.48265895953757226,
209
+ "acc_stderr": 0.026902900458666647,
210
+ "acc_norm": 0.48265895953757226,
211
+ "acc_norm_stderr": 0.026902900458666647
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4049079754601227,
215
+ "acc_stderr": 0.03856672163548914,
216
+ "acc_norm": 0.4049079754601227,
217
+ "acc_norm_stderr": 0.03856672163548914
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4845679012345679,
221
+ "acc_stderr": 0.027807490044276198,
222
+ "acc_norm": 0.4845679012345679,
223
+ "acc_norm_stderr": 0.027807490044276198
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.29,
227
+ "acc_stderr": 0.045604802157206845,
228
+ "acc_norm": 0.29,
229
+ "acc_norm_stderr": 0.045604802157206845
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.533678756476684,
233
+ "acc_stderr": 0.03600244069867178,
234
+ "acc_norm": 0.533678756476684,
235
+ "acc_norm_stderr": 0.03600244069867178
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.30701754385964913,
239
+ "acc_stderr": 0.0433913832257986,
240
+ "acc_norm": 0.30701754385964913,
241
+ "acc_norm_stderr": 0.0433913832257986
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.46055045871559636,
245
+ "acc_stderr": 0.021370494609995093,
246
+ "acc_norm": 0.46055045871559636,
247
+ "acc_norm_stderr": 0.021370494609995093
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.23809523809523808,
251
+ "acc_stderr": 0.03809523809523811,
252
+ "acc_norm": 0.23809523809523808,
253
+ "acc_norm_stderr": 0.03809523809523811
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4411764705882353,
257
+ "acc_stderr": 0.028431095444176643,
258
+ "acc_norm": 0.4411764705882353,
259
+ "acc_norm_stderr": 0.028431095444176643
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.41,
263
+ "acc_stderr": 0.049431107042371025,
264
+ "acc_norm": 0.41,
265
+ "acc_norm_stderr": 0.049431107042371025
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6528925619834711,
269
+ "acc_stderr": 0.04345724570292534,
270
+ "acc_norm": 0.6528925619834711,
271
+ "acc_norm_stderr": 0.04345724570292534
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40789473684210525,
275
+ "acc_stderr": 0.039993097127774734,
276
+ "acc_norm": 0.40789473684210525,
277
+ "acc_norm_stderr": 0.039993097127774734
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3790849673202614,
281
+ "acc_stderr": 0.019627444748412243,
282
+ "acc_norm": 0.3790849673202614,
283
+ "acc_norm_stderr": 0.019627444748412243
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3333333333333333,
287
+ "acc_stderr": 0.028121636040639875,
288
+ "acc_norm": 0.3333333333333333,
289
+ "acc_norm_stderr": 0.028121636040639875
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.2767857142857143,
293
+ "acc_stderr": 0.042466243366976256,
294
+ "acc_norm": 0.2767857142857143,
295
+ "acc_norm_stderr": 0.042466243366976256
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3611111111111111,
299
+ "acc_stderr": 0.03275773486100999,
300
+ "acc_norm": 0.3611111111111111,
301
+ "acc_norm_stderr": 0.03275773486100999
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.24692737430167597,
305
+ "acc_stderr": 0.014422292204808852,
306
+ "acc_norm": 0.24692737430167597,
307
+ "acc_norm_stderr": 0.014422292204808852
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.39,
311
+ "acc_stderr": 0.04902071300001975,
312
+ "acc_norm": 0.39,
313
+ "acc_norm_stderr": 0.04902071300001975
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.54,
317
+ "acc_stderr": 0.05009082659620332,
318
+ "acc_norm": 0.54,
319
+ "acc_norm_stderr": 0.05009082659620332
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3639705882352941,
323
+ "acc_stderr": 0.029227192460032025,
324
+ "acc_norm": 0.3639705882352941,
325
+ "acc_norm_stderr": 0.029227192460032025
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4122448979591837,
329
+ "acc_stderr": 0.03151236044674281,
330
+ "acc_norm": 0.4122448979591837,
331
+ "acc_norm_stderr": 0.03151236044674281
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5569620253164557,
335
+ "acc_stderr": 0.03233532777533484,
336
+ "acc_norm": 0.5569620253164557,
337
+ "acc_norm_stderr": 0.03233532777533484
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3213820078226858,
341
+ "acc_stderr": 0.01192758135226508,
342
+ "acc_norm": 0.3213820078226858,
343
+ "acc_norm_stderr": 0.01192758135226508
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4117647058823529,
347
+ "acc_stderr": 0.03454236585380611,
348
+ "acc_norm": 0.4117647058823529,
349
+ "acc_norm_stderr": 0.03454236585380611
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.44242424242424244,
353
+ "acc_stderr": 0.038783721137112745,
354
+ "acc_norm": 0.44242424242424244,
355
+ "acc_norm_stderr": 0.038783721137112745
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2802937576499388,
359
+ "mc1_stderr": 0.015723139524608753,
360
+ "mc2": 0.44405801770483816,
361
+ "mc2_stderr": 0.015315267499738446
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.32585596221959856,
365
+ "acc_stderr": 0.016114023894800333,
366
+ "acc_norm": 0.46162927981109797,
367
+ "acc_norm_stderr": 0.017139660221845564
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.0.3",
436
+ "model_sha": "5d368e894e2091bd003ae65710009016b7bafaa0",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.0/result_2023-11-07 08:10:27.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.35409556313993173,
5
+ "acc_stderr": 0.013975454122756562,
6
+ "acc_norm": 0.4087030716723549,
7
+ "acc_norm_stderr": 0.014365750345427
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3875721967735511,
11
+ "acc_stderr": 0.004862003566798545,
12
+ "acc_norm": 0.504282015534754,
13
+ "acc_norm_stderr": 0.004989598426249547
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4853801169590643,
17
+ "acc_stderr": 0.038331852752130205,
18
+ "acc_norm": 0.4853801169590643,
19
+ "acc_norm_stderr": 0.038331852752130205
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5533980582524272,
23
+ "acc_stderr": 0.04922424153458933,
24
+ "acc_norm": 0.5533980582524272,
25
+ "acc_norm_stderr": 0.04922424153458933
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.51213282247765,
29
+ "acc_stderr": 0.01787469866749133,
30
+ "acc_norm": 0.51213282247765,
31
+ "acc_norm_stderr": 0.01787469866749133
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4444444444444444,
35
+ "acc_stderr": 0.04292596718256981,
36
+ "acc_norm": 0.4444444444444444,
37
+ "acc_norm_stderr": 0.04292596718256981
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.28,
41
+ "acc_stderr": 0.045126085985421255,
42
+ "acc_norm": 0.28,
43
+ "acc_norm_stderr": 0.045126085985421255
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.41702127659574467,
47
+ "acc_stderr": 0.03223276266711712,
48
+ "acc_norm": 0.41702127659574467,
49
+ "acc_norm_stderr": 0.03223276266711712
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.43373493975903615,
53
+ "acc_stderr": 0.03858158940685516,
54
+ "acc_norm": 0.43373493975903615,
55
+ "acc_norm_stderr": 0.03858158940685516
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.49517684887459806,
59
+ "acc_stderr": 0.028396770444111288,
60
+ "acc_norm": 0.49517684887459806,
61
+ "acc_norm_stderr": 0.028396770444111288
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.48878923766816146,
65
+ "acc_stderr": 0.033549366530984746,
66
+ "acc_norm": 0.48878923766816146,
67
+ "acc_norm_stderr": 0.033549366530984746
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4122137404580153,
71
+ "acc_stderr": 0.04317171194870254,
72
+ "acc_norm": 0.4122137404580153,
73
+ "acc_norm_stderr": 0.04317171194870254
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.44,
77
+ "acc_stderr": 0.04988876515698589,
78
+ "acc_norm": 0.44,
79
+ "acc_norm_stderr": 0.04988876515698589
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5454545454545454,
83
+ "acc_stderr": 0.03547601494006937,
84
+ "acc_norm": 0.5454545454545454,
85
+ "acc_norm_stderr": 0.03547601494006937
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4482758620689655,
89
+ "acc_stderr": 0.04144311810878152,
90
+ "acc_norm": 0.4482758620689655,
91
+ "acc_norm_stderr": 0.04144311810878152
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.23529411764705882,
95
+ "acc_stderr": 0.04220773659171452,
96
+ "acc_norm": 0.23529411764705882,
97
+ "acc_norm_stderr": 0.04220773659171452
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.40756302521008403,
101
+ "acc_stderr": 0.03191863374478465,
102
+ "acc_norm": 0.40756302521008403,
103
+ "acc_norm_stderr": 0.03191863374478465
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4307692307692308,
107
+ "acc_stderr": 0.02510682066053975,
108
+ "acc_norm": 0.4307692307692308,
109
+ "acc_norm_stderr": 0.02510682066053975
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.66,
113
+ "acc_stderr": 0.04760952285695237,
114
+ "acc_norm": 0.66,
115
+ "acc_norm_stderr": 0.04760952285695237
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.38,
119
+ "acc_stderr": 0.04878317312145632,
120
+ "acc_norm": 0.38,
121
+ "acc_norm_stderr": 0.04878317312145632
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.4722222222222222,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.4722222222222222,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.43349753694581283,
131
+ "acc_stderr": 0.03486731727419872,
132
+ "acc_norm": 0.43349753694581283,
133
+ "acc_norm_stderr": 0.03486731727419872
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.45806451612903226,
137
+ "acc_stderr": 0.02834378725054062,
138
+ "acc_norm": 0.45806451612903226,
139
+ "acc_norm_stderr": 0.02834378725054062
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6709401709401709,
143
+ "acc_stderr": 0.030782321577688173,
144
+ "acc_norm": 0.6709401709401709,
145
+ "acc_norm_stderr": 0.030782321577688173
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.41509433962264153,
149
+ "acc_stderr": 0.03032594578928611,
150
+ "acc_norm": 0.41509433962264153,
151
+ "acc_norm_stderr": 0.03032594578928611
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.43636363636363634,
155
+ "acc_stderr": 0.04750185058907297,
156
+ "acc_norm": 0.43636363636363634,
157
+ "acc_norm_stderr": 0.04750185058907297
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.26666666666666666,
161
+ "acc_stderr": 0.02696242432507382,
162
+ "acc_norm": 0.26666666666666666,
163
+ "acc_norm_stderr": 0.02696242432507382
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.24503311258278146,
167
+ "acc_stderr": 0.03511807571804723,
168
+ "acc_norm": 0.24503311258278146,
169
+ "acc_norm_stderr": 0.03511807571804723
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5870646766169154,
173
+ "acc_stderr": 0.03481520803367348,
174
+ "acc_norm": 0.5870646766169154,
175
+ "acc_norm_stderr": 0.03481520803367348
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3930635838150289,
179
+ "acc_stderr": 0.03724249595817729,
180
+ "acc_norm": 0.3930635838150289,
181
+ "acc_norm_stderr": 0.03724249595817729
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.35714285714285715,
185
+ "acc_stderr": 0.024677862841332786,
186
+ "acc_norm": 0.35714285714285715,
187
+ "acc_norm_stderr": 0.024677862841332786
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.375,
191
+ "acc_stderr": 0.04048439222695598,
192
+ "acc_norm": 0.375,
193
+ "acc_norm_stderr": 0.04048439222695598
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.3,
197
+ "acc_stderr": 0.046056618647183814,
198
+ "acc_norm": 0.3,
199
+ "acc_norm_stderr": 0.046056618647183814
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.57,
203
+ "acc_stderr": 0.04975698519562427,
204
+ "acc_norm": 0.57,
205
+ "acc_norm_stderr": 0.04975698519562427
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.47398843930635837,
209
+ "acc_stderr": 0.026882643434022895,
210
+ "acc_norm": 0.47398843930635837,
211
+ "acc_norm_stderr": 0.026882643434022895
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.43558282208588955,
215
+ "acc_stderr": 0.03895632464138937,
216
+ "acc_norm": 0.43558282208588955,
217
+ "acc_norm_stderr": 0.03895632464138937
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4567901234567901,
221
+ "acc_stderr": 0.027716661650194038,
222
+ "acc_norm": 0.4567901234567901,
223
+ "acc_norm_stderr": 0.027716661650194038
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.32,
227
+ "acc_stderr": 0.046882617226215034,
228
+ "acc_norm": 0.32,
229
+ "acc_norm_stderr": 0.046882617226215034
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5233160621761658,
233
+ "acc_stderr": 0.03604513672442202,
234
+ "acc_norm": 0.5233160621761658,
235
+ "acc_norm_stderr": 0.03604513672442202
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.3333333333333333,
239
+ "acc_stderr": 0.044346007015849245,
240
+ "acc_norm": 0.3333333333333333,
241
+ "acc_norm_stderr": 0.044346007015849245
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.48256880733944957,
245
+ "acc_stderr": 0.02142429187185315,
246
+ "acc_norm": 0.48256880733944957,
247
+ "acc_norm_stderr": 0.02142429187185315
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.23809523809523808,
251
+ "acc_stderr": 0.03809523809523811,
252
+ "acc_norm": 0.23809523809523808,
253
+ "acc_norm_stderr": 0.03809523809523811
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.434640522875817,
257
+ "acc_stderr": 0.028384256704883037,
258
+ "acc_norm": 0.434640522875817,
259
+ "acc_norm_stderr": 0.028384256704883037
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.41,
263
+ "acc_stderr": 0.049431107042371025,
264
+ "acc_norm": 0.41,
265
+ "acc_norm_stderr": 0.049431107042371025
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6528925619834711,
269
+ "acc_stderr": 0.04345724570292534,
270
+ "acc_norm": 0.6528925619834711,
271
+ "acc_norm_stderr": 0.04345724570292534
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3881578947368421,
275
+ "acc_stderr": 0.03965842097512744,
276
+ "acc_norm": 0.3881578947368421,
277
+ "acc_norm_stderr": 0.03965842097512744
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.37254901960784315,
281
+ "acc_stderr": 0.019559646809215934,
282
+ "acc_norm": 0.37254901960784315,
283
+ "acc_norm_stderr": 0.019559646809215934
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.33687943262411346,
287
+ "acc_stderr": 0.02819553487396673,
288
+ "acc_norm": 0.33687943262411346,
289
+ "acc_norm_stderr": 0.02819553487396673
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.32142857142857145,
293
+ "acc_stderr": 0.04432804055291518,
294
+ "acc_norm": 0.32142857142857145,
295
+ "acc_norm_stderr": 0.04432804055291518
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.375,
299
+ "acc_stderr": 0.033016908987210894,
300
+ "acc_norm": 0.375,
301
+ "acc_norm_stderr": 0.033016908987210894
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.24134078212290502,
305
+ "acc_stderr": 0.014310999547961443,
306
+ "acc_norm": 0.24134078212290502,
307
+ "acc_norm_stderr": 0.014310999547961443
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.37,
311
+ "acc_stderr": 0.04852365870939099,
312
+ "acc_norm": 0.37,
313
+ "acc_norm_stderr": 0.04852365870939099
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.6,
317
+ "acc_stderr": 0.049236596391733084,
318
+ "acc_norm": 0.6,
319
+ "acc_norm_stderr": 0.049236596391733084
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3382352941176471,
323
+ "acc_stderr": 0.028739328513983576,
324
+ "acc_norm": 0.3382352941176471,
325
+ "acc_norm_stderr": 0.028739328513983576
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.39183673469387753,
329
+ "acc_stderr": 0.03125127591089165,
330
+ "acc_norm": 0.39183673469387753,
331
+ "acc_norm_stderr": 0.03125127591089165
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5780590717299579,
335
+ "acc_stderr": 0.032148146302403695,
336
+ "acc_norm": 0.5780590717299579,
337
+ "acc_norm_stderr": 0.032148146302403695
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.32985658409387225,
341
+ "acc_stderr": 0.012008129938540472,
342
+ "acc_norm": 0.32985658409387225,
343
+ "acc_norm_stderr": 0.012008129938540472
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4215686274509804,
347
+ "acc_stderr": 0.03465868196380758,
348
+ "acc_norm": 0.4215686274509804,
349
+ "acc_norm_stderr": 0.03465868196380758
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.4727272727272727,
353
+ "acc_stderr": 0.03898531605579419,
354
+ "acc_norm": 0.4727272727272727,
355
+ "acc_norm_stderr": 0.03898531605579419
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2802937576499388,
359
+ "mc1_stderr": 0.015723139524608742,
360
+ "mc2": 0.4447858809482175,
361
+ "mc2_stderr": 0.015211057250300537
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3860684769775679,
365
+ "acc_stderr": 0.016738130760321743,
366
+ "acc_norm": 0.5100354191263282,
367
+ "acc_norm_stderr": 0.017186891286894067
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.0",
436
+ "model_sha": "f5bfb9dc4f4dd8b64d45c9a158e3982959b18035",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.1.0/result_2023-11-13 23:21:20.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.36177474402730375,
5
+ "acc_stderr": 0.014041957945038078,
6
+ "acc_norm": 0.4232081911262799,
7
+ "acc_norm_stderr": 0.01443803622084802
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.37223660625373434,
11
+ "acc_stderr": 0.004824130528590597,
12
+ "acc_norm": 0.47610037841067515,
13
+ "acc_norm_stderr": 0.004984077906216103
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4327485380116959,
17
+ "acc_stderr": 0.03799978644370607,
18
+ "acc_norm": 0.4327485380116959,
19
+ "acc_norm_stderr": 0.03799978644370607
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5533980582524272,
23
+ "acc_stderr": 0.04922424153458934,
24
+ "acc_norm": 0.5533980582524272,
25
+ "acc_norm_stderr": 0.04922424153458934
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4840357598978289,
29
+ "acc_stderr": 0.017870847506081727,
30
+ "acc_norm": 0.4840357598978289,
31
+ "acc_norm_stderr": 0.017870847506081727
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4074074074074074,
35
+ "acc_stderr": 0.04244633238353229,
36
+ "acc_norm": 0.4074074074074074,
37
+ "acc_norm_stderr": 0.04244633238353229
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.29,
41
+ "acc_stderr": 0.04560480215720684,
42
+ "acc_norm": 0.29,
43
+ "acc_norm_stderr": 0.04560480215720684
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.35319148936170214,
47
+ "acc_stderr": 0.031245325202761926,
48
+ "acc_norm": 0.35319148936170214,
49
+ "acc_norm_stderr": 0.031245325202761926
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39759036144578314,
53
+ "acc_stderr": 0.038099730845402184,
54
+ "acc_norm": 0.39759036144578314,
55
+ "acc_norm_stderr": 0.038099730845402184
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4212218649517685,
59
+ "acc_stderr": 0.028043399858210628,
60
+ "acc_norm": 0.4212218649517685,
61
+ "acc_norm_stderr": 0.028043399858210628
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.40358744394618834,
65
+ "acc_stderr": 0.03292802819330315,
66
+ "acc_norm": 0.40358744394618834,
67
+ "acc_norm_stderr": 0.03292802819330315
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.44274809160305345,
71
+ "acc_stderr": 0.04356447202665069,
72
+ "acc_norm": 0.44274809160305345,
73
+ "acc_norm_stderr": 0.04356447202665069
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.37,
77
+ "acc_stderr": 0.04852365870939099,
78
+ "acc_norm": 0.37,
79
+ "acc_norm_stderr": 0.04852365870939099
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5353535353535354,
83
+ "acc_stderr": 0.03553436368828063,
84
+ "acc_norm": 0.5353535353535354,
85
+ "acc_norm_stderr": 0.03553436368828063
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.42758620689655175,
89
+ "acc_stderr": 0.041227371113703316,
90
+ "acc_norm": 0.42758620689655175,
91
+ "acc_norm_stderr": 0.041227371113703316
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.23529411764705882,
95
+ "acc_stderr": 0.042207736591714534,
96
+ "acc_norm": 0.23529411764705882,
97
+ "acc_norm_stderr": 0.042207736591714534
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.3907563025210084,
101
+ "acc_stderr": 0.031693802357129965,
102
+ "acc_norm": 0.3907563025210084,
103
+ "acc_norm_stderr": 0.031693802357129965
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4025641025641026,
107
+ "acc_stderr": 0.024864995159767762,
108
+ "acc_norm": 0.4025641025641026,
109
+ "acc_norm_stderr": 0.024864995159767762
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.56,
113
+ "acc_stderr": 0.04988876515698589,
114
+ "acc_norm": 0.56,
115
+ "acc_norm_stderr": 0.04988876515698589
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.26,
119
+ "acc_stderr": 0.0440844002276808,
120
+ "acc_norm": 0.26,
121
+ "acc_norm_stderr": 0.0440844002276808
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5,
125
+ "acc_stderr": 0.04833682445228318,
126
+ "acc_norm": 0.5,
127
+ "acc_norm_stderr": 0.04833682445228318
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4236453201970443,
131
+ "acc_stderr": 0.034767257476490364,
132
+ "acc_norm": 0.4236453201970443,
133
+ "acc_norm_stderr": 0.034767257476490364
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.44193548387096776,
137
+ "acc_stderr": 0.02825155790684974,
138
+ "acc_norm": 0.44193548387096776,
139
+ "acc_norm_stderr": 0.02825155790684974
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6111111111111112,
143
+ "acc_stderr": 0.031937057262002924,
144
+ "acc_norm": 0.6111111111111112,
145
+ "acc_norm_stderr": 0.031937057262002924
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4226415094339623,
149
+ "acc_stderr": 0.030402331445769537,
150
+ "acc_norm": 0.4226415094339623,
151
+ "acc_norm_stderr": 0.030402331445769537
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4909090909090909,
155
+ "acc_stderr": 0.04788339768702861,
156
+ "acc_norm": 0.4909090909090909,
157
+ "acc_norm_stderr": 0.04788339768702861
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.26666666666666666,
161
+ "acc_stderr": 0.026962424325073828,
162
+ "acc_norm": 0.26666666666666666,
163
+ "acc_norm_stderr": 0.026962424325073828
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.32450331125827814,
167
+ "acc_stderr": 0.038227469376587525,
168
+ "acc_norm": 0.32450331125827814,
169
+ "acc_norm_stderr": 0.038227469376587525
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5572139303482587,
173
+ "acc_stderr": 0.03512310964123935,
174
+ "acc_norm": 0.5572139303482587,
175
+ "acc_norm_stderr": 0.03512310964123935
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3699421965317919,
179
+ "acc_stderr": 0.0368122963339432,
180
+ "acc_norm": 0.3699421965317919,
181
+ "acc_norm_stderr": 0.0368122963339432
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.3439153439153439,
185
+ "acc_stderr": 0.024464426625596433,
186
+ "acc_norm": 0.3439153439153439,
187
+ "acc_norm_stderr": 0.024464426625596433
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3680555555555556,
191
+ "acc_stderr": 0.04032999053960719,
192
+ "acc_norm": 0.3680555555555556,
193
+ "acc_norm_stderr": 0.04032999053960719
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.37,
197
+ "acc_stderr": 0.04852365870939099,
198
+ "acc_norm": 0.37,
199
+ "acc_norm_stderr": 0.04852365870939099
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.57,
203
+ "acc_stderr": 0.04975698519562426,
204
+ "acc_norm": 0.57,
205
+ "acc_norm_stderr": 0.04975698519562426
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.44508670520231214,
209
+ "acc_stderr": 0.026756255129663765,
210
+ "acc_norm": 0.44508670520231214,
211
+ "acc_norm_stderr": 0.026756255129663765
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.44171779141104295,
215
+ "acc_stderr": 0.03901591825836184,
216
+ "acc_norm": 0.44171779141104295,
217
+ "acc_norm_stderr": 0.03901591825836184
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.41975308641975306,
221
+ "acc_stderr": 0.027460099557005138,
222
+ "acc_norm": 0.41975308641975306,
223
+ "acc_norm_stderr": 0.027460099557005138
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.33,
227
+ "acc_stderr": 0.04725815626252605,
228
+ "acc_norm": 0.33,
229
+ "acc_norm_stderr": 0.04725815626252605
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.47150259067357514,
233
+ "acc_stderr": 0.03602573571288441,
234
+ "acc_norm": 0.47150259067357514,
235
+ "acc_norm_stderr": 0.03602573571288441
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2631578947368421,
239
+ "acc_stderr": 0.041424397194893596,
240
+ "acc_norm": 0.2631578947368421,
241
+ "acc_norm_stderr": 0.041424397194893596
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.46422018348623856,
245
+ "acc_stderr": 0.0213823647757019,
246
+ "acc_norm": 0.46422018348623856,
247
+ "acc_norm_stderr": 0.0213823647757019
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.29365079365079366,
251
+ "acc_stderr": 0.04073524322147125,
252
+ "acc_norm": 0.29365079365079366,
253
+ "acc_norm_stderr": 0.04073524322147125
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4444444444444444,
257
+ "acc_stderr": 0.02845263998508801,
258
+ "acc_norm": 0.4444444444444444,
259
+ "acc_norm_stderr": 0.02845263998508801
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.39,
263
+ "acc_stderr": 0.04902071300001975,
264
+ "acc_norm": 0.39,
265
+ "acc_norm_stderr": 0.04902071300001975
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6198347107438017,
269
+ "acc_stderr": 0.04431324501968431,
270
+ "acc_norm": 0.6198347107438017,
271
+ "acc_norm_stderr": 0.04431324501968431
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3881578947368421,
275
+ "acc_stderr": 0.03965842097512744,
276
+ "acc_norm": 0.3881578947368421,
277
+ "acc_norm_stderr": 0.03965842097512744
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.35294117647058826,
281
+ "acc_stderr": 0.01933314202079706,
282
+ "acc_norm": 0.35294117647058826,
283
+ "acc_norm_stderr": 0.01933314202079706
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.2978723404255319,
287
+ "acc_stderr": 0.027281608344469417,
288
+ "acc_norm": 0.2978723404255319,
289
+ "acc_norm_stderr": 0.027281608344469417
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.25,
293
+ "acc_stderr": 0.04109974682633932,
294
+ "acc_norm": 0.25,
295
+ "acc_norm_stderr": 0.04109974682633932
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.375,
299
+ "acc_stderr": 0.033016908987210894,
300
+ "acc_norm": 0.375,
301
+ "acc_norm_stderr": 0.033016908987210894
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.24134078212290502,
305
+ "acc_stderr": 0.014310999547961464,
306
+ "acc_norm": 0.24134078212290502,
307
+ "acc_norm_stderr": 0.014310999547961464
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.41,
311
+ "acc_stderr": 0.049431107042371025,
312
+ "acc_norm": 0.41,
313
+ "acc_norm_stderr": 0.049431107042371025
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.44,
317
+ "acc_stderr": 0.04988876515698589,
318
+ "acc_norm": 0.44,
319
+ "acc_norm_stderr": 0.04988876515698589
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.39338235294117646,
323
+ "acc_stderr": 0.029674288281311183,
324
+ "acc_norm": 0.39338235294117646,
325
+ "acc_norm_stderr": 0.029674288281311183
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.32653061224489793,
329
+ "acc_stderr": 0.030021056238440303,
330
+ "acc_norm": 0.32653061224489793,
331
+ "acc_norm_stderr": 0.030021056238440303
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5274261603375527,
335
+ "acc_stderr": 0.03249822718301303,
336
+ "acc_norm": 0.5274261603375527,
337
+ "acc_norm_stderr": 0.03249822718301303
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.29791395045632335,
341
+ "acc_stderr": 0.01168071734040005,
342
+ "acc_norm": 0.29791395045632335,
343
+ "acc_norm_stderr": 0.01168071734040005
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4215686274509804,
347
+ "acc_stderr": 0.03465868196380757,
348
+ "acc_norm": 0.4215686274509804,
349
+ "acc_norm_stderr": 0.03465868196380757
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.45454545454545453,
353
+ "acc_stderr": 0.03888176921674099,
354
+ "acc_norm": 0.45454545454545453,
355
+ "acc_norm_stderr": 0.03888176921674099
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2778457772337821,
359
+ "mc1_stderr": 0.01568092936402462,
360
+ "mc2": 0.4553516695896828,
361
+ "mc2_stderr": 0.01619950826163877
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3069657615112161,
365
+ "acc_stderr": 0.015857588095362814,
366
+ "acc_norm": 0.34710743801652894,
367
+ "acc_norm_stderr": 0.016366945603281273
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.1.0",
436
+ "model_sha": "7cf13a6ab9a4f0231b168a8102d784fc6c22510a",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.2.0/result_2023-11-14 04:32:36.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.33532423208191126,
5
+ "acc_stderr": 0.013796182947785564,
6
+ "acc_norm": 0.38139931740614336,
7
+ "acc_norm_stderr": 0.014194389086685251
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3736307508464449,
11
+ "acc_stderr": 0.00482778628907485,
12
+ "acc_norm": 0.4671380203146783,
13
+ "acc_norm_stderr": 0.004978992721242828
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.39766081871345027,
17
+ "acc_stderr": 0.0375363895576169,
18
+ "acc_norm": 0.39766081871345027,
19
+ "acc_norm_stderr": 0.0375363895576169
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5825242718446602,
23
+ "acc_stderr": 0.048828405482122375,
24
+ "acc_norm": 0.5825242718446602,
25
+ "acc_norm_stderr": 0.048828405482122375
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.42911877394636017,
29
+ "acc_stderr": 0.017699388483126795,
30
+ "acc_norm": 0.42911877394636017,
31
+ "acc_norm_stderr": 0.017699388483126795
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.37777777777777777,
35
+ "acc_stderr": 0.04188307537595853,
36
+ "acc_norm": 0.37777777777777777,
37
+ "acc_norm_stderr": 0.04188307537595853
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.35,
41
+ "acc_stderr": 0.047937248544110196,
42
+ "acc_norm": 0.35,
43
+ "acc_norm_stderr": 0.047937248544110196
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.35319148936170214,
47
+ "acc_stderr": 0.031245325202761926,
48
+ "acc_norm": 0.35319148936170214,
49
+ "acc_norm_stderr": 0.031245325202761926
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39156626506024095,
53
+ "acc_stderr": 0.03799857454479636,
54
+ "acc_norm": 0.39156626506024095,
55
+ "acc_norm_stderr": 0.03799857454479636
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.42765273311897106,
59
+ "acc_stderr": 0.02809924077580955,
60
+ "acc_norm": 0.42765273311897106,
61
+ "acc_norm_stderr": 0.02809924077580955
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.37668161434977576,
65
+ "acc_stderr": 0.03252113489929187,
66
+ "acc_norm": 0.37668161434977576,
67
+ "acc_norm_stderr": 0.03252113489929187
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.45038167938931295,
71
+ "acc_stderr": 0.04363643698524779,
72
+ "acc_norm": 0.45038167938931295,
73
+ "acc_norm_stderr": 0.04363643698524779
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.37,
77
+ "acc_stderr": 0.048523658709391,
78
+ "acc_norm": 0.37,
79
+ "acc_norm_stderr": 0.048523658709391
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5303030303030303,
83
+ "acc_stderr": 0.0355580405176393,
84
+ "acc_norm": 0.5303030303030303,
85
+ "acc_norm_stderr": 0.0355580405176393
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.42758620689655175,
89
+ "acc_stderr": 0.04122737111370332,
90
+ "acc_norm": 0.42758620689655175,
91
+ "acc_norm_stderr": 0.04122737111370332
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.20588235294117646,
95
+ "acc_stderr": 0.040233822736177476,
96
+ "acc_norm": 0.20588235294117646,
97
+ "acc_norm_stderr": 0.040233822736177476
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.4117647058823529,
101
+ "acc_stderr": 0.031968769891957786,
102
+ "acc_norm": 0.4117647058823529,
103
+ "acc_norm_stderr": 0.031968769891957786
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.43846153846153846,
107
+ "acc_stderr": 0.02515826601686856,
108
+ "acc_norm": 0.43846153846153846,
109
+ "acc_norm_stderr": 0.02515826601686856
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.53,
113
+ "acc_stderr": 0.05016135580465919,
114
+ "acc_norm": 0.53,
115
+ "acc_norm_stderr": 0.05016135580465919
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.24,
119
+ "acc_stderr": 0.042923469599092816,
120
+ "acc_norm": 0.24,
121
+ "acc_norm_stderr": 0.042923469599092816
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5648148148148148,
125
+ "acc_stderr": 0.04792898170907061,
126
+ "acc_norm": 0.5648148148148148,
127
+ "acc_norm_stderr": 0.04792898170907061
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.37438423645320196,
131
+ "acc_stderr": 0.03405155380561952,
132
+ "acc_norm": 0.37438423645320196,
133
+ "acc_norm_stderr": 0.03405155380561952
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.432258064516129,
137
+ "acc_stderr": 0.028181739720019416,
138
+ "acc_norm": 0.432258064516129,
139
+ "acc_norm_stderr": 0.028181739720019416
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6196581196581197,
143
+ "acc_stderr": 0.031804252043841,
144
+ "acc_norm": 0.6196581196581197,
145
+ "acc_norm_stderr": 0.031804252043841
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.41509433962264153,
149
+ "acc_stderr": 0.030325945789286102,
150
+ "acc_norm": 0.41509433962264153,
151
+ "acc_norm_stderr": 0.030325945789286102
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5727272727272728,
155
+ "acc_stderr": 0.047381987035454834,
156
+ "acc_norm": 0.5727272727272728,
157
+ "acc_norm_stderr": 0.047381987035454834
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.3037037037037037,
161
+ "acc_stderr": 0.028037929969114975,
162
+ "acc_norm": 0.3037037037037037,
163
+ "acc_norm_stderr": 0.028037929969114975
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2781456953642384,
167
+ "acc_stderr": 0.036586032627637426,
168
+ "acc_norm": 0.2781456953642384,
169
+ "acc_norm_stderr": 0.036586032627637426
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5572139303482587,
173
+ "acc_stderr": 0.03512310964123937,
174
+ "acc_norm": 0.5572139303482587,
175
+ "acc_norm_stderr": 0.03512310964123937
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3468208092485549,
179
+ "acc_stderr": 0.036291466701596636,
180
+ "acc_norm": 0.3468208092485549,
181
+ "acc_norm_stderr": 0.036291466701596636
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.373015873015873,
185
+ "acc_stderr": 0.02490699045899257,
186
+ "acc_norm": 0.373015873015873,
187
+ "acc_norm_stderr": 0.02490699045899257
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3472222222222222,
191
+ "acc_stderr": 0.039812405437178615,
192
+ "acc_norm": 0.3472222222222222,
193
+ "acc_norm_stderr": 0.039812405437178615
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.33,
197
+ "acc_stderr": 0.047258156262526045,
198
+ "acc_norm": 0.33,
199
+ "acc_norm_stderr": 0.047258156262526045
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.58,
203
+ "acc_stderr": 0.04960449637488585,
204
+ "acc_norm": 0.58,
205
+ "acc_norm_stderr": 0.04960449637488585
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.4797687861271676,
209
+ "acc_stderr": 0.026897049996382875,
210
+ "acc_norm": 0.4797687861271676,
211
+ "acc_norm_stderr": 0.026897049996382875
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4171779141104294,
215
+ "acc_stderr": 0.0387410285981808,
216
+ "acc_norm": 0.4171779141104294,
217
+ "acc_norm_stderr": 0.0387410285981808
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.38271604938271603,
221
+ "acc_stderr": 0.027044538138402616,
222
+ "acc_norm": 0.38271604938271603,
223
+ "acc_norm_stderr": 0.027044538138402616
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.35,
227
+ "acc_stderr": 0.0479372485441102,
228
+ "acc_norm": 0.35,
229
+ "acc_norm_stderr": 0.0479372485441102
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.46113989637305697,
233
+ "acc_stderr": 0.03597524411734578,
234
+ "acc_norm": 0.46113989637305697,
235
+ "acc_norm_stderr": 0.03597524411734578
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2807017543859649,
239
+ "acc_stderr": 0.04227054451232199,
240
+ "acc_norm": 0.2807017543859649,
241
+ "acc_norm_stderr": 0.04227054451232199
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.45871559633027525,
245
+ "acc_stderr": 0.02136412253388169,
246
+ "acc_norm": 0.45871559633027525,
247
+ "acc_norm_stderr": 0.02136412253388169
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3492063492063492,
251
+ "acc_stderr": 0.04263906892795133,
252
+ "acc_norm": 0.3492063492063492,
253
+ "acc_norm_stderr": 0.04263906892795133
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.43790849673202614,
257
+ "acc_stderr": 0.02840830202033269,
258
+ "acc_norm": 0.43790849673202614,
259
+ "acc_norm_stderr": 0.02840830202033269
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.38,
263
+ "acc_stderr": 0.048783173121456316,
264
+ "acc_norm": 0.38,
265
+ "acc_norm_stderr": 0.048783173121456316
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.5867768595041323,
269
+ "acc_stderr": 0.04495087843548408,
270
+ "acc_norm": 0.5867768595041323,
271
+ "acc_norm_stderr": 0.04495087843548408
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4605263157894737,
275
+ "acc_stderr": 0.04056242252249033,
276
+ "acc_norm": 0.4605263157894737,
277
+ "acc_norm_stderr": 0.04056242252249033
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.35294117647058826,
281
+ "acc_stderr": 0.01933314202079706,
282
+ "acc_norm": 0.35294117647058826,
283
+ "acc_norm_stderr": 0.01933314202079706
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.32269503546099293,
287
+ "acc_stderr": 0.027889139300534778,
288
+ "acc_norm": 0.32269503546099293,
289
+ "acc_norm_stderr": 0.027889139300534778
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.23214285714285715,
293
+ "acc_stderr": 0.04007341809755807,
294
+ "acc_norm": 0.23214285714285715,
295
+ "acc_norm_stderr": 0.04007341809755807
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.36574074074074076,
299
+ "acc_stderr": 0.03284738857647207,
300
+ "acc_norm": 0.36574074074074076,
301
+ "acc_norm_stderr": 0.03284738857647207
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2435754189944134,
305
+ "acc_stderr": 0.014355911964767857,
306
+ "acc_norm": 0.2435754189944134,
307
+ "acc_norm_stderr": 0.014355911964767857
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.35,
311
+ "acc_stderr": 0.0479372485441102,
312
+ "acc_norm": 0.35,
313
+ "acc_norm_stderr": 0.0479372485441102
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.54,
317
+ "acc_stderr": 0.05009082659620332,
318
+ "acc_norm": 0.54,
319
+ "acc_norm_stderr": 0.05009082659620332
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.31985294117647056,
323
+ "acc_stderr": 0.028332959514031225,
324
+ "acc_norm": 0.31985294117647056,
325
+ "acc_norm_stderr": 0.028332959514031225
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.4448979591836735,
329
+ "acc_stderr": 0.031814251181977865,
330
+ "acc_norm": 0.4448979591836735,
331
+ "acc_norm_stderr": 0.031814251181977865
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.569620253164557,
335
+ "acc_stderr": 0.03223017195937598,
336
+ "acc_norm": 0.569620253164557,
337
+ "acc_norm_stderr": 0.03223017195937598
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3076923076923077,
341
+ "acc_stderr": 0.011787910251664587,
342
+ "acc_norm": 0.3076923076923077,
343
+ "acc_norm_stderr": 0.011787910251664587
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4019607843137255,
347
+ "acc_stderr": 0.034411900234824655,
348
+ "acc_norm": 0.4019607843137255,
349
+ "acc_norm_stderr": 0.034411900234824655
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.42424242424242425,
353
+ "acc_stderr": 0.038592681420702615,
354
+ "acc_norm": 0.42424242424242425,
355
+ "acc_norm_stderr": 0.038592681420702615
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.27906976744186046,
359
+ "mc1_stderr": 0.015702107090627877,
360
+ "mc2": 0.4571128110826051,
361
+ "mc2_stderr": 0.0163313732350845
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.30460448642266824,
365
+ "acc_stderr": 0.015823367273129395,
366
+ "acc_norm": 0.3293978748524203,
367
+ "acc_norm_stderr": 0.016158746868147146
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.2.0",
436
+ "model_sha": "d233f62c06362a1008b268f72e919bd6eaf96166",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.3.0/result_2023-11-15 15:01:55.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3455631399317406,
5
+ "acc_stderr": 0.013896938461145687,
6
+ "acc_norm": 0.3873720136518771,
7
+ "acc_norm_stderr": 0.014235872487909872
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.382194781915953,
11
+ "acc_stderr": 0.004849306998727764,
12
+ "acc_norm": 0.4986058554072894,
13
+ "acc_norm_stderr": 0.004989762014739187
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5029239766081871,
17
+ "acc_stderr": 0.03834759370936839,
18
+ "acc_norm": 0.5029239766081871,
19
+ "acc_norm_stderr": 0.03834759370936839
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.6019417475728155,
23
+ "acc_stderr": 0.04846748253977238,
24
+ "acc_norm": 0.6019417475728155,
25
+ "acc_norm_stderr": 0.04846748253977238
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.47126436781609193,
29
+ "acc_stderr": 0.01785041079438017,
30
+ "acc_norm": 0.47126436781609193,
31
+ "acc_norm_stderr": 0.01785041079438017
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.34074074074074073,
35
+ "acc_stderr": 0.040943762699967946,
36
+ "acc_norm": 0.34074074074074073,
37
+ "acc_norm_stderr": 0.040943762699967946
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.22,
41
+ "acc_stderr": 0.04163331998932269,
42
+ "acc_norm": 0.22,
43
+ "acc_norm_stderr": 0.04163331998932269
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.44680851063829785,
47
+ "acc_stderr": 0.0325005368436584,
48
+ "acc_norm": 0.44680851063829785,
49
+ "acc_norm_stderr": 0.0325005368436584
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39156626506024095,
53
+ "acc_stderr": 0.037998574544796354,
54
+ "acc_norm": 0.39156626506024095,
55
+ "acc_norm_stderr": 0.037998574544796354
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.45980707395498394,
59
+ "acc_stderr": 0.028306190403305696,
60
+ "acc_norm": 0.45980707395498394,
61
+ "acc_norm_stderr": 0.028306190403305696
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.452914798206278,
65
+ "acc_stderr": 0.03340867501923325,
66
+ "acc_norm": 0.452914798206278,
67
+ "acc_norm_stderr": 0.03340867501923325
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.46564885496183206,
71
+ "acc_stderr": 0.04374928560599738,
72
+ "acc_norm": 0.46564885496183206,
73
+ "acc_norm_stderr": 0.04374928560599738
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.4,
77
+ "acc_stderr": 0.049236596391733084,
78
+ "acc_norm": 0.4,
79
+ "acc_norm_stderr": 0.049236596391733084
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5353535353535354,
83
+ "acc_stderr": 0.035534363688280626,
84
+ "acc_norm": 0.5353535353535354,
85
+ "acc_norm_stderr": 0.035534363688280626
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4689655172413793,
89
+ "acc_stderr": 0.04158632762097828,
90
+ "acc_norm": 0.4689655172413793,
91
+ "acc_norm_stderr": 0.04158632762097828
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.28431372549019607,
95
+ "acc_stderr": 0.04488482852329017,
96
+ "acc_norm": 0.28431372549019607,
97
+ "acc_norm_stderr": 0.04488482852329017
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.5336134453781513,
101
+ "acc_stderr": 0.03240501447690071,
102
+ "acc_norm": 0.5336134453781513,
103
+ "acc_norm_stderr": 0.03240501447690071
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.46923076923076923,
107
+ "acc_stderr": 0.025302958890850154,
108
+ "acc_norm": 0.46923076923076923,
109
+ "acc_norm_stderr": 0.025302958890850154
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.57,
113
+ "acc_stderr": 0.049756985195624284,
114
+ "acc_norm": 0.57,
115
+ "acc_norm_stderr": 0.049756985195624284
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.33,
119
+ "acc_stderr": 0.04725815626252604,
120
+ "acc_norm": 0.33,
121
+ "acc_norm_stderr": 0.04725815626252604
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.48148148148148145,
125
+ "acc_stderr": 0.04830366024635331,
126
+ "acc_norm": 0.48148148148148145,
127
+ "acc_norm_stderr": 0.04830366024635331
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4433497536945813,
131
+ "acc_stderr": 0.034953345821629324,
132
+ "acc_norm": 0.4433497536945813,
133
+ "acc_norm_stderr": 0.034953345821629324
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4612903225806452,
137
+ "acc_stderr": 0.028358634859836914,
138
+ "acc_norm": 0.4612903225806452,
139
+ "acc_norm_stderr": 0.028358634859836914
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.7435897435897436,
143
+ "acc_stderr": 0.028605953702004243,
144
+ "acc_norm": 0.7435897435897436,
145
+ "acc_norm_stderr": 0.028605953702004243
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4490566037735849,
149
+ "acc_stderr": 0.030612730713641095,
150
+ "acc_norm": 0.4490566037735849,
151
+ "acc_norm_stderr": 0.030612730713641095
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5,
155
+ "acc_stderr": 0.04789131426105757,
156
+ "acc_norm": 0.5,
157
+ "acc_norm_stderr": 0.04789131426105757
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.32592592592592595,
161
+ "acc_stderr": 0.028578348365473075,
162
+ "acc_norm": 0.32592592592592595,
163
+ "acc_norm_stderr": 0.028578348365473075
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2582781456953642,
167
+ "acc_stderr": 0.035737053147634576,
168
+ "acc_norm": 0.2582781456953642,
169
+ "acc_norm_stderr": 0.035737053147634576
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5920398009950248,
173
+ "acc_stderr": 0.03475116365194092,
174
+ "acc_norm": 0.5920398009950248,
175
+ "acc_norm_stderr": 0.03475116365194092
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3930635838150289,
179
+ "acc_stderr": 0.03724249595817729,
180
+ "acc_norm": 0.3930635838150289,
181
+ "acc_norm_stderr": 0.03724249595817729
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.36772486772486773,
185
+ "acc_stderr": 0.02483383982556242,
186
+ "acc_norm": 0.36772486772486773,
187
+ "acc_norm_stderr": 0.02483383982556242
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3472222222222222,
191
+ "acc_stderr": 0.039812405437178615,
192
+ "acc_norm": 0.3472222222222222,
193
+ "acc_norm_stderr": 0.039812405437178615
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.35,
197
+ "acc_stderr": 0.0479372485441102,
198
+ "acc_norm": 0.35,
199
+ "acc_norm_stderr": 0.0479372485441102
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.6,
203
+ "acc_stderr": 0.04923659639173309,
204
+ "acc_norm": 0.6,
205
+ "acc_norm_stderr": 0.04923659639173309
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.48265895953757226,
209
+ "acc_stderr": 0.02690290045866664,
210
+ "acc_norm": 0.48265895953757226,
211
+ "acc_norm_stderr": 0.02690290045866664
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.50920245398773,
215
+ "acc_stderr": 0.039277056007874414,
216
+ "acc_norm": 0.50920245398773,
217
+ "acc_norm_stderr": 0.039277056007874414
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.44753086419753085,
221
+ "acc_stderr": 0.02766713856942271,
222
+ "acc_norm": 0.44753086419753085,
223
+ "acc_norm_stderr": 0.02766713856942271
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.31,
227
+ "acc_stderr": 0.04648231987117316,
228
+ "acc_norm": 0.31,
229
+ "acc_norm_stderr": 0.04648231987117316
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5233160621761658,
233
+ "acc_stderr": 0.03604513672442203,
234
+ "acc_norm": 0.5233160621761658,
235
+ "acc_norm_stderr": 0.03604513672442203
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2894736842105263,
239
+ "acc_stderr": 0.04266339443159395,
240
+ "acc_norm": 0.2894736842105263,
241
+ "acc_norm_stderr": 0.04266339443159395
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.4972477064220184,
245
+ "acc_stderr": 0.02143699835976532,
246
+ "acc_norm": 0.4972477064220184,
247
+ "acc_norm_stderr": 0.02143699835976532
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.4444444444444444,
251
+ "acc_stderr": 0.044444444444444495,
252
+ "acc_norm": 0.4444444444444444,
253
+ "acc_norm_stderr": 0.044444444444444495
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4934640522875817,
257
+ "acc_stderr": 0.028627470550556054,
258
+ "acc_norm": 0.4934640522875817,
259
+ "acc_norm_stderr": 0.028627470550556054
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.46,
263
+ "acc_stderr": 0.05009082659620332,
264
+ "acc_norm": 0.46,
265
+ "acc_norm_stderr": 0.05009082659620332
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6859504132231405,
269
+ "acc_stderr": 0.04236964753041018,
270
+ "acc_norm": 0.6859504132231405,
271
+ "acc_norm_stderr": 0.04236964753041018
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4144736842105263,
275
+ "acc_stderr": 0.04008973785779206,
276
+ "acc_norm": 0.4144736842105263,
277
+ "acc_norm_stderr": 0.04008973785779206
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.39215686274509803,
281
+ "acc_stderr": 0.019751726508762637,
282
+ "acc_norm": 0.39215686274509803,
283
+ "acc_norm_stderr": 0.019751726508762637
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3262411347517731,
287
+ "acc_stderr": 0.027968453043563168,
288
+ "acc_norm": 0.3262411347517731,
289
+ "acc_norm_stderr": 0.027968453043563168
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.3482142857142857,
293
+ "acc_stderr": 0.045218299028335865,
294
+ "acc_norm": 0.3482142857142857,
295
+ "acc_norm_stderr": 0.045218299028335865
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.44907407407407407,
299
+ "acc_stderr": 0.03392238405321617,
300
+ "acc_norm": 0.44907407407407407,
301
+ "acc_norm_stderr": 0.03392238405321617
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.30837988826815643,
305
+ "acc_stderr": 0.01544571691099888,
306
+ "acc_norm": 0.30837988826815643,
307
+ "acc_norm_stderr": 0.01544571691099888
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.48,
311
+ "acc_stderr": 0.050211673156867795,
312
+ "acc_norm": 0.48,
313
+ "acc_norm_stderr": 0.050211673156867795
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.62,
317
+ "acc_stderr": 0.048783173121456316,
318
+ "acc_norm": 0.62,
319
+ "acc_norm_stderr": 0.048783173121456316
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3786764705882353,
323
+ "acc_stderr": 0.029465133639776125,
324
+ "acc_norm": 0.3786764705882353,
325
+ "acc_norm_stderr": 0.029465133639776125
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5510204081632653,
329
+ "acc_stderr": 0.03184213866687578,
330
+ "acc_norm": 0.5510204081632653,
331
+ "acc_norm_stderr": 0.03184213866687578
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.510548523206751,
335
+ "acc_stderr": 0.032539983791662855,
336
+ "acc_norm": 0.510548523206751,
337
+ "acc_norm_stderr": 0.032539983791662855
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3318122555410691,
341
+ "acc_stderr": 0.012026088259897634,
342
+ "acc_norm": 0.3318122555410691,
343
+ "acc_norm_stderr": 0.012026088259897634
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4117647058823529,
347
+ "acc_stderr": 0.03454236585380609,
348
+ "acc_norm": 0.4117647058823529,
349
+ "acc_norm_stderr": 0.03454236585380609
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.46060606060606063,
353
+ "acc_stderr": 0.03892207016552013,
354
+ "acc_norm": 0.46060606060606063,
355
+ "acc_norm_stderr": 0.03892207016552013
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2802937576499388,
359
+ "mc1_stderr": 0.015723139524608742,
360
+ "mc2": 0.4538855040890016,
361
+ "mc2_stderr": 0.015473472871845475
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5371900826446281,
365
+ "acc_stderr": 0.017142736117643304,
366
+ "acc_norm": 0.5796930342384888,
367
+ "acc_norm_stderr": 0.01697059828117771
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.3.0",
436
+ "model_sha": "995ca1c4360613685103c646f290b0062770ec7b",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.5.0/result_2023-11-16 23:47:33.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.34044368600682595,
5
+ "acc_stderr": 0.01384746051889298,
6
+ "acc_norm": 0.37457337883959047,
7
+ "acc_norm_stderr": 0.014144193471893446
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3906592312288389,
11
+ "acc_stderr": 0.004869010152280755,
12
+ "acc_norm": 0.5010953993228441,
13
+ "acc_norm_stderr": 0.004989769436956922
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.47368421052631576,
17
+ "acc_stderr": 0.038295098689947286,
18
+ "acc_norm": 0.47368421052631576,
19
+ "acc_norm_stderr": 0.038295098689947286
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5922330097087378,
23
+ "acc_stderr": 0.04865777570410768,
24
+ "acc_norm": 0.5922330097087378,
25
+ "acc_norm_stderr": 0.04865777570410768
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4648786717752235,
29
+ "acc_stderr": 0.01783579880629064,
30
+ "acc_norm": 0.4648786717752235,
31
+ "acc_norm_stderr": 0.01783579880629064
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.362962962962963,
35
+ "acc_stderr": 0.04153948404742398,
36
+ "acc_norm": 0.362962962962963,
37
+ "acc_norm_stderr": 0.04153948404742398
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.25,
41
+ "acc_stderr": 0.04351941398892446,
42
+ "acc_norm": 0.25,
43
+ "acc_norm_stderr": 0.04351941398892446
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.40425531914893614,
47
+ "acc_stderr": 0.03208115750788684,
48
+ "acc_norm": 0.40425531914893614,
49
+ "acc_norm_stderr": 0.03208115750788684
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.42168674698795183,
53
+ "acc_stderr": 0.03844453181770917,
54
+ "acc_norm": 0.42168674698795183,
55
+ "acc_norm_stderr": 0.03844453181770917
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4919614147909968,
59
+ "acc_stderr": 0.028394421370984548,
60
+ "acc_norm": 0.4919614147909968,
61
+ "acc_norm_stderr": 0.028394421370984548
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.452914798206278,
65
+ "acc_stderr": 0.03340867501923325,
66
+ "acc_norm": 0.452914798206278,
67
+ "acc_norm_stderr": 0.03340867501923325
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.5038167938931297,
71
+ "acc_stderr": 0.043851623256015534,
72
+ "acc_norm": 0.5038167938931297,
73
+ "acc_norm_stderr": 0.043851623256015534
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.4,
77
+ "acc_stderr": 0.049236596391733084,
78
+ "acc_norm": 0.4,
79
+ "acc_norm_stderr": 0.049236596391733084
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5909090909090909,
83
+ "acc_stderr": 0.03502975799413008,
84
+ "acc_norm": 0.5909090909090909,
85
+ "acc_norm_stderr": 0.03502975799413008
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.496551724137931,
89
+ "acc_stderr": 0.041665675771015785,
90
+ "acc_norm": 0.496551724137931,
91
+ "acc_norm_stderr": 0.041665675771015785
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.2647058823529412,
95
+ "acc_stderr": 0.04389869956808778,
96
+ "acc_norm": 0.2647058823529412,
97
+ "acc_norm_stderr": 0.04389869956808778
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.5084033613445378,
101
+ "acc_stderr": 0.0324739027656967,
102
+ "acc_norm": 0.5084033613445378,
103
+ "acc_norm_stderr": 0.0324739027656967
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4641025641025641,
107
+ "acc_stderr": 0.025285585990017834,
108
+ "acc_norm": 0.4641025641025641,
109
+ "acc_norm_stderr": 0.025285585990017834
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.53,
113
+ "acc_stderr": 0.05016135580465919,
114
+ "acc_norm": 0.53,
115
+ "acc_norm_stderr": 0.05016135580465919
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.36,
119
+ "acc_stderr": 0.048241815132442176,
120
+ "acc_norm": 0.36,
121
+ "acc_norm_stderr": 0.048241815132442176
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.4722222222222222,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.4722222222222222,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4236453201970443,
131
+ "acc_stderr": 0.03476725747649038,
132
+ "acc_norm": 0.4236453201970443,
133
+ "acc_norm_stderr": 0.03476725747649038
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.49032258064516127,
137
+ "acc_stderr": 0.02843867799890955,
138
+ "acc_norm": 0.49032258064516127,
139
+ "acc_norm_stderr": 0.02843867799890955
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6965811965811965,
143
+ "acc_stderr": 0.030118210106942638,
144
+ "acc_norm": 0.6965811965811965,
145
+ "acc_norm_stderr": 0.030118210106942638
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4528301886792453,
149
+ "acc_stderr": 0.03063562795796183,
150
+ "acc_norm": 0.4528301886792453,
151
+ "acc_norm_stderr": 0.03063562795796183
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.509090909090909,
155
+ "acc_stderr": 0.04788339768702861,
156
+ "acc_norm": 0.509090909090909,
157
+ "acc_norm_stderr": 0.04788339768702861
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.32592592592592595,
161
+ "acc_stderr": 0.02857834836547308,
162
+ "acc_norm": 0.32592592592592595,
163
+ "acc_norm_stderr": 0.02857834836547308
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2980132450331126,
167
+ "acc_stderr": 0.037345356767871984,
168
+ "acc_norm": 0.2980132450331126,
169
+ "acc_norm_stderr": 0.037345356767871984
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.6069651741293532,
173
+ "acc_stderr": 0.0345368246603156,
174
+ "acc_norm": 0.6069651741293532,
175
+ "acc_norm_stderr": 0.0345368246603156
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3872832369942196,
179
+ "acc_stderr": 0.03714325906302065,
180
+ "acc_norm": 0.3872832369942196,
181
+ "acc_norm_stderr": 0.03714325906302065
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.34656084656084657,
185
+ "acc_stderr": 0.024508777521028417,
186
+ "acc_norm": 0.34656084656084657,
187
+ "acc_norm_stderr": 0.024508777521028417
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3263888888888889,
191
+ "acc_stderr": 0.03921067198982266,
192
+ "acc_norm": 0.3263888888888889,
193
+ "acc_norm_stderr": 0.03921067198982266
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.33,
197
+ "acc_stderr": 0.047258156262526045,
198
+ "acc_norm": 0.33,
199
+ "acc_norm_stderr": 0.047258156262526045
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.6,
203
+ "acc_stderr": 0.04923659639173309,
204
+ "acc_norm": 0.6,
205
+ "acc_norm_stderr": 0.04923659639173309
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5260115606936416,
209
+ "acc_stderr": 0.026882643434022895,
210
+ "acc_norm": 0.5260115606936416,
211
+ "acc_norm_stderr": 0.026882643434022895
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.48466257668711654,
215
+ "acc_stderr": 0.03926522378708843,
216
+ "acc_norm": 0.48466257668711654,
217
+ "acc_norm_stderr": 0.03926522378708843
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4660493827160494,
221
+ "acc_stderr": 0.027756535257347666,
222
+ "acc_norm": 0.4660493827160494,
223
+ "acc_norm_stderr": 0.027756535257347666
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.35,
227
+ "acc_stderr": 0.0479372485441102,
228
+ "acc_norm": 0.35,
229
+ "acc_norm_stderr": 0.0479372485441102
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.538860103626943,
233
+ "acc_stderr": 0.03597524411734578,
234
+ "acc_norm": 0.538860103626943,
235
+ "acc_norm_stderr": 0.03597524411734578
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.3508771929824561,
239
+ "acc_stderr": 0.044895393502706986,
240
+ "acc_norm": 0.3508771929824561,
241
+ "acc_norm_stderr": 0.044895393502706986
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.47706422018348627,
245
+ "acc_stderr": 0.021414757058175506,
246
+ "acc_norm": 0.47706422018348627,
247
+ "acc_norm_stderr": 0.021414757058175506
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.36507936507936506,
251
+ "acc_stderr": 0.04306241259127153,
252
+ "acc_norm": 0.36507936507936506,
253
+ "acc_norm_stderr": 0.04306241259127153
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4738562091503268,
257
+ "acc_stderr": 0.028590752958852387,
258
+ "acc_norm": 0.4738562091503268,
259
+ "acc_norm_stderr": 0.028590752958852387
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.47,
263
+ "acc_stderr": 0.05016135580465919,
264
+ "acc_norm": 0.47,
265
+ "acc_norm_stderr": 0.05016135580465919
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6859504132231405,
269
+ "acc_stderr": 0.04236964753041018,
270
+ "acc_norm": 0.6859504132231405,
271
+ "acc_norm_stderr": 0.04236964753041018
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40131578947368424,
275
+ "acc_stderr": 0.03988903703336285,
276
+ "acc_norm": 0.40131578947368424,
277
+ "acc_norm_stderr": 0.03988903703336285
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3758169934640523,
281
+ "acc_stderr": 0.01959402113657745,
282
+ "acc_norm": 0.3758169934640523,
283
+ "acc_norm_stderr": 0.01959402113657745
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.36879432624113473,
287
+ "acc_stderr": 0.028782227561347233,
288
+ "acc_norm": 0.36879432624113473,
289
+ "acc_norm_stderr": 0.028782227561347233
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.45535714285714285,
293
+ "acc_stderr": 0.04726835553719099,
294
+ "acc_norm": 0.45535714285714285,
295
+ "acc_norm_stderr": 0.04726835553719099
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3888888888888889,
299
+ "acc_stderr": 0.033247089118091176,
300
+ "acc_norm": 0.3888888888888889,
301
+ "acc_norm_stderr": 0.033247089118091176
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.27039106145251396,
305
+ "acc_stderr": 0.014854993938010078,
306
+ "acc_norm": 0.27039106145251396,
307
+ "acc_norm_stderr": 0.014854993938010078
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.36,
311
+ "acc_stderr": 0.04824181513244218,
312
+ "acc_norm": 0.36,
313
+ "acc_norm_stderr": 0.04824181513244218
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.58,
317
+ "acc_stderr": 0.049604496374885836,
318
+ "acc_norm": 0.58,
319
+ "acc_norm_stderr": 0.049604496374885836
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3860294117647059,
323
+ "acc_stderr": 0.029573269134411124,
324
+ "acc_norm": 0.3860294117647059,
325
+ "acc_norm_stderr": 0.029573269134411124
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5673469387755102,
329
+ "acc_stderr": 0.03171752824062665,
330
+ "acc_norm": 0.5673469387755102,
331
+ "acc_norm_stderr": 0.03171752824062665
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5611814345991561,
335
+ "acc_stderr": 0.032302649315470375,
336
+ "acc_norm": 0.5611814345991561,
337
+ "acc_norm_stderr": 0.032302649315470375
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3363754889178618,
341
+ "acc_stderr": 0.012067083079452225,
342
+ "acc_norm": 0.3363754889178618,
343
+ "acc_norm_stderr": 0.012067083079452225
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.43137254901960786,
347
+ "acc_stderr": 0.03476099060501636,
348
+ "acc_norm": 0.43137254901960786,
349
+ "acc_norm_stderr": 0.03476099060501636
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.45454545454545453,
353
+ "acc_stderr": 0.03888176921674098,
354
+ "acc_norm": 0.45454545454545453,
355
+ "acc_norm_stderr": 0.03888176921674098
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2692778457772338,
359
+ "mc1_stderr": 0.015528566637087307,
360
+ "mc2": 0.4312897833619012,
361
+ "mc2_stderr": 0.015536038118672747
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.4037780401416765,
365
+ "acc_stderr": 0.01686903154029863,
366
+ "acc_norm": 0.42502951593860683,
367
+ "acc_norm_stderr": 0.016996016308362887
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.5.0",
436
+ "model_sha": "59b094a8741371d220147b53e7536af0fcf27d2e",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.6.0/result_2023-11-19 13:15:31.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.34044368600682595,
5
+ "acc_stderr": 0.01384746051889298,
6
+ "acc_norm": 0.3728668941979522,
7
+ "acc_norm_stderr": 0.014131176760131163
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3867755427205736,
11
+ "acc_stderr": 0.00486016207633099,
12
+ "acc_norm": 0.4992033459470225,
13
+ "acc_norm_stderr": 0.0049897750778356495
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4619883040935672,
17
+ "acc_stderr": 0.03823727092882307,
18
+ "acc_norm": 0.4619883040935672,
19
+ "acc_norm_stderr": 0.03823727092882307
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.6310679611650486,
23
+ "acc_stderr": 0.0477761518115674,
24
+ "acc_norm": 0.6310679611650486,
25
+ "acc_norm_stderr": 0.0477761518115674
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.47509578544061304,
29
+ "acc_stderr": 0.017857770704901018,
30
+ "acc_norm": 0.47509578544061304,
31
+ "acc_norm_stderr": 0.017857770704901018
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.34814814814814815,
35
+ "acc_stderr": 0.041153246103369526,
36
+ "acc_norm": 0.34814814814814815,
37
+ "acc_norm_stderr": 0.041153246103369526
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.25,
41
+ "acc_stderr": 0.04351941398892446,
42
+ "acc_norm": 0.25,
43
+ "acc_norm_stderr": 0.04351941398892446
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.37446808510638296,
47
+ "acc_stderr": 0.03163910665367291,
48
+ "acc_norm": 0.37446808510638296,
49
+ "acc_norm_stderr": 0.03163910665367291
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3674698795180723,
53
+ "acc_stderr": 0.03753267402120574,
54
+ "acc_norm": 0.3674698795180723,
55
+ "acc_norm_stderr": 0.03753267402120574
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.44694533762057875,
59
+ "acc_stderr": 0.028237769422085335,
60
+ "acc_norm": 0.44694533762057875,
61
+ "acc_norm_stderr": 0.028237769422085335
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.4260089686098655,
65
+ "acc_stderr": 0.033188332862172806,
66
+ "acc_norm": 0.4260089686098655,
67
+ "acc_norm_stderr": 0.033188332862172806
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.42748091603053434,
71
+ "acc_stderr": 0.04338920305792401,
72
+ "acc_norm": 0.42748091603053434,
73
+ "acc_norm_stderr": 0.04338920305792401
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.37,
77
+ "acc_stderr": 0.04852365870939098,
78
+ "acc_norm": 0.37,
79
+ "acc_norm_stderr": 0.04852365870939098
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5505050505050505,
83
+ "acc_stderr": 0.035441324919479704,
84
+ "acc_norm": 0.5505050505050505,
85
+ "acc_norm_stderr": 0.035441324919479704
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4896551724137931,
89
+ "acc_stderr": 0.041657747757287644,
90
+ "acc_norm": 0.4896551724137931,
91
+ "acc_norm_stderr": 0.041657747757287644
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.27450980392156865,
95
+ "acc_stderr": 0.04440521906179328,
96
+ "acc_norm": 0.27450980392156865,
97
+ "acc_norm_stderr": 0.04440521906179328
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.5210084033613446,
101
+ "acc_stderr": 0.03244980849990029,
102
+ "acc_norm": 0.5210084033613446,
103
+ "acc_norm_stderr": 0.03244980849990029
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.46153846153846156,
107
+ "acc_stderr": 0.025275892070240634,
108
+ "acc_norm": 0.46153846153846156,
109
+ "acc_norm_stderr": 0.025275892070240634
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.58,
113
+ "acc_stderr": 0.04960449637488583,
114
+ "acc_norm": 0.58,
115
+ "acc_norm_stderr": 0.04960449637488583
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.32,
119
+ "acc_stderr": 0.04688261722621504,
120
+ "acc_norm": 0.32,
121
+ "acc_norm_stderr": 0.04688261722621504
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.4444444444444444,
125
+ "acc_stderr": 0.04803752235190193,
126
+ "acc_norm": 0.4444444444444444,
127
+ "acc_norm_stderr": 0.04803752235190193
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3842364532019704,
131
+ "acc_stderr": 0.03422398565657553,
132
+ "acc_norm": 0.3842364532019704,
133
+ "acc_norm_stderr": 0.03422398565657553
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.44193548387096776,
137
+ "acc_stderr": 0.028251557906849724,
138
+ "acc_norm": 0.44193548387096776,
139
+ "acc_norm_stderr": 0.028251557906849724
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6923076923076923,
143
+ "acc_stderr": 0.030236389942173095,
144
+ "acc_norm": 0.6923076923076923,
145
+ "acc_norm_stderr": 0.030236389942173095
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.42641509433962266,
149
+ "acc_stderr": 0.030437794342983045,
150
+ "acc_norm": 0.42641509433962266,
151
+ "acc_norm_stderr": 0.030437794342983045
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5,
155
+ "acc_stderr": 0.04789131426105757,
156
+ "acc_norm": 0.5,
157
+ "acc_norm_stderr": 0.04789131426105757
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.32222222222222224,
161
+ "acc_stderr": 0.028493465091028593,
162
+ "acc_norm": 0.32222222222222224,
163
+ "acc_norm_stderr": 0.028493465091028593
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.24503311258278146,
167
+ "acc_stderr": 0.03511807571804724,
168
+ "acc_norm": 0.24503311258278146,
169
+ "acc_norm_stderr": 0.03511807571804724
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5870646766169154,
173
+ "acc_stderr": 0.03481520803367348,
174
+ "acc_norm": 0.5870646766169154,
175
+ "acc_norm_stderr": 0.03481520803367348
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.37572254335260113,
179
+ "acc_stderr": 0.03692820767264867,
180
+ "acc_norm": 0.37572254335260113,
181
+ "acc_norm_stderr": 0.03692820767264867
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.36243386243386244,
185
+ "acc_stderr": 0.024757473902752056,
186
+ "acc_norm": 0.36243386243386244,
187
+ "acc_norm_stderr": 0.024757473902752056
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3611111111111111,
191
+ "acc_stderr": 0.040166600304512336,
192
+ "acc_norm": 0.3611111111111111,
193
+ "acc_norm_stderr": 0.040166600304512336
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.37,
197
+ "acc_stderr": 0.04852365870939099,
198
+ "acc_norm": 0.37,
199
+ "acc_norm_stderr": 0.04852365870939099
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.55,
203
+ "acc_stderr": 0.05,
204
+ "acc_norm": 0.55,
205
+ "acc_norm_stderr": 0.05
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.476878612716763,
209
+ "acc_stderr": 0.026890297881303125,
210
+ "acc_norm": 0.476878612716763,
211
+ "acc_norm_stderr": 0.026890297881303125
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4539877300613497,
215
+ "acc_stderr": 0.0391170190467718,
216
+ "acc_norm": 0.4539877300613497,
217
+ "acc_norm_stderr": 0.0391170190467718
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.42592592592592593,
221
+ "acc_stderr": 0.027513747284379414,
222
+ "acc_norm": 0.42592592592592593,
223
+ "acc_norm_stderr": 0.027513747284379414
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.29,
227
+ "acc_stderr": 0.045604802157206845,
228
+ "acc_norm": 0.29,
229
+ "acc_norm_stderr": 0.045604802157206845
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.49222797927461137,
233
+ "acc_stderr": 0.03608003225569654,
234
+ "acc_norm": 0.49222797927461137,
235
+ "acc_norm_stderr": 0.03608003225569654
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.3508771929824561,
239
+ "acc_stderr": 0.04489539350270699,
240
+ "acc_norm": 0.3508771929824561,
241
+ "acc_norm_stderr": 0.04489539350270699
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.47522935779816516,
245
+ "acc_stderr": 0.021410999753635918,
246
+ "acc_norm": 0.47522935779816516,
247
+ "acc_norm_stderr": 0.021410999753635918
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3968253968253968,
251
+ "acc_stderr": 0.043758884927270605,
252
+ "acc_norm": 0.3968253968253968,
253
+ "acc_norm_stderr": 0.043758884927270605
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.4477124183006536,
257
+ "acc_stderr": 0.028472938478033522,
258
+ "acc_norm": 0.4477124183006536,
259
+ "acc_norm_stderr": 0.028472938478033522
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.46,
263
+ "acc_stderr": 0.05009082659620332,
264
+ "acc_norm": 0.46,
265
+ "acc_norm_stderr": 0.05009082659620332
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6446280991735537,
269
+ "acc_stderr": 0.0436923632657398,
270
+ "acc_norm": 0.6446280991735537,
271
+ "acc_norm_stderr": 0.0436923632657398
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40131578947368424,
275
+ "acc_stderr": 0.03988903703336284,
276
+ "acc_norm": 0.40131578947368424,
277
+ "acc_norm_stderr": 0.03988903703336284
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.39052287581699346,
281
+ "acc_stderr": 0.019737008998094597,
282
+ "acc_norm": 0.39052287581699346,
283
+ "acc_norm_stderr": 0.019737008998094597
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.31560283687943264,
287
+ "acc_stderr": 0.027724989449509314,
288
+ "acc_norm": 0.31560283687943264,
289
+ "acc_norm_stderr": 0.027724989449509314
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.29464285714285715,
293
+ "acc_stderr": 0.04327040932578728,
294
+ "acc_norm": 0.29464285714285715,
295
+ "acc_norm_stderr": 0.04327040932578728
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.41203703703703703,
299
+ "acc_stderr": 0.03356787758160835,
300
+ "acc_norm": 0.41203703703703703,
301
+ "acc_norm_stderr": 0.03356787758160835
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2860335195530726,
305
+ "acc_stderr": 0.015113972129062125,
306
+ "acc_norm": 0.2860335195530726,
307
+ "acc_norm_stderr": 0.015113972129062125
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.37,
311
+ "acc_stderr": 0.04852365870939099,
312
+ "acc_norm": 0.37,
313
+ "acc_norm_stderr": 0.04852365870939099
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.59,
317
+ "acc_stderr": 0.04943110704237102,
318
+ "acc_norm": 0.59,
319
+ "acc_norm_stderr": 0.04943110704237102
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3235294117647059,
323
+ "acc_stderr": 0.02841820861940679,
324
+ "acc_norm": 0.3235294117647059,
325
+ "acc_norm_stderr": 0.02841820861940679
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5142857142857142,
329
+ "acc_stderr": 0.03199615232806286,
330
+ "acc_norm": 0.5142857142857142,
331
+ "acc_norm_stderr": 0.03199615232806286
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.540084388185654,
335
+ "acc_stderr": 0.03244246810187913,
336
+ "acc_norm": 0.540084388185654,
337
+ "acc_norm_stderr": 0.03244246810187913
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3363754889178618,
341
+ "acc_stderr": 0.01206708307945223,
342
+ "acc_norm": 0.3363754889178618,
343
+ "acc_norm_stderr": 0.01206708307945223
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.38235294117647056,
347
+ "acc_stderr": 0.034107853389047184,
348
+ "acc_norm": 0.38235294117647056,
349
+ "acc_norm_stderr": 0.034107853389047184
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.44242424242424244,
353
+ "acc_stderr": 0.03878372113711274,
354
+ "acc_norm": 0.44242424242424244,
355
+ "acc_norm_stderr": 0.03878372113711274
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2802937576499388,
359
+ "mc1_stderr": 0.015723139524608742,
360
+ "mc2": 0.4436091279270421,
361
+ "mc2_stderr": 0.015994798162179236
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5194805194805194,
365
+ "acc_stderr": 0.01717730199234254,
366
+ "acc_norm": 0.5430932703659976,
367
+ "acc_norm_stderr": 0.017126389093086777
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.6.0",
436
+ "model_sha": "b149b065fe748591389f5ce440e424a118880e26",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.7.0/result_2023-11-20 09:22:14.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3447098976109215,
5
+ "acc_stderr": 0.01388881628678211,
6
+ "acc_norm": 0.386518771331058,
7
+ "acc_norm_stderr": 0.014230084761910474
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3873730332603067,
11
+ "acc_stderr": 0.004861544478451848,
12
+ "acc_norm": 0.5089623580959968,
13
+ "acc_norm_stderr": 0.004988979750014428
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.49122807017543857,
17
+ "acc_stderr": 0.038342347441649924,
18
+ "acc_norm": 0.49122807017543857,
19
+ "acc_norm_stderr": 0.038342347441649924
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.6019417475728155,
23
+ "acc_stderr": 0.04846748253977238,
24
+ "acc_norm": 0.6019417475728155,
25
+ "acc_norm_stderr": 0.04846748253977238
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4827586206896552,
29
+ "acc_stderr": 0.017869330154003705,
30
+ "acc_norm": 0.4827586206896552,
31
+ "acc_norm_stderr": 0.017869330154003705
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.362962962962963,
35
+ "acc_stderr": 0.041539484047424,
36
+ "acc_norm": 0.362962962962963,
37
+ "acc_norm_stderr": 0.041539484047424
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.26,
41
+ "acc_stderr": 0.0440844002276808,
42
+ "acc_norm": 0.26,
43
+ "acc_norm_stderr": 0.0440844002276808
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.41702127659574467,
47
+ "acc_stderr": 0.03223276266711712,
48
+ "acc_norm": 0.41702127659574467,
49
+ "acc_norm_stderr": 0.03223276266711712
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39156626506024095,
53
+ "acc_stderr": 0.03799857454479636,
54
+ "acc_norm": 0.39156626506024095,
55
+ "acc_norm_stderr": 0.03799857454479636
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4662379421221865,
59
+ "acc_stderr": 0.02833327710956278,
60
+ "acc_norm": 0.4662379421221865,
61
+ "acc_norm_stderr": 0.02833327710956278
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.43946188340807174,
65
+ "acc_stderr": 0.03331092511038179,
66
+ "acc_norm": 0.43946188340807174,
67
+ "acc_norm_stderr": 0.03331092511038179
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.42748091603053434,
71
+ "acc_stderr": 0.043389203057924,
72
+ "acc_norm": 0.42748091603053434,
73
+ "acc_norm_stderr": 0.043389203057924
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.34,
77
+ "acc_stderr": 0.04760952285695235,
78
+ "acc_norm": 0.34,
79
+ "acc_norm_stderr": 0.04760952285695235
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5353535353535354,
83
+ "acc_stderr": 0.03553436368828063,
84
+ "acc_norm": 0.5353535353535354,
85
+ "acc_norm_stderr": 0.03553436368828063
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4896551724137931,
89
+ "acc_stderr": 0.041657747757287644,
90
+ "acc_norm": 0.4896551724137931,
91
+ "acc_norm_stderr": 0.041657747757287644
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.29411764705882354,
95
+ "acc_stderr": 0.04533838195929778,
96
+ "acc_norm": 0.29411764705882354,
97
+ "acc_norm_stderr": 0.04533838195929778
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.5546218487394958,
101
+ "acc_stderr": 0.0322841062671639,
102
+ "acc_norm": 0.5546218487394958,
103
+ "acc_norm_stderr": 0.0322841062671639
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.45897435897435895,
107
+ "acc_stderr": 0.025265525491284295,
108
+ "acc_norm": 0.45897435897435895,
109
+ "acc_norm_stderr": 0.025265525491284295
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.56,
113
+ "acc_stderr": 0.0498887651569859,
114
+ "acc_norm": 0.56,
115
+ "acc_norm_stderr": 0.0498887651569859
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.34,
119
+ "acc_stderr": 0.04760952285695235,
120
+ "acc_norm": 0.34,
121
+ "acc_norm_stderr": 0.04760952285695235
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.49074074074074076,
125
+ "acc_stderr": 0.04832853553437055,
126
+ "acc_norm": 0.49074074074074076,
127
+ "acc_norm_stderr": 0.04832853553437055
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4088669950738916,
131
+ "acc_stderr": 0.0345905881588323,
132
+ "acc_norm": 0.4088669950738916,
133
+ "acc_norm_stderr": 0.0345905881588323
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4774193548387097,
137
+ "acc_stderr": 0.028414985019707868,
138
+ "acc_norm": 0.4774193548387097,
139
+ "acc_norm_stderr": 0.028414985019707868
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.7307692307692307,
143
+ "acc_stderr": 0.029058588303748842,
144
+ "acc_norm": 0.7307692307692307,
145
+ "acc_norm_stderr": 0.029058588303748842
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.41509433962264153,
149
+ "acc_stderr": 0.030325945789286105,
150
+ "acc_norm": 0.41509433962264153,
151
+ "acc_norm_stderr": 0.030325945789286105
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.509090909090909,
155
+ "acc_stderr": 0.0478833976870286,
156
+ "acc_norm": 0.509090909090909,
157
+ "acc_norm_stderr": 0.0478833976870286
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.3333333333333333,
161
+ "acc_stderr": 0.028742040903948482,
162
+ "acc_norm": 0.3333333333333333,
163
+ "acc_norm_stderr": 0.028742040903948482
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.26490066225165565,
167
+ "acc_stderr": 0.03603038545360384,
168
+ "acc_norm": 0.26490066225165565,
169
+ "acc_norm_stderr": 0.03603038545360384
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.6119402985074627,
173
+ "acc_stderr": 0.03445789964362749,
174
+ "acc_norm": 0.6119402985074627,
175
+ "acc_norm_stderr": 0.03445789964362749
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3930635838150289,
179
+ "acc_stderr": 0.03724249595817729,
180
+ "acc_norm": 0.3930635838150289,
181
+ "acc_norm_stderr": 0.03724249595817729
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.35714285714285715,
185
+ "acc_stderr": 0.024677862841332783,
186
+ "acc_norm": 0.35714285714285715,
187
+ "acc_norm_stderr": 0.024677862841332783
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3888888888888889,
191
+ "acc_stderr": 0.04076663253918567,
192
+ "acc_norm": 0.3888888888888889,
193
+ "acc_norm_stderr": 0.04076663253918567
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.38,
197
+ "acc_stderr": 0.04878317312145632,
198
+ "acc_norm": 0.38,
199
+ "acc_norm_stderr": 0.04878317312145632
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.61,
203
+ "acc_stderr": 0.04902071300001975,
204
+ "acc_norm": 0.61,
205
+ "acc_norm_stderr": 0.04902071300001975
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.47398843930635837,
209
+ "acc_stderr": 0.026882643434022885,
210
+ "acc_norm": 0.47398843930635837,
211
+ "acc_norm_stderr": 0.026882643434022885
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.49693251533742333,
215
+ "acc_stderr": 0.03928297078179662,
216
+ "acc_norm": 0.49693251533742333,
217
+ "acc_norm_stderr": 0.03928297078179662
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4537037037037037,
221
+ "acc_stderr": 0.0277012284685426,
222
+ "acc_norm": 0.4537037037037037,
223
+ "acc_norm_stderr": 0.0277012284685426
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.3,
227
+ "acc_stderr": 0.046056618647183814,
228
+ "acc_norm": 0.3,
229
+ "acc_norm_stderr": 0.046056618647183814
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5129533678756477,
233
+ "acc_stderr": 0.03607228061047749,
234
+ "acc_norm": 0.5129533678756477,
235
+ "acc_norm_stderr": 0.03607228061047749
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2719298245614035,
239
+ "acc_stderr": 0.041857744240220575,
240
+ "acc_norm": 0.2719298245614035,
241
+ "acc_norm_stderr": 0.041857744240220575
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.48990825688073397,
245
+ "acc_stderr": 0.021432956203453316,
246
+ "acc_norm": 0.48990825688073397,
247
+ "acc_norm_stderr": 0.021432956203453316
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.4603174603174603,
251
+ "acc_stderr": 0.04458029125470973,
252
+ "acc_norm": 0.4603174603174603,
253
+ "acc_norm_stderr": 0.04458029125470973
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.5032679738562091,
257
+ "acc_stderr": 0.02862930519400355,
258
+ "acc_norm": 0.5032679738562091,
259
+ "acc_norm_stderr": 0.02862930519400355
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.47,
263
+ "acc_stderr": 0.05016135580465919,
264
+ "acc_norm": 0.47,
265
+ "acc_norm_stderr": 0.05016135580465919
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6859504132231405,
269
+ "acc_stderr": 0.04236964753041018,
270
+ "acc_norm": 0.6859504132231405,
271
+ "acc_norm_stderr": 0.04236964753041018
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.42105263157894735,
275
+ "acc_stderr": 0.04017901275981748,
276
+ "acc_norm": 0.42105263157894735,
277
+ "acc_norm_stderr": 0.04017901275981748
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.39869281045751637,
281
+ "acc_stderr": 0.01980828131744984,
282
+ "acc_norm": 0.39869281045751637,
283
+ "acc_norm_stderr": 0.01980828131744984
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.31560283687943264,
287
+ "acc_stderr": 0.027724989449509314,
288
+ "acc_norm": 0.31560283687943264,
289
+ "acc_norm_stderr": 0.027724989449509314
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.3482142857142857,
293
+ "acc_stderr": 0.045218299028335865,
294
+ "acc_norm": 0.3482142857142857,
295
+ "acc_norm_stderr": 0.045218299028335865
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.4398148148148148,
299
+ "acc_stderr": 0.03385177976044811,
300
+ "acc_norm": 0.4398148148148148,
301
+ "acc_norm_stderr": 0.03385177976044811
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.2748603351955307,
305
+ "acc_stderr": 0.014931316703220517,
306
+ "acc_norm": 0.2748603351955307,
307
+ "acc_norm_stderr": 0.014931316703220517
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.43,
311
+ "acc_stderr": 0.049756985195624284,
312
+ "acc_norm": 0.43,
313
+ "acc_norm_stderr": 0.049756985195624284
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.61,
317
+ "acc_stderr": 0.04902071300001974,
318
+ "acc_norm": 0.61,
319
+ "acc_norm_stderr": 0.04902071300001974
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3492647058823529,
323
+ "acc_stderr": 0.028959755196824862,
324
+ "acc_norm": 0.3492647058823529,
325
+ "acc_norm_stderr": 0.028959755196824862
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5346938775510204,
329
+ "acc_stderr": 0.03193207024425314,
330
+ "acc_norm": 0.5346938775510204,
331
+ "acc_norm_stderr": 0.03193207024425314
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5738396624472574,
335
+ "acc_stderr": 0.03219035703131774,
336
+ "acc_norm": 0.5738396624472574,
337
+ "acc_norm_stderr": 0.03219035703131774
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.33116036505867014,
341
+ "acc_stderr": 0.012020128195985774,
342
+ "acc_norm": 0.33116036505867014,
343
+ "acc_norm_stderr": 0.012020128195985774
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4411764705882353,
347
+ "acc_stderr": 0.034849415144292316,
348
+ "acc_norm": 0.4411764705882353,
349
+ "acc_norm_stderr": 0.034849415144292316
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.4666666666666667,
353
+ "acc_stderr": 0.03895658065271846,
354
+ "acc_norm": 0.4666666666666667,
355
+ "acc_norm_stderr": 0.03895658065271846
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.3023255813953488,
359
+ "mc1_stderr": 0.016077509266133033,
360
+ "mc2": 0.4714881280704747,
361
+ "mc2_stderr": 0.01554412183162796
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5230224321133412,
365
+ "acc_stderr": 0.017172121546727634,
366
+ "acc_norm": 0.577331759149941,
367
+ "acc_norm_stderr": 0.016983506079577604
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.7.0",
436
+ "model_sha": "f62174ae285bf46cc453305f1e0b76899a8bcf82",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.7.1/result_2023-11-21 00:39:48.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3455631399317406,
5
+ "acc_stderr": 0.013896938461145685,
6
+ "acc_norm": 0.39590443686006827,
7
+ "acc_norm_stderr": 0.014291228393536587
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.38717386974706236,
11
+ "acc_stderr": 0.00486108453408704,
12
+ "acc_norm": 0.5087631945827524,
13
+ "acc_norm_stderr": 0.004989014986235632
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4853801169590643,
17
+ "acc_stderr": 0.038331852752130205,
18
+ "acc_norm": 0.4853801169590643,
19
+ "acc_norm_stderr": 0.038331852752130205
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.6116504854368932,
23
+ "acc_stderr": 0.048257293373563895,
24
+ "acc_norm": 0.6116504854368932,
25
+ "acc_norm_stderr": 0.048257293373563895
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4763729246487867,
29
+ "acc_stderr": 0.017859989765176453,
30
+ "acc_norm": 0.4763729246487867,
31
+ "acc_norm_stderr": 0.017859989765176453
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.35555555555555557,
35
+ "acc_stderr": 0.04135176749720386,
36
+ "acc_norm": 0.35555555555555557,
37
+ "acc_norm_stderr": 0.04135176749720386
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.26,
41
+ "acc_stderr": 0.04408440022768079,
42
+ "acc_norm": 0.26,
43
+ "acc_norm_stderr": 0.04408440022768079
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.41702127659574467,
47
+ "acc_stderr": 0.032232762667117124,
48
+ "acc_norm": 0.41702127659574467,
49
+ "acc_norm_stderr": 0.032232762667117124
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3795180722891566,
53
+ "acc_stderr": 0.037777988227480165,
54
+ "acc_norm": 0.3795180722891566,
55
+ "acc_norm_stderr": 0.037777988227480165
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4983922829581994,
59
+ "acc_stderr": 0.02839794490780661,
60
+ "acc_norm": 0.4983922829581994,
61
+ "acc_norm_stderr": 0.02839794490780661
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.4349775784753363,
65
+ "acc_stderr": 0.03327283370271344,
66
+ "acc_norm": 0.4349775784753363,
67
+ "acc_norm_stderr": 0.03327283370271344
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4580152671755725,
71
+ "acc_stderr": 0.04369802690578756,
72
+ "acc_norm": 0.4580152671755725,
73
+ "acc_norm_stderr": 0.04369802690578756
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.35,
77
+ "acc_stderr": 0.047937248544110196,
78
+ "acc_norm": 0.35,
79
+ "acc_norm_stderr": 0.047937248544110196
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5454545454545454,
83
+ "acc_stderr": 0.03547601494006938,
84
+ "acc_norm": 0.5454545454545454,
85
+ "acc_norm_stderr": 0.03547601494006938
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.45517241379310347,
89
+ "acc_stderr": 0.04149886942192117,
90
+ "acc_norm": 0.45517241379310347,
91
+ "acc_norm_stderr": 0.04149886942192117
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.2549019607843137,
95
+ "acc_stderr": 0.043364327079931785,
96
+ "acc_norm": 0.2549019607843137,
97
+ "acc_norm_stderr": 0.043364327079931785
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.5294117647058824,
101
+ "acc_stderr": 0.03242225027115006,
102
+ "acc_norm": 0.5294117647058824,
103
+ "acc_norm_stderr": 0.03242225027115006
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.45384615384615384,
107
+ "acc_stderr": 0.025242770987126174,
108
+ "acc_norm": 0.45384615384615384,
109
+ "acc_norm_stderr": 0.025242770987126174
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.55,
113
+ "acc_stderr": 0.04999999999999998,
114
+ "acc_norm": 0.55,
115
+ "acc_norm_stderr": 0.04999999999999998
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.33,
119
+ "acc_stderr": 0.04725815626252604,
120
+ "acc_norm": 0.33,
121
+ "acc_norm_stderr": 0.04725815626252604
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5,
125
+ "acc_stderr": 0.04833682445228318,
126
+ "acc_norm": 0.5,
127
+ "acc_norm_stderr": 0.04833682445228318
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.42857142857142855,
131
+ "acc_stderr": 0.03481904844438803,
132
+ "acc_norm": 0.42857142857142855,
133
+ "acc_norm_stderr": 0.03481904844438803
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.46774193548387094,
137
+ "acc_stderr": 0.02838474778881333,
138
+ "acc_norm": 0.46774193548387094,
139
+ "acc_norm_stderr": 0.02838474778881333
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.7393162393162394,
143
+ "acc_stderr": 0.02876034895652341,
144
+ "acc_norm": 0.7393162393162394,
145
+ "acc_norm_stderr": 0.02876034895652341
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.44528301886792454,
149
+ "acc_stderr": 0.030588052974270658,
150
+ "acc_norm": 0.44528301886792454,
151
+ "acc_norm_stderr": 0.030588052974270658
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4909090909090909,
155
+ "acc_stderr": 0.04788339768702861,
156
+ "acc_norm": 0.4909090909090909,
157
+ "acc_norm_stderr": 0.04788339768702861
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.34074074074074073,
161
+ "acc_stderr": 0.02889774874113114,
162
+ "acc_norm": 0.34074074074074073,
163
+ "acc_norm_stderr": 0.02889774874113114
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2847682119205298,
167
+ "acc_stderr": 0.03684881521389024,
168
+ "acc_norm": 0.2847682119205298,
169
+ "acc_norm_stderr": 0.03684881521389024
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5870646766169154,
173
+ "acc_stderr": 0.03481520803367348,
174
+ "acc_norm": 0.5870646766169154,
175
+ "acc_norm_stderr": 0.03481520803367348
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.4161849710982659,
179
+ "acc_stderr": 0.03758517775404947,
180
+ "acc_norm": 0.4161849710982659,
181
+ "acc_norm_stderr": 0.03758517775404947
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.36507936507936506,
185
+ "acc_stderr": 0.024796060602699954,
186
+ "acc_norm": 0.36507936507936506,
187
+ "acc_norm_stderr": 0.024796060602699954
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3680555555555556,
191
+ "acc_stderr": 0.040329990539607195,
192
+ "acc_norm": 0.3680555555555556,
193
+ "acc_norm_stderr": 0.040329990539607195
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.37,
197
+ "acc_stderr": 0.048523658709391,
198
+ "acc_norm": 0.37,
199
+ "acc_norm_stderr": 0.048523658709391
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.64,
203
+ "acc_stderr": 0.048241815132442176,
204
+ "acc_norm": 0.64,
205
+ "acc_norm_stderr": 0.048241815132442176
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.4624277456647399,
209
+ "acc_stderr": 0.026842985519615375,
210
+ "acc_norm": 0.4624277456647399,
211
+ "acc_norm_stderr": 0.026842985519615375
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.5521472392638037,
215
+ "acc_stderr": 0.03906947479456607,
216
+ "acc_norm": 0.5521472392638037,
217
+ "acc_norm_stderr": 0.03906947479456607
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.44135802469135804,
221
+ "acc_stderr": 0.027628737155668777,
222
+ "acc_norm": 0.44135802469135804,
223
+ "acc_norm_stderr": 0.027628737155668777
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.29,
227
+ "acc_stderr": 0.045604802157206845,
228
+ "acc_norm": 0.29,
229
+ "acc_norm_stderr": 0.045604802157206845
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.538860103626943,
233
+ "acc_stderr": 0.035975244117345775,
234
+ "acc_norm": 0.538860103626943,
235
+ "acc_norm_stderr": 0.035975244117345775
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2719298245614035,
239
+ "acc_stderr": 0.041857744240220575,
240
+ "acc_norm": 0.2719298245614035,
241
+ "acc_norm_stderr": 0.041857744240220575
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.4990825688073395,
245
+ "acc_stderr": 0.021437287056051215,
246
+ "acc_norm": 0.4990825688073395,
247
+ "acc_norm_stderr": 0.021437287056051215
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.4126984126984127,
251
+ "acc_stderr": 0.04403438954768177,
252
+ "acc_norm": 0.4126984126984127,
253
+ "acc_norm_stderr": 0.04403438954768177
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.48366013071895425,
257
+ "acc_stderr": 0.028614624752805407,
258
+ "acc_norm": 0.48366013071895425,
259
+ "acc_norm_stderr": 0.028614624752805407
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.43,
263
+ "acc_stderr": 0.049756985195624284,
264
+ "acc_norm": 0.43,
265
+ "acc_norm_stderr": 0.049756985195624284
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6694214876033058,
269
+ "acc_stderr": 0.04294340845212094,
270
+ "acc_norm": 0.6694214876033058,
271
+ "acc_norm_stderr": 0.04294340845212094
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40789473684210525,
275
+ "acc_stderr": 0.03999309712777472,
276
+ "acc_norm": 0.40789473684210525,
277
+ "acc_norm_stderr": 0.03999309712777472
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.4035947712418301,
281
+ "acc_stderr": 0.019848280168401147,
282
+ "acc_norm": 0.4035947712418301,
283
+ "acc_norm_stderr": 0.019848280168401147
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.34397163120567376,
287
+ "acc_stderr": 0.028338017428611306,
288
+ "acc_norm": 0.34397163120567376,
289
+ "acc_norm_stderr": 0.028338017428611306
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.375,
293
+ "acc_stderr": 0.04595091388086298,
294
+ "acc_norm": 0.375,
295
+ "acc_norm_stderr": 0.04595091388086298
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.4305555555555556,
299
+ "acc_stderr": 0.03376922151252336,
300
+ "acc_norm": 0.4305555555555556,
301
+ "acc_norm_stderr": 0.03376922151252336
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.27150837988826815,
305
+ "acc_stderr": 0.014874252168095285,
306
+ "acc_norm": 0.27150837988826815,
307
+ "acc_norm_stderr": 0.014874252168095285
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.42,
311
+ "acc_stderr": 0.049604496374885836,
312
+ "acc_norm": 0.42,
313
+ "acc_norm_stderr": 0.049604496374885836
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.6,
317
+ "acc_stderr": 0.049236596391733084,
318
+ "acc_norm": 0.6,
319
+ "acc_norm_stderr": 0.049236596391733084
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3860294117647059,
323
+ "acc_stderr": 0.029573269134411127,
324
+ "acc_norm": 0.3860294117647059,
325
+ "acc_norm_stderr": 0.029573269134411127
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5551020408163265,
329
+ "acc_stderr": 0.031814251181977865,
330
+ "acc_norm": 0.5551020408163265,
331
+ "acc_norm_stderr": 0.031814251181977865
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5569620253164557,
335
+ "acc_stderr": 0.03233532777533484,
336
+ "acc_norm": 0.5569620253164557,
337
+ "acc_norm_stderr": 0.03233532777533484
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3259452411994785,
341
+ "acc_stderr": 0.01197150729498278,
342
+ "acc_norm": 0.3259452411994785,
343
+ "acc_norm_stderr": 0.01197150729498278
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4264705882352941,
347
+ "acc_stderr": 0.034711579079534254,
348
+ "acc_norm": 0.4264705882352941,
349
+ "acc_norm_stderr": 0.034711579079534254
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.46060606060606063,
353
+ "acc_stderr": 0.03892207016552013,
354
+ "acc_norm": 0.46060606060606063,
355
+ "acc_norm_stderr": 0.03892207016552013
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.28886168910648713,
359
+ "mc1_stderr": 0.0158663464013843,
360
+ "mc2": 0.46609022121434857,
361
+ "mc2_stderr": 0.015577378664296664
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5112160566706021,
365
+ "acc_stderr": 0.01718602846948929,
366
+ "acc_norm": 0.564344746162928,
367
+ "acc_norm_stderr": 0.017047415229476323
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.7.1",
436
+ "model_sha": "d9b8182352893418c91400683819f41d7be1e292",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.7.2/result_2023-11-30 06:16:02.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3660409556313993,
5
+ "acc_stderr": 0.01407722310847014,
6
+ "acc_norm": 0.41467576791808874,
7
+ "acc_norm_stderr": 0.014397070564409175
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.3996215893248357,
11
+ "acc_stderr": 0.004888194985997395,
12
+ "acc_norm": 0.5103565026887075,
13
+ "acc_norm_stderr": 0.004988710917169331
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.543859649122807,
17
+ "acc_stderr": 0.038200425866029654,
18
+ "acc_norm": 0.543859649122807,
19
+ "acc_norm_stderr": 0.038200425866029654
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5631067961165048,
23
+ "acc_stderr": 0.04911147107365777,
24
+ "acc_norm": 0.5631067961165048,
25
+ "acc_norm_stderr": 0.04911147107365777
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.5134099616858238,
29
+ "acc_stderr": 0.017873531736510354,
30
+ "acc_norm": 0.5134099616858238,
31
+ "acc_norm_stderr": 0.017873531736510354
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4074074074074074,
35
+ "acc_stderr": 0.04244633238353229,
36
+ "acc_norm": 0.4074074074074074,
37
+ "acc_norm_stderr": 0.04244633238353229
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.28,
41
+ "acc_stderr": 0.04512608598542128,
42
+ "acc_norm": 0.28,
43
+ "acc_norm_stderr": 0.04512608598542128
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.42127659574468085,
47
+ "acc_stderr": 0.03227834510146268,
48
+ "acc_norm": 0.42127659574468085,
49
+ "acc_norm_stderr": 0.03227834510146268
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.42168674698795183,
53
+ "acc_stderr": 0.03844453181770917,
54
+ "acc_norm": 0.42168674698795183,
55
+ "acc_norm_stderr": 0.03844453181770917
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4887459807073955,
59
+ "acc_stderr": 0.02839089739686352,
60
+ "acc_norm": 0.4887459807073955,
61
+ "acc_norm_stderr": 0.02839089739686352
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.452914798206278,
65
+ "acc_stderr": 0.033408675019233246,
66
+ "acc_norm": 0.452914798206278,
67
+ "acc_norm_stderr": 0.033408675019233246
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.4122137404580153,
71
+ "acc_stderr": 0.04317171194870255,
72
+ "acc_norm": 0.4122137404580153,
73
+ "acc_norm_stderr": 0.04317171194870255
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.41,
77
+ "acc_stderr": 0.049431107042371025,
78
+ "acc_norm": 0.41,
79
+ "acc_norm_stderr": 0.049431107042371025
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5808080808080808,
83
+ "acc_stderr": 0.03515520728670417,
84
+ "acc_norm": 0.5808080808080808,
85
+ "acc_norm_stderr": 0.03515520728670417
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4689655172413793,
89
+ "acc_stderr": 0.04158632762097828,
90
+ "acc_norm": 0.4689655172413793,
91
+ "acc_norm_stderr": 0.04158632762097828
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.2549019607843137,
95
+ "acc_stderr": 0.043364327079931785,
96
+ "acc_norm": 0.2549019607843137,
97
+ "acc_norm_stderr": 0.043364327079931785
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.48739495798319327,
101
+ "acc_stderr": 0.03246816765752174,
102
+ "acc_norm": 0.48739495798319327,
103
+ "acc_norm_stderr": 0.03246816765752174
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.45897435897435895,
107
+ "acc_stderr": 0.025265525491284295,
108
+ "acc_norm": 0.45897435897435895,
109
+ "acc_norm_stderr": 0.025265525491284295
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.56,
113
+ "acc_stderr": 0.04988876515698589,
114
+ "acc_norm": 0.56,
115
+ "acc_norm_stderr": 0.04988876515698589
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.33,
119
+ "acc_stderr": 0.04725815626252604,
120
+ "acc_norm": 0.33,
121
+ "acc_norm_stderr": 0.04725815626252604
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5092592592592593,
125
+ "acc_stderr": 0.04832853553437055,
126
+ "acc_norm": 0.5092592592592593,
127
+ "acc_norm_stderr": 0.04832853553437055
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4187192118226601,
131
+ "acc_stderr": 0.03471192860518468,
132
+ "acc_norm": 0.4187192118226601,
133
+ "acc_norm_stderr": 0.03471192860518468
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.47419354838709676,
137
+ "acc_stderr": 0.028406095057653315,
138
+ "acc_norm": 0.47419354838709676,
139
+ "acc_norm_stderr": 0.028406095057653315
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.7094017094017094,
143
+ "acc_stderr": 0.029745048572674078,
144
+ "acc_norm": 0.7094017094017094,
145
+ "acc_norm_stderr": 0.029745048572674078
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4528301886792453,
149
+ "acc_stderr": 0.03063562795796182,
150
+ "acc_norm": 0.4528301886792453,
151
+ "acc_norm_stderr": 0.03063562795796182
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4818181818181818,
155
+ "acc_stderr": 0.04785964010794916,
156
+ "acc_norm": 0.4818181818181818,
157
+ "acc_norm_stderr": 0.04785964010794916
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.34074074074074073,
161
+ "acc_stderr": 0.02889774874113115,
162
+ "acc_norm": 0.34074074074074073,
163
+ "acc_norm_stderr": 0.02889774874113115
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2582781456953642,
167
+ "acc_stderr": 0.035737053147634576,
168
+ "acc_norm": 0.2582781456953642,
169
+ "acc_norm_stderr": 0.035737053147634576
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5970149253731343,
173
+ "acc_stderr": 0.034683432951111266,
174
+ "acc_norm": 0.5970149253731343,
175
+ "acc_norm_stderr": 0.034683432951111266
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3988439306358382,
179
+ "acc_stderr": 0.03733626655383509,
180
+ "acc_norm": 0.3988439306358382,
181
+ "acc_norm_stderr": 0.03733626655383509
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.3333333333333333,
185
+ "acc_stderr": 0.0242785680243077,
186
+ "acc_norm": 0.3333333333333333,
187
+ "acc_norm_stderr": 0.0242785680243077
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.375,
191
+ "acc_stderr": 0.04048439222695598,
192
+ "acc_norm": 0.375,
193
+ "acc_norm_stderr": 0.04048439222695598
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.33,
197
+ "acc_stderr": 0.047258156262526045,
198
+ "acc_norm": 0.33,
199
+ "acc_norm_stderr": 0.047258156262526045
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.63,
203
+ "acc_stderr": 0.048523658709391,
204
+ "acc_norm": 0.63,
205
+ "acc_norm_stderr": 0.048523658709391
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5086705202312138,
209
+ "acc_stderr": 0.026915047355369804,
210
+ "acc_norm": 0.5086705202312138,
211
+ "acc_norm_stderr": 0.026915047355369804
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.4662576687116564,
215
+ "acc_stderr": 0.03919415545048409,
216
+ "acc_norm": 0.4662576687116564,
217
+ "acc_norm_stderr": 0.03919415545048409
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.46296296296296297,
221
+ "acc_stderr": 0.02774431344337654,
222
+ "acc_norm": 0.46296296296296297,
223
+ "acc_norm_stderr": 0.02774431344337654
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.25,
227
+ "acc_stderr": 0.04351941398892446,
228
+ "acc_norm": 0.25,
229
+ "acc_norm_stderr": 0.04351941398892446
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5492227979274611,
233
+ "acc_stderr": 0.035909109522355244,
234
+ "acc_norm": 0.5492227979274611,
235
+ "acc_norm_stderr": 0.035909109522355244
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2894736842105263,
239
+ "acc_stderr": 0.04266339443159394,
240
+ "acc_norm": 0.2894736842105263,
241
+ "acc_norm_stderr": 0.04266339443159394
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.4972477064220184,
245
+ "acc_stderr": 0.02143699835976532,
246
+ "acc_norm": 0.4972477064220184,
247
+ "acc_norm_stderr": 0.02143699835976532
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.3412698412698413,
251
+ "acc_stderr": 0.04240799327574924,
252
+ "acc_norm": 0.3412698412698413,
253
+ "acc_norm_stderr": 0.04240799327574924
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.5065359477124183,
257
+ "acc_stderr": 0.028627470550556047,
258
+ "acc_norm": 0.5065359477124183,
259
+ "acc_norm_stderr": 0.028627470550556047
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.47,
263
+ "acc_stderr": 0.05016135580465919,
264
+ "acc_norm": 0.47,
265
+ "acc_norm_stderr": 0.05016135580465919
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.7107438016528925,
269
+ "acc_stderr": 0.04139112727635464,
270
+ "acc_norm": 0.7107438016528925,
271
+ "acc_norm_stderr": 0.04139112727635464
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.40789473684210525,
275
+ "acc_stderr": 0.03999309712777471,
276
+ "acc_norm": 0.40789473684210525,
277
+ "acc_norm_stderr": 0.03999309712777471
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.40522875816993464,
281
+ "acc_stderr": 0.019861155193829153,
282
+ "acc_norm": 0.40522875816993464,
283
+ "acc_norm_stderr": 0.019861155193829153
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3404255319148936,
287
+ "acc_stderr": 0.028267657482650137,
288
+ "acc_norm": 0.3404255319148936,
289
+ "acc_norm_stderr": 0.028267657482650137
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.36607142857142855,
293
+ "acc_stderr": 0.0457237235873743,
294
+ "acc_norm": 0.36607142857142855,
295
+ "acc_norm_stderr": 0.0457237235873743
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.3333333333333333,
299
+ "acc_stderr": 0.03214952147802749,
300
+ "acc_norm": 0.3333333333333333,
301
+ "acc_norm_stderr": 0.03214952147802749
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.29608938547486036,
305
+ "acc_stderr": 0.015268677317602298,
306
+ "acc_norm": 0.29608938547486036,
307
+ "acc_norm_stderr": 0.015268677317602298
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.36,
311
+ "acc_stderr": 0.04824181513244218,
312
+ "acc_norm": 0.36,
313
+ "acc_norm_stderr": 0.04824181513244218
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.58,
317
+ "acc_stderr": 0.049604496374885836,
318
+ "acc_norm": 0.58,
319
+ "acc_norm_stderr": 0.049604496374885836
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.41911764705882354,
323
+ "acc_stderr": 0.029972807170464626,
324
+ "acc_norm": 0.41911764705882354,
325
+ "acc_norm_stderr": 0.029972807170464626
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5265306122448979,
329
+ "acc_stderr": 0.03196412734523272,
330
+ "acc_norm": 0.5265306122448979,
331
+ "acc_norm_stderr": 0.03196412734523272
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.6160337552742616,
335
+ "acc_stderr": 0.031658678064106674,
336
+ "acc_norm": 0.6160337552742616,
337
+ "acc_norm_stderr": 0.031658678064106674
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3435462842242503,
341
+ "acc_stderr": 0.012128961174190156,
342
+ "acc_norm": 0.3435462842242503,
343
+ "acc_norm_stderr": 0.012128961174190156
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.47549019607843135,
347
+ "acc_stderr": 0.035050931943487976,
348
+ "acc_norm": 0.47549019607843135,
349
+ "acc_norm_stderr": 0.035050931943487976
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.46060606060606063,
353
+ "acc_stderr": 0.03892207016552013,
354
+ "acc_norm": 0.46060606060606063,
355
+ "acc_norm_stderr": 0.03892207016552013
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2717258261933905,
359
+ "mc1_stderr": 0.015572840452875835,
360
+ "mc2": 0.4265196372341518,
361
+ "mc2_stderr": 0.015607268833983732
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.3707201889020071,
365
+ "acc_stderr": 0.0166058012892126,
366
+ "acc_norm": 0.3955135773317591,
367
+ "acc_norm_stderr": 0.01681081590220604
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.7.2",
436
+ "model_sha": "dbbe8fc186f5371e9ca608b9650e387b20de7ae3",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-7b-it-v1.7.3/result_2023-12-04 23:53:23.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.34215017064846415,
5
+ "acc_stderr": 0.013864152159177278,
6
+ "acc_norm": 0.3771331058020478,
7
+ "acc_norm_stderr": 0.014163366896192598
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.37880900219079866,
11
+ "acc_stderr": 0.00484099059349469,
12
+ "acc_norm": 0.4870543716391157,
13
+ "acc_norm_stderr": 0.004988108663179766
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.5029239766081871,
17
+ "acc_stderr": 0.03834759370936839,
18
+ "acc_norm": 0.5029239766081871,
19
+ "acc_norm_stderr": 0.03834759370936839
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5339805825242718,
23
+ "acc_stderr": 0.0493929144727348,
24
+ "acc_norm": 0.5339805825242718,
25
+ "acc_norm_stderr": 0.0493929144727348
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4623243933588761,
29
+ "acc_stderr": 0.017829131764287198,
30
+ "acc_norm": 0.4623243933588761,
31
+ "acc_norm_stderr": 0.017829131764287198
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.362962962962963,
35
+ "acc_stderr": 0.041539484047424,
36
+ "acc_norm": 0.362962962962963,
37
+ "acc_norm_stderr": 0.041539484047424
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.26,
41
+ "acc_stderr": 0.04408440022768079,
42
+ "acc_norm": 0.26,
43
+ "acc_norm_stderr": 0.04408440022768079
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.3404255319148936,
47
+ "acc_stderr": 0.03097669299853441,
48
+ "acc_norm": 0.3404255319148936,
49
+ "acc_norm_stderr": 0.03097669299853441
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.3192771084337349,
53
+ "acc_stderr": 0.03629335329947859,
54
+ "acc_norm": 0.3192771084337349,
55
+ "acc_norm_stderr": 0.03629335329947859
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4437299035369775,
59
+ "acc_stderr": 0.02821768355665231,
60
+ "acc_norm": 0.4437299035369775,
61
+ "acc_norm_stderr": 0.02821768355665231
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.37668161434977576,
65
+ "acc_stderr": 0.03252113489929186,
66
+ "acc_norm": 0.37668161434977576,
67
+ "acc_norm_stderr": 0.03252113489929186
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.40458015267175573,
71
+ "acc_stderr": 0.043046937953806645,
72
+ "acc_norm": 0.40458015267175573,
73
+ "acc_norm_stderr": 0.043046937953806645
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.33,
77
+ "acc_stderr": 0.047258156262526045,
78
+ "acc_norm": 0.33,
79
+ "acc_norm_stderr": 0.047258156262526045
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5353535353535354,
83
+ "acc_stderr": 0.03553436368828061,
84
+ "acc_norm": 0.5353535353535354,
85
+ "acc_norm_stderr": 0.03553436368828061
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.42758620689655175,
89
+ "acc_stderr": 0.04122737111370333,
90
+ "acc_norm": 0.42758620689655175,
91
+ "acc_norm_stderr": 0.04122737111370333
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.2647058823529412,
95
+ "acc_stderr": 0.04389869956808778,
96
+ "acc_norm": 0.2647058823529412,
97
+ "acc_norm_stderr": 0.04389869956808778
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.49159663865546216,
101
+ "acc_stderr": 0.03247390276569669,
102
+ "acc_norm": 0.49159663865546216,
103
+ "acc_norm_stderr": 0.03247390276569669
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.3871794871794872,
107
+ "acc_stderr": 0.02469721693087895,
108
+ "acc_norm": 0.3871794871794872,
109
+ "acc_norm_stderr": 0.02469721693087895
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.5,
113
+ "acc_stderr": 0.050251890762960605,
114
+ "acc_norm": 0.5,
115
+ "acc_norm_stderr": 0.050251890762960605
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.31,
119
+ "acc_stderr": 0.04648231987117316,
120
+ "acc_norm": 0.31,
121
+ "acc_norm_stderr": 0.04648231987117316
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.4722222222222222,
125
+ "acc_stderr": 0.04826217294139894,
126
+ "acc_norm": 0.4722222222222222,
127
+ "acc_norm_stderr": 0.04826217294139894
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.3842364532019704,
131
+ "acc_stderr": 0.034223985656575515,
132
+ "acc_norm": 0.3842364532019704,
133
+ "acc_norm_stderr": 0.034223985656575515
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4161290322580645,
137
+ "acc_stderr": 0.028040981380761543,
138
+ "acc_norm": 0.4161290322580645,
139
+ "acc_norm_stderr": 0.028040981380761543
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.6923076923076923,
143
+ "acc_stderr": 0.03023638994217309,
144
+ "acc_norm": 0.6923076923076923,
145
+ "acc_norm_stderr": 0.03023638994217309
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.44150943396226416,
149
+ "acc_stderr": 0.030561590426731833,
150
+ "acc_norm": 0.44150943396226416,
151
+ "acc_norm_stderr": 0.030561590426731833
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.4909090909090909,
155
+ "acc_stderr": 0.04788339768702862,
156
+ "acc_norm": 0.4909090909090909,
157
+ "acc_norm_stderr": 0.04788339768702862
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.3,
161
+ "acc_stderr": 0.027940457136228405,
162
+ "acc_norm": 0.3,
163
+ "acc_norm_stderr": 0.027940457136228405
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.33112582781456956,
167
+ "acc_stderr": 0.038425817186598696,
168
+ "acc_norm": 0.33112582781456956,
169
+ "acc_norm_stderr": 0.038425817186598696
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5920398009950248,
173
+ "acc_stderr": 0.03475116365194092,
174
+ "acc_norm": 0.5920398009950248,
175
+ "acc_norm_stderr": 0.03475116365194092
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3236994219653179,
179
+ "acc_stderr": 0.035676037996391706,
180
+ "acc_norm": 0.3236994219653179,
181
+ "acc_norm_stderr": 0.035676037996391706
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.30687830687830686,
185
+ "acc_stderr": 0.02375292871211214,
186
+ "acc_norm": 0.30687830687830686,
187
+ "acc_norm_stderr": 0.02375292871211214
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.375,
191
+ "acc_stderr": 0.04048439222695598,
192
+ "acc_norm": 0.375,
193
+ "acc_norm_stderr": 0.04048439222695598
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.37,
197
+ "acc_stderr": 0.048523658709391,
198
+ "acc_norm": 0.37,
199
+ "acc_norm_stderr": 0.048523658709391
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.63,
203
+ "acc_stderr": 0.048523658709391,
204
+ "acc_norm": 0.63,
205
+ "acc_norm_stderr": 0.048523658709391
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.44508670520231214,
209
+ "acc_stderr": 0.026756255129663765,
210
+ "acc_norm": 0.44508670520231214,
211
+ "acc_norm_stderr": 0.026756255129663765
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.49079754601226994,
215
+ "acc_stderr": 0.03927705600787443,
216
+ "acc_norm": 0.49079754601226994,
217
+ "acc_norm_stderr": 0.03927705600787443
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.41358024691358025,
221
+ "acc_stderr": 0.02740204204026994,
222
+ "acc_norm": 0.41358024691358025,
223
+ "acc_norm_stderr": 0.02740204204026994
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.28,
227
+ "acc_stderr": 0.04512608598542127,
228
+ "acc_norm": 0.28,
229
+ "acc_norm_stderr": 0.04512608598542127
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.46113989637305697,
233
+ "acc_stderr": 0.035975244117345775,
234
+ "acc_norm": 0.46113989637305697,
235
+ "acc_norm_stderr": 0.035975244117345775
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.2894736842105263,
239
+ "acc_stderr": 0.04266339443159394,
240
+ "acc_norm": 0.2894736842105263,
241
+ "acc_norm_stderr": 0.04266339443159394
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.4917431192660551,
245
+ "acc_stderr": 0.021434399918214334,
246
+ "acc_norm": 0.4917431192660551,
247
+ "acc_norm_stderr": 0.021434399918214334
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.38095238095238093,
251
+ "acc_stderr": 0.04343525428949098,
252
+ "acc_norm": 0.38095238095238093,
253
+ "acc_norm_stderr": 0.04343525428949098
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.45751633986928103,
257
+ "acc_stderr": 0.028526383452142628,
258
+ "acc_norm": 0.45751633986928103,
259
+ "acc_norm_stderr": 0.028526383452142628
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.48,
263
+ "acc_stderr": 0.050211673156867795,
264
+ "acc_norm": 0.48,
265
+ "acc_norm_stderr": 0.050211673156867795
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6363636363636364,
269
+ "acc_stderr": 0.043913262867240704,
270
+ "acc_norm": 0.6363636363636364,
271
+ "acc_norm_stderr": 0.043913262867240704
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.4407894736842105,
275
+ "acc_stderr": 0.04040311062490436,
276
+ "acc_norm": 0.4407894736842105,
277
+ "acc_norm_stderr": 0.04040311062490436
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.3741830065359477,
281
+ "acc_stderr": 0.01957695312208883,
282
+ "acc_norm": 0.3741830065359477,
283
+ "acc_norm_stderr": 0.01957695312208883
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3120567375886525,
287
+ "acc_stderr": 0.02764012054516993,
288
+ "acc_norm": 0.3120567375886525,
289
+ "acc_norm_stderr": 0.02764012054516993
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.3392857142857143,
293
+ "acc_stderr": 0.04493949068613539,
294
+ "acc_norm": 0.3392857142857143,
295
+ "acc_norm_stderr": 0.04493949068613539
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.4074074074074074,
299
+ "acc_stderr": 0.03350991604696043,
300
+ "acc_norm": 0.4074074074074074,
301
+ "acc_norm_stderr": 0.03350991604696043
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.22346368715083798,
305
+ "acc_stderr": 0.013932068638579771,
306
+ "acc_norm": 0.22346368715083798,
307
+ "acc_norm_stderr": 0.013932068638579771
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.36,
311
+ "acc_stderr": 0.048241815132442176,
312
+ "acc_norm": 0.36,
313
+ "acc_norm_stderr": 0.048241815132442176
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.61,
317
+ "acc_stderr": 0.04902071300001974,
318
+ "acc_norm": 0.61,
319
+ "acc_norm_stderr": 0.04902071300001974
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.2757352941176471,
323
+ "acc_stderr": 0.02714627193662517,
324
+ "acc_norm": 0.2757352941176471,
325
+ "acc_norm_stderr": 0.02714627193662517
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.3836734693877551,
329
+ "acc_stderr": 0.031130880396235926,
330
+ "acc_norm": 0.3836734693877551,
331
+ "acc_norm_stderr": 0.031130880396235926
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5232067510548524,
335
+ "acc_stderr": 0.032512152011410174,
336
+ "acc_norm": 0.5232067510548524,
337
+ "acc_norm_stderr": 0.032512152011410174
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3324641460234681,
341
+ "acc_stderr": 0.012032022332260514,
342
+ "acc_norm": 0.3324641460234681,
343
+ "acc_norm_stderr": 0.012032022332260514
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.43137254901960786,
347
+ "acc_stderr": 0.03476099060501636,
348
+ "acc_norm": 0.43137254901960786,
349
+ "acc_norm_stderr": 0.03476099060501636
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.4484848484848485,
353
+ "acc_stderr": 0.03883565977956929,
354
+ "acc_norm": 0.4484848484848485,
355
+ "acc_norm_stderr": 0.03883565977956929
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.26560587515299877,
359
+ "mc1_stderr": 0.0154610276272536,
360
+ "mc2": 0.42128232388140774,
361
+ "mc2_stderr": 0.015666521111746597
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.45336481700118064,
365
+ "acc_stderr": 0.017115418225226872,
366
+ "acc_norm": 0.49940968122786306,
367
+ "acc_norm_stderr": 0.017190342123448662
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-7b-it-v1.7.3",
436
+ "model_sha": "485019f55449b1adffd92e017a85bfe922cac2dd",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-ko-7b-it-v2.0.0/result_2023-12-23 00:56:47.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.34812286689419797,
5
+ "acc_stderr": 0.013921008595179344,
6
+ "acc_norm": 0.39078498293515357,
7
+ "acc_norm_stderr": 0.014258563880513778
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.37970523800039835,
11
+ "acc_stderr": 0.0048432163250902655,
12
+ "acc_norm": 0.4901414060944035,
13
+ "acc_norm_stderr": 0.004988811384747425
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.49122807017543857,
17
+ "acc_stderr": 0.038342347441649924,
18
+ "acc_norm": 0.49122807017543857,
19
+ "acc_norm_stderr": 0.038342347441649924
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5728155339805825,
23
+ "acc_stderr": 0.04897957737781168,
24
+ "acc_norm": 0.5728155339805825,
25
+ "acc_norm_stderr": 0.04897957737781168
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.4725415070242657,
29
+ "acc_stderr": 0.017852981266633955,
30
+ "acc_norm": 0.4725415070242657,
31
+ "acc_norm_stderr": 0.017852981266633955
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4222222222222222,
35
+ "acc_stderr": 0.04266763404099582,
36
+ "acc_norm": 0.4222222222222222,
37
+ "acc_norm_stderr": 0.04266763404099582
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.27,
41
+ "acc_stderr": 0.0446196043338474,
42
+ "acc_norm": 0.27,
43
+ "acc_norm_stderr": 0.0446196043338474
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.39574468085106385,
47
+ "acc_stderr": 0.03196758697835362,
48
+ "acc_norm": 0.39574468085106385,
49
+ "acc_norm_stderr": 0.03196758697835362
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.41566265060240964,
53
+ "acc_stderr": 0.03836722176598052,
54
+ "acc_norm": 0.41566265060240964,
55
+ "acc_norm_stderr": 0.03836722176598052
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4662379421221865,
59
+ "acc_stderr": 0.028333277109562786,
60
+ "acc_norm": 0.4662379421221865,
61
+ "acc_norm_stderr": 0.028333277109562786
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.452914798206278,
65
+ "acc_stderr": 0.03340867501923325,
66
+ "acc_norm": 0.452914798206278,
67
+ "acc_norm_stderr": 0.03340867501923325
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.46564885496183206,
71
+ "acc_stderr": 0.04374928560599738,
72
+ "acc_norm": 0.46564885496183206,
73
+ "acc_norm_stderr": 0.04374928560599738
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.36,
77
+ "acc_stderr": 0.04824181513244218,
78
+ "acc_norm": 0.36,
79
+ "acc_norm_stderr": 0.04824181513244218
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5555555555555556,
83
+ "acc_stderr": 0.035402943770953675,
84
+ "acc_norm": 0.5555555555555556,
85
+ "acc_norm_stderr": 0.035402943770953675
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4689655172413793,
89
+ "acc_stderr": 0.04158632762097828,
90
+ "acc_norm": 0.4689655172413793,
91
+ "acc_norm_stderr": 0.04158632762097828
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.24509803921568626,
95
+ "acc_stderr": 0.04280105837364395,
96
+ "acc_norm": 0.24509803921568626,
97
+ "acc_norm_stderr": 0.04280105837364395
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.5168067226890757,
101
+ "acc_stderr": 0.03246013680375308,
102
+ "acc_norm": 0.5168067226890757,
103
+ "acc_norm_stderr": 0.03246013680375308
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4641025641025641,
107
+ "acc_stderr": 0.025285585990017834,
108
+ "acc_norm": 0.4641025641025641,
109
+ "acc_norm_stderr": 0.025285585990017834
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.58,
113
+ "acc_stderr": 0.04960449637488583,
114
+ "acc_norm": 0.58,
115
+ "acc_norm_stderr": 0.04960449637488583
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.33,
119
+ "acc_stderr": 0.04725815626252604,
120
+ "acc_norm": 0.33,
121
+ "acc_norm_stderr": 0.04725815626252604
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5555555555555556,
125
+ "acc_stderr": 0.04803752235190192,
126
+ "acc_norm": 0.5555555555555556,
127
+ "acc_norm_stderr": 0.04803752235190192
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.42857142857142855,
131
+ "acc_stderr": 0.03481904844438803,
132
+ "acc_norm": 0.42857142857142855,
133
+ "acc_norm_stderr": 0.03481904844438803
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.45806451612903226,
137
+ "acc_stderr": 0.028343787250540636,
138
+ "acc_norm": 0.45806451612903226,
139
+ "acc_norm_stderr": 0.028343787250540636
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.7307692307692307,
143
+ "acc_stderr": 0.029058588303748845,
144
+ "acc_norm": 0.7307692307692307,
145
+ "acc_norm_stderr": 0.029058588303748845
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.44528301886792454,
149
+ "acc_stderr": 0.030588052974270655,
150
+ "acc_norm": 0.44528301886792454,
151
+ "acc_norm_stderr": 0.030588052974270655
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5363636363636364,
155
+ "acc_stderr": 0.04776449162396197,
156
+ "acc_norm": 0.5363636363636364,
157
+ "acc_norm_stderr": 0.04776449162396197
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.3333333333333333,
161
+ "acc_stderr": 0.028742040903948485,
162
+ "acc_norm": 0.3333333333333333,
163
+ "acc_norm_stderr": 0.028742040903948485
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.31788079470198677,
167
+ "acc_stderr": 0.038020397601079024,
168
+ "acc_norm": 0.31788079470198677,
169
+ "acc_norm_stderr": 0.038020397601079024
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.6119402985074627,
173
+ "acc_stderr": 0.03445789964362749,
174
+ "acc_norm": 0.6119402985074627,
175
+ "acc_norm_stderr": 0.03445789964362749
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.3583815028901734,
179
+ "acc_stderr": 0.036563436533531585,
180
+ "acc_norm": 0.3583815028901734,
181
+ "acc_norm_stderr": 0.036563436533531585
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.35978835978835977,
185
+ "acc_stderr": 0.024718075944129274,
186
+ "acc_norm": 0.35978835978835977,
187
+ "acc_norm_stderr": 0.024718075944129274
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.4097222222222222,
191
+ "acc_stderr": 0.04112490974670787,
192
+ "acc_norm": 0.4097222222222222,
193
+ "acc_norm_stderr": 0.04112490974670787
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.38,
197
+ "acc_stderr": 0.04878317312145633,
198
+ "acc_norm": 0.38,
199
+ "acc_norm_stderr": 0.04878317312145633
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.57,
203
+ "acc_stderr": 0.04975698519562426,
204
+ "acc_norm": 0.57,
205
+ "acc_norm_stderr": 0.04975698519562426
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.5317919075144508,
209
+ "acc_stderr": 0.02686462436675666,
210
+ "acc_norm": 0.5317919075144508,
211
+ "acc_norm_stderr": 0.02686462436675666
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.5030674846625767,
215
+ "acc_stderr": 0.03928297078179663,
216
+ "acc_norm": 0.5030674846625767,
217
+ "acc_norm_stderr": 0.03928297078179663
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.42901234567901236,
221
+ "acc_stderr": 0.027538925613470867,
222
+ "acc_norm": 0.42901234567901236,
223
+ "acc_norm_stderr": 0.027538925613470867
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.32,
227
+ "acc_stderr": 0.04688261722621504,
228
+ "acc_norm": 0.32,
229
+ "acc_norm_stderr": 0.04688261722621504
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5233160621761658,
233
+ "acc_stderr": 0.03604513672442202,
234
+ "acc_norm": 0.5233160621761658,
235
+ "acc_norm_stderr": 0.03604513672442202
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.30701754385964913,
239
+ "acc_stderr": 0.043391383225798594,
240
+ "acc_norm": 0.30701754385964913,
241
+ "acc_norm_stderr": 0.043391383225798594
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.4917431192660551,
245
+ "acc_stderr": 0.021434399918214334,
246
+ "acc_norm": 0.4917431192660551,
247
+ "acc_norm_stderr": 0.021434399918214334
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.42063492063492064,
251
+ "acc_stderr": 0.04415438226743744,
252
+ "acc_norm": 0.42063492063492064,
253
+ "acc_norm_stderr": 0.04415438226743744
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.48366013071895425,
257
+ "acc_stderr": 0.028614624752805407,
258
+ "acc_norm": 0.48366013071895425,
259
+ "acc_norm_stderr": 0.028614624752805407
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.47,
263
+ "acc_stderr": 0.050161355804659205,
264
+ "acc_norm": 0.47,
265
+ "acc_norm_stderr": 0.050161355804659205
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.7603305785123967,
269
+ "acc_stderr": 0.03896878985070417,
270
+ "acc_norm": 0.7603305785123967,
271
+ "acc_norm_stderr": 0.03896878985070417
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.3881578947368421,
275
+ "acc_stderr": 0.03965842097512744,
276
+ "acc_norm": 0.3881578947368421,
277
+ "acc_norm_stderr": 0.03965842097512744
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.4150326797385621,
281
+ "acc_stderr": 0.01993362777685742,
282
+ "acc_norm": 0.4150326797385621,
283
+ "acc_norm_stderr": 0.01993362777685742
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.34397163120567376,
287
+ "acc_stderr": 0.028338017428611317,
288
+ "acc_norm": 0.34397163120567376,
289
+ "acc_norm_stderr": 0.028338017428611317
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.41964285714285715,
293
+ "acc_stderr": 0.04684099321077106,
294
+ "acc_norm": 0.41964285714285715,
295
+ "acc_norm_stderr": 0.04684099321077106
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.4074074074074074,
299
+ "acc_stderr": 0.03350991604696043,
300
+ "acc_norm": 0.4074074074074074,
301
+ "acc_norm_stderr": 0.03350991604696043
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.3039106145251397,
305
+ "acc_stderr": 0.015382845587584518,
306
+ "acc_norm": 0.3039106145251397,
307
+ "acc_norm_stderr": 0.015382845587584518
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.45,
311
+ "acc_stderr": 0.05,
312
+ "acc_norm": 0.45,
313
+ "acc_norm_stderr": 0.05
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.6,
317
+ "acc_stderr": 0.049236596391733084,
318
+ "acc_norm": 0.6,
319
+ "acc_norm_stderr": 0.049236596391733084
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.3492647058823529,
323
+ "acc_stderr": 0.028959755196824866,
324
+ "acc_norm": 0.3492647058823529,
325
+ "acc_norm_stderr": 0.028959755196824866
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5673469387755102,
329
+ "acc_stderr": 0.031717528240626645,
330
+ "acc_norm": 0.5673469387755102,
331
+ "acc_norm_stderr": 0.031717528240626645
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5569620253164557,
335
+ "acc_stderr": 0.03233532777533484,
336
+ "acc_norm": 0.5569620253164557,
337
+ "acc_norm_stderr": 0.03233532777533484
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.32790091264667537,
341
+ "acc_stderr": 0.011989936640666544,
342
+ "acc_norm": 0.32790091264667537,
343
+ "acc_norm_stderr": 0.011989936640666544
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.4215686274509804,
347
+ "acc_stderr": 0.03465868196380758,
348
+ "acc_norm": 0.4215686274509804,
349
+ "acc_norm_stderr": 0.03465868196380758
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.45454545454545453,
353
+ "acc_stderr": 0.03888176921674098,
354
+ "acc_norm": 0.45454545454545453,
355
+ "acc_norm_stderr": 0.03888176921674098
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.26438188494492043,
359
+ "mc1_stderr": 0.015438211119522512,
360
+ "mc2": 0.43306568977437526,
361
+ "mc2_stderr": 0.015345933860590263
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.4935064935064935,
365
+ "acc_stderr": 0.01718890435907731,
366
+ "acc_norm": 0.5348288075560803,
367
+ "acc_norm_stderr": 0.017148598015747425
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-ko-7b-it-v2.0.0",
436
+ "model_sha": "344b5f989128dc9f7a1bb3a1e8bbfe8b50a4159d",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }
BM-K/mistral-ko-7b-it-v2.0.1/result_2023-12-26 12:34:11.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "harness|ko_arc_challenge|25": {
4
+ "acc": 0.3395904436860068,
5
+ "acc_stderr": 0.013839039762820164,
6
+ "acc_norm": 0.378839590443686,
7
+ "acc_norm_stderr": 0.014175915490000322
8
+ },
9
+ "harness|ko_hellaswag|10": {
10
+ "acc": 0.37890858394742083,
11
+ "acc_stderr": 0.004841238763529383,
12
+ "acc_norm": 0.4918342959569807,
13
+ "acc_norm_stderr": 0.004989115942570063
14
+ },
15
+ "harness|ko_mmlu_world_religions|5": {
16
+ "acc": 0.4853801169590643,
17
+ "acc_stderr": 0.038331852752130205,
18
+ "acc_norm": 0.4853801169590643,
19
+ "acc_norm_stderr": 0.038331852752130205
20
+ },
21
+ "harness|ko_mmlu_management|5": {
22
+ "acc": 0.5728155339805825,
23
+ "acc_stderr": 0.04897957737781168,
24
+ "acc_norm": 0.5728155339805825,
25
+ "acc_norm_stderr": 0.04897957737781168
26
+ },
27
+ "harness|ko_mmlu_miscellaneous|5": {
28
+ "acc": 0.47509578544061304,
29
+ "acc_stderr": 0.017857770704901018,
30
+ "acc_norm": 0.47509578544061304,
31
+ "acc_norm_stderr": 0.017857770704901018
32
+ },
33
+ "harness|ko_mmlu_anatomy|5": {
34
+ "acc": 0.4,
35
+ "acc_stderr": 0.04232073695151589,
36
+ "acc_norm": 0.4,
37
+ "acc_norm_stderr": 0.04232073695151589
38
+ },
39
+ "harness|ko_mmlu_abstract_algebra|5": {
40
+ "acc": 0.23,
41
+ "acc_stderr": 0.04229525846816505,
42
+ "acc_norm": 0.23,
43
+ "acc_norm_stderr": 0.04229525846816505
44
+ },
45
+ "harness|ko_mmlu_conceptual_physics|5": {
46
+ "acc": 0.40425531914893614,
47
+ "acc_stderr": 0.03208115750788684,
48
+ "acc_norm": 0.40425531914893614,
49
+ "acc_norm_stderr": 0.03208115750788684
50
+ },
51
+ "harness|ko_mmlu_virology|5": {
52
+ "acc": 0.39759036144578314,
53
+ "acc_stderr": 0.038099730845402184,
54
+ "acc_norm": 0.39759036144578314,
55
+ "acc_norm_stderr": 0.038099730845402184
56
+ },
57
+ "harness|ko_mmlu_philosophy|5": {
58
+ "acc": 0.4919614147909968,
59
+ "acc_stderr": 0.028394421370984545,
60
+ "acc_norm": 0.4919614147909968,
61
+ "acc_norm_stderr": 0.028394421370984545
62
+ },
63
+ "harness|ko_mmlu_human_aging|5": {
64
+ "acc": 0.4484304932735426,
65
+ "acc_stderr": 0.03337883736255099,
66
+ "acc_norm": 0.4484304932735426,
67
+ "acc_norm_stderr": 0.03337883736255099
68
+ },
69
+ "harness|ko_mmlu_human_sexuality|5": {
70
+ "acc": 0.46564885496183206,
71
+ "acc_stderr": 0.04374928560599738,
72
+ "acc_norm": 0.46564885496183206,
73
+ "acc_norm_stderr": 0.04374928560599738
74
+ },
75
+ "harness|ko_mmlu_medical_genetics|5": {
76
+ "acc": 0.4,
77
+ "acc_stderr": 0.049236596391733084,
78
+ "acc_norm": 0.4,
79
+ "acc_norm_stderr": 0.049236596391733084
80
+ },
81
+ "harness|ko_mmlu_high_school_geography|5": {
82
+ "acc": 0.5454545454545454,
83
+ "acc_stderr": 0.03547601494006938,
84
+ "acc_norm": 0.5454545454545454,
85
+ "acc_norm_stderr": 0.03547601494006938
86
+ },
87
+ "harness|ko_mmlu_electrical_engineering|5": {
88
+ "acc": 0.4689655172413793,
89
+ "acc_stderr": 0.04158632762097828,
90
+ "acc_norm": 0.4689655172413793,
91
+ "acc_norm_stderr": 0.04158632762097828
92
+ },
93
+ "harness|ko_mmlu_college_physics|5": {
94
+ "acc": 0.2647058823529412,
95
+ "acc_stderr": 0.04389869956808777,
96
+ "acc_norm": 0.2647058823529412,
97
+ "acc_norm_stderr": 0.04389869956808777
98
+ },
99
+ "harness|ko_mmlu_high_school_microeconomics|5": {
100
+ "acc": 0.5504201680672269,
101
+ "acc_stderr": 0.03231293497137707,
102
+ "acc_norm": 0.5504201680672269,
103
+ "acc_norm_stderr": 0.03231293497137707
104
+ },
105
+ "harness|ko_mmlu_high_school_macroeconomics|5": {
106
+ "acc": 0.4794871794871795,
107
+ "acc_stderr": 0.025329663163489943,
108
+ "acc_norm": 0.4794871794871795,
109
+ "acc_norm_stderr": 0.025329663163489943
110
+ },
111
+ "harness|ko_mmlu_computer_security|5": {
112
+ "acc": 0.57,
113
+ "acc_stderr": 0.04975698519562429,
114
+ "acc_norm": 0.57,
115
+ "acc_norm_stderr": 0.04975698519562429
116
+ },
117
+ "harness|ko_mmlu_global_facts|5": {
118
+ "acc": 0.34,
119
+ "acc_stderr": 0.04760952285695235,
120
+ "acc_norm": 0.34,
121
+ "acc_norm_stderr": 0.04760952285695235
122
+ },
123
+ "harness|ko_mmlu_jurisprudence|5": {
124
+ "acc": 0.5462962962962963,
125
+ "acc_stderr": 0.04812917324536823,
126
+ "acc_norm": 0.5462962962962963,
127
+ "acc_norm_stderr": 0.04812917324536823
128
+ },
129
+ "harness|ko_mmlu_high_school_chemistry|5": {
130
+ "acc": 0.4482758620689655,
131
+ "acc_stderr": 0.034991131376767445,
132
+ "acc_norm": 0.4482758620689655,
133
+ "acc_norm_stderr": 0.034991131376767445
134
+ },
135
+ "harness|ko_mmlu_high_school_biology|5": {
136
+ "acc": 0.4838709677419355,
137
+ "acc_stderr": 0.028429203176724555,
138
+ "acc_norm": 0.4838709677419355,
139
+ "acc_norm_stderr": 0.028429203176724555
140
+ },
141
+ "harness|ko_mmlu_marketing|5": {
142
+ "acc": 0.7435897435897436,
143
+ "acc_stderr": 0.028605953702004243,
144
+ "acc_norm": 0.7435897435897436,
145
+ "acc_norm_stderr": 0.028605953702004243
146
+ },
147
+ "harness|ko_mmlu_clinical_knowledge|5": {
148
+ "acc": 0.4377358490566038,
149
+ "acc_stderr": 0.030533338430467512,
150
+ "acc_norm": 0.4377358490566038,
151
+ "acc_norm_stderr": 0.030533338430467512
152
+ },
153
+ "harness|ko_mmlu_public_relations|5": {
154
+ "acc": 0.5,
155
+ "acc_stderr": 0.04789131426105757,
156
+ "acc_norm": 0.5,
157
+ "acc_norm_stderr": 0.04789131426105757
158
+ },
159
+ "harness|ko_mmlu_high_school_mathematics|5": {
160
+ "acc": 0.32222222222222224,
161
+ "acc_stderr": 0.028493465091028593,
162
+ "acc_norm": 0.32222222222222224,
163
+ "acc_norm_stderr": 0.028493465091028593
164
+ },
165
+ "harness|ko_mmlu_high_school_physics|5": {
166
+ "acc": 0.2847682119205298,
167
+ "acc_stderr": 0.03684881521389024,
168
+ "acc_norm": 0.2847682119205298,
169
+ "acc_norm_stderr": 0.03684881521389024
170
+ },
171
+ "harness|ko_mmlu_sociology|5": {
172
+ "acc": 0.5771144278606966,
173
+ "acc_stderr": 0.034932317774212816,
174
+ "acc_norm": 0.5771144278606966,
175
+ "acc_norm_stderr": 0.034932317774212816
176
+ },
177
+ "harness|ko_mmlu_college_medicine|5": {
178
+ "acc": 0.4046242774566474,
179
+ "acc_stderr": 0.03742461193887248,
180
+ "acc_norm": 0.4046242774566474,
181
+ "acc_norm_stderr": 0.03742461193887248
182
+ },
183
+ "harness|ko_mmlu_elementary_mathematics|5": {
184
+ "acc": 0.36243386243386244,
185
+ "acc_stderr": 0.024757473902752045,
186
+ "acc_norm": 0.36243386243386244,
187
+ "acc_norm_stderr": 0.024757473902752045
188
+ },
189
+ "harness|ko_mmlu_college_biology|5": {
190
+ "acc": 0.3611111111111111,
191
+ "acc_stderr": 0.040166600304512336,
192
+ "acc_norm": 0.3611111111111111,
193
+ "acc_norm_stderr": 0.040166600304512336
194
+ },
195
+ "harness|ko_mmlu_college_chemistry|5": {
196
+ "acc": 0.43,
197
+ "acc_stderr": 0.049756985195624284,
198
+ "acc_norm": 0.43,
199
+ "acc_norm_stderr": 0.049756985195624284
200
+ },
201
+ "harness|ko_mmlu_us_foreign_policy|5": {
202
+ "acc": 0.57,
203
+ "acc_stderr": 0.04975698519562426,
204
+ "acc_norm": 0.57,
205
+ "acc_norm_stderr": 0.04975698519562426
206
+ },
207
+ "harness|ko_mmlu_moral_disputes|5": {
208
+ "acc": 0.4913294797687861,
209
+ "acc_stderr": 0.026915047355369804,
210
+ "acc_norm": 0.4913294797687861,
211
+ "acc_norm_stderr": 0.026915047355369804
212
+ },
213
+ "harness|ko_mmlu_logical_fallacies|5": {
214
+ "acc": 0.50920245398773,
215
+ "acc_stderr": 0.03927705600787443,
216
+ "acc_norm": 0.50920245398773,
217
+ "acc_norm_stderr": 0.03927705600787443
218
+ },
219
+ "harness|ko_mmlu_prehistory|5": {
220
+ "acc": 0.4444444444444444,
221
+ "acc_stderr": 0.027648477877413317,
222
+ "acc_norm": 0.4444444444444444,
223
+ "acc_norm_stderr": 0.027648477877413317
224
+ },
225
+ "harness|ko_mmlu_college_mathematics|5": {
226
+ "acc": 0.34,
227
+ "acc_stderr": 0.04760952285695235,
228
+ "acc_norm": 0.34,
229
+ "acc_norm_stderr": 0.04760952285695235
230
+ },
231
+ "harness|ko_mmlu_high_school_government_and_politics|5": {
232
+ "acc": 0.5129533678756477,
233
+ "acc_stderr": 0.0360722806104775,
234
+ "acc_norm": 0.5129533678756477,
235
+ "acc_norm_stderr": 0.0360722806104775
236
+ },
237
+ "harness|ko_mmlu_econometrics|5": {
238
+ "acc": 0.32456140350877194,
239
+ "acc_stderr": 0.044045561573747685,
240
+ "acc_norm": 0.32456140350877194,
241
+ "acc_norm_stderr": 0.044045561573747685
242
+ },
243
+ "harness|ko_mmlu_high_school_psychology|5": {
244
+ "acc": 0.48623853211009177,
245
+ "acc_stderr": 0.02142920208987408,
246
+ "acc_norm": 0.48623853211009177,
247
+ "acc_norm_stderr": 0.02142920208987408
248
+ },
249
+ "harness|ko_mmlu_formal_logic|5": {
250
+ "acc": 0.42857142857142855,
251
+ "acc_stderr": 0.0442626668137991,
252
+ "acc_norm": 0.42857142857142855,
253
+ "acc_norm_stderr": 0.0442626668137991
254
+ },
255
+ "harness|ko_mmlu_nutrition|5": {
256
+ "acc": 0.47058823529411764,
257
+ "acc_stderr": 0.028580341065138286,
258
+ "acc_norm": 0.47058823529411764,
259
+ "acc_norm_stderr": 0.028580341065138286
260
+ },
261
+ "harness|ko_mmlu_business_ethics|5": {
262
+ "acc": 0.48,
263
+ "acc_stderr": 0.050211673156867795,
264
+ "acc_norm": 0.48,
265
+ "acc_norm_stderr": 0.050211673156867795
266
+ },
267
+ "harness|ko_mmlu_international_law|5": {
268
+ "acc": 0.6776859504132231,
269
+ "acc_stderr": 0.042664163633521685,
270
+ "acc_norm": 0.6776859504132231,
271
+ "acc_norm_stderr": 0.042664163633521685
272
+ },
273
+ "harness|ko_mmlu_astronomy|5": {
274
+ "acc": 0.42105263157894735,
275
+ "acc_stderr": 0.04017901275981748,
276
+ "acc_norm": 0.42105263157894735,
277
+ "acc_norm_stderr": 0.04017901275981748
278
+ },
279
+ "harness|ko_mmlu_professional_psychology|5": {
280
+ "acc": 0.4035947712418301,
281
+ "acc_stderr": 0.019848280168401154,
282
+ "acc_norm": 0.4035947712418301,
283
+ "acc_norm_stderr": 0.019848280168401154
284
+ },
285
+ "harness|ko_mmlu_professional_accounting|5": {
286
+ "acc": 0.3475177304964539,
287
+ "acc_stderr": 0.028406627809590947,
288
+ "acc_norm": 0.3475177304964539,
289
+ "acc_norm_stderr": 0.028406627809590947
290
+ },
291
+ "harness|ko_mmlu_machine_learning|5": {
292
+ "acc": 0.35714285714285715,
293
+ "acc_stderr": 0.04547960999764376,
294
+ "acc_norm": 0.35714285714285715,
295
+ "acc_norm_stderr": 0.04547960999764376
296
+ },
297
+ "harness|ko_mmlu_high_school_statistics|5": {
298
+ "acc": 0.4722222222222222,
299
+ "acc_stderr": 0.0340470532865388,
300
+ "acc_norm": 0.4722222222222222,
301
+ "acc_norm_stderr": 0.0340470532865388
302
+ },
303
+ "harness|ko_mmlu_moral_scenarios|5": {
304
+ "acc": 0.29608938547486036,
305
+ "acc_stderr": 0.01526867731760228,
306
+ "acc_norm": 0.29608938547486036,
307
+ "acc_norm_stderr": 0.01526867731760228
308
+ },
309
+ "harness|ko_mmlu_college_computer_science|5": {
310
+ "acc": 0.46,
311
+ "acc_stderr": 0.05009082659620333,
312
+ "acc_norm": 0.46,
313
+ "acc_norm_stderr": 0.05009082659620333
314
+ },
315
+ "harness|ko_mmlu_high_school_computer_science|5": {
316
+ "acc": 0.63,
317
+ "acc_stderr": 0.04852365870939099,
318
+ "acc_norm": 0.63,
319
+ "acc_norm_stderr": 0.04852365870939099
320
+ },
321
+ "harness|ko_mmlu_professional_medicine|5": {
322
+ "acc": 0.41544117647058826,
323
+ "acc_stderr": 0.029935342707877746,
324
+ "acc_norm": 0.41544117647058826,
325
+ "acc_norm_stderr": 0.029935342707877746
326
+ },
327
+ "harness|ko_mmlu_security_studies|5": {
328
+ "acc": 0.5510204081632653,
329
+ "acc_stderr": 0.03184213866687578,
330
+ "acc_norm": 0.5510204081632653,
331
+ "acc_norm_stderr": 0.03184213866687578
332
+ },
333
+ "harness|ko_mmlu_high_school_world_history|5": {
334
+ "acc": 0.5654008438818565,
335
+ "acc_stderr": 0.03226759995510145,
336
+ "acc_norm": 0.5654008438818565,
337
+ "acc_norm_stderr": 0.03226759995510145
338
+ },
339
+ "harness|ko_mmlu_professional_law|5": {
340
+ "acc": 0.3305084745762712,
341
+ "acc_stderr": 0.01201414210184298,
342
+ "acc_norm": 0.3305084745762712,
343
+ "acc_norm_stderr": 0.01201414210184298
344
+ },
345
+ "harness|ko_mmlu_high_school_us_history|5": {
346
+ "acc": 0.47549019607843135,
347
+ "acc_stderr": 0.035050931943487976,
348
+ "acc_norm": 0.47549019607843135,
349
+ "acc_norm_stderr": 0.035050931943487976
350
+ },
351
+ "harness|ko_mmlu_high_school_european_history|5": {
352
+ "acc": 0.45454545454545453,
353
+ "acc_stderr": 0.038881769216740976,
354
+ "acc_norm": 0.45454545454545453,
355
+ "acc_norm_stderr": 0.038881769216740976
356
+ },
357
+ "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.2741738066095471,
359
+ "mc1_stderr": 0.015616518497219376,
360
+ "mc2": 0.44088319088488914,
361
+ "mc2_stderr": 0.01533044885511757
362
+ },
363
+ "harness|ko_commongen_v2|2": {
364
+ "acc": 0.5277449822904369,
365
+ "acc_stderr": 0.01716386797945602,
366
+ "acc_norm": 0.5749704840613932,
367
+ "acc_norm_stderr": 0.016996016308362887
368
+ }
369
+ },
370
+ "versions": {
371
+ "all": 0,
372
+ "harness|ko_arc_challenge|25": 0,
373
+ "harness|ko_hellaswag|10": 0,
374
+ "harness|ko_mmlu_world_religions|5": 1,
375
+ "harness|ko_mmlu_management|5": 1,
376
+ "harness|ko_mmlu_miscellaneous|5": 1,
377
+ "harness|ko_mmlu_anatomy|5": 1,
378
+ "harness|ko_mmlu_abstract_algebra|5": 1,
379
+ "harness|ko_mmlu_conceptual_physics|5": 1,
380
+ "harness|ko_mmlu_virology|5": 1,
381
+ "harness|ko_mmlu_philosophy|5": 1,
382
+ "harness|ko_mmlu_human_aging|5": 1,
383
+ "harness|ko_mmlu_human_sexuality|5": 1,
384
+ "harness|ko_mmlu_medical_genetics|5": 1,
385
+ "harness|ko_mmlu_high_school_geography|5": 1,
386
+ "harness|ko_mmlu_electrical_engineering|5": 1,
387
+ "harness|ko_mmlu_college_physics|5": 1,
388
+ "harness|ko_mmlu_high_school_microeconomics|5": 1,
389
+ "harness|ko_mmlu_high_school_macroeconomics|5": 1,
390
+ "harness|ko_mmlu_computer_security|5": 1,
391
+ "harness|ko_mmlu_global_facts|5": 1,
392
+ "harness|ko_mmlu_jurisprudence|5": 1,
393
+ "harness|ko_mmlu_high_school_chemistry|5": 1,
394
+ "harness|ko_mmlu_high_school_biology|5": 1,
395
+ "harness|ko_mmlu_marketing|5": 1,
396
+ "harness|ko_mmlu_clinical_knowledge|5": 1,
397
+ "harness|ko_mmlu_public_relations|5": 1,
398
+ "harness|ko_mmlu_high_school_mathematics|5": 1,
399
+ "harness|ko_mmlu_high_school_physics|5": 1,
400
+ "harness|ko_mmlu_sociology|5": 1,
401
+ "harness|ko_mmlu_college_medicine|5": 1,
402
+ "harness|ko_mmlu_elementary_mathematics|5": 1,
403
+ "harness|ko_mmlu_college_biology|5": 1,
404
+ "harness|ko_mmlu_college_chemistry|5": 1,
405
+ "harness|ko_mmlu_us_foreign_policy|5": 1,
406
+ "harness|ko_mmlu_moral_disputes|5": 1,
407
+ "harness|ko_mmlu_logical_fallacies|5": 1,
408
+ "harness|ko_mmlu_prehistory|5": 1,
409
+ "harness|ko_mmlu_college_mathematics|5": 1,
410
+ "harness|ko_mmlu_high_school_government_and_politics|5": 1,
411
+ "harness|ko_mmlu_econometrics|5": 1,
412
+ "harness|ko_mmlu_high_school_psychology|5": 1,
413
+ "harness|ko_mmlu_formal_logic|5": 1,
414
+ "harness|ko_mmlu_nutrition|5": 1,
415
+ "harness|ko_mmlu_business_ethics|5": 1,
416
+ "harness|ko_mmlu_international_law|5": 1,
417
+ "harness|ko_mmlu_astronomy|5": 1,
418
+ "harness|ko_mmlu_professional_psychology|5": 1,
419
+ "harness|ko_mmlu_professional_accounting|5": 1,
420
+ "harness|ko_mmlu_machine_learning|5": 1,
421
+ "harness|ko_mmlu_high_school_statistics|5": 1,
422
+ "harness|ko_mmlu_moral_scenarios|5": 1,
423
+ "harness|ko_mmlu_college_computer_science|5": 1,
424
+ "harness|ko_mmlu_high_school_computer_science|5": 1,
425
+ "harness|ko_mmlu_professional_medicine|5": 1,
426
+ "harness|ko_mmlu_security_studies|5": 1,
427
+ "harness|ko_mmlu_high_school_world_history|5": 1,
428
+ "harness|ko_mmlu_professional_law|5": 1,
429
+ "harness|ko_mmlu_high_school_us_history|5": 1,
430
+ "harness|ko_mmlu_high_school_european_history|5": 1,
431
+ "harness|ko_truthfulqa_mc|0": 0,
432
+ "harness|ko_commongen_v2|2": 1
433
+ },
434
+ "config_general": {
435
+ "model_name": "BM-K/mistral-ko-7b-it-v2.0.1",
436
+ "model_sha": "5482aa57e129559221c5109620df556b75e70f3a",
437
+ "model_dtype": "torch.float16",
438
+ "lighteval_sha": "",
439
+ "num_few_shot_default": 0,
440
+ "num_fewshot_seeds": 1,
441
+ "override_batch_size": 1,
442
+ "max_samples": null
443
+ }
444
+ }