barthfab commited on
Commit
aa76947
1 Parent(s): 210a557

Delete DiscoResearch/Llama3-German-8B

Browse files
DiscoResearch/Llama3-German-8B/results_2024_06_03T12-52-58.json DELETED
@@ -1,570 +0,0 @@
1
- {
2
- "config_general": {
3
- "lighteval_sha": "",
4
- "num_few_shot_default": 0,
5
- "num_fewshot_seeds": 1,
6
- "override_batch_size": "auto:6",
7
- "max_samples": "null",
8
- "job_id": "",
9
- "model_name": "DiscoResearch/Llama3-German-8B",
10
- "model_sha": "",
11
- "model_dtype": "torch.bfloat16",
12
- "model_size": ""
13
- },
14
- "results": {
15
- "harness|hellaswag|10": {
16
- "acc,none": 0.59699263095001,
17
- "acc_stderr,none": 0.0048949977367190355,
18
- "acc_norm,none": 0.7961561441943836,
19
- "acc_norm_stderr,none": 0.004020309191221623,
20
- "alias": "hellaswag"
21
- },
22
- "harness|truthfulqa_mc2_m_es|0": {
23
- "acc,none": 0.2788339670468948,
24
- "acc_stderr,none": 0.01597450323607343,
25
- "alias": "truthfulqa_mc2_m_es"
26
- },
27
- "harness|hellaswag_es|10": {
28
- "acc,none": 0.4829315126946874,
29
- "acc_stderr,none": 0.005161518632240655,
30
- "acc_norm,none": 0.6499893321954342,
31
- "acc_norm_stderr,none": 0.004926681042105101,
32
- "alias": "hellaswag_es"
33
- },
34
- "harness|arc_challenge_m_es|25": {
35
- "acc,none": 0.4478632478632479,
36
- "acc_stderr,none": 0.014544164741853641,
37
- "acc_norm,none": 0.47863247863247865,
38
- "acc_norm_stderr,none": 0.014610524729617737,
39
- "alias": "arc_challenge_m_es"
40
- },
41
- "harness|hellaswag_de|10": {
42
- "acc,none": 0.5053373185311699,
43
- "acc_stderr,none": 0.0051658881571013246,
44
- "acc_norm,none": 0.6699402220324508,
45
- "acc_norm_stderr,none": 0.004858632690278109,
46
- "alias": "hellaswag_de"
47
- },
48
- "harness|hellaswag_fr|10": {
49
- "acc,none": 0.4706575283786678,
50
- "acc_stderr,none": 0.005165557449021636,
51
- "acc_norm,none": 0.6366459627329193,
52
- "acc_norm_stderr,none": 0.004977489015703268,
53
- "alias": "hellaswag_fr"
54
- },
55
- "harness|belebele_spa_Latn|5": {
56
- "acc,none": 0.7588888888888888,
57
- "acc_stderr,none": 0.014266513886578925,
58
- "acc_norm,none": 0.7588888888888888,
59
- "acc_norm_stderr,none": 0.014266513886578925,
60
- "alias": "belebele_spa_Latn"
61
- },
62
- "harness|belebele_deu_Latn|5": {
63
- "acc,none": 0.7877777777777778,
64
- "acc_stderr,none": 0.013636956209422681,
65
- "acc_norm,none": 0.7877777777777778,
66
- "acc_norm_stderr,none": 0.013636956209422681,
67
- "alias": "belebele_deu_Latn"
68
- },
69
- "harness|arc_challenge|25": {
70
- "acc,none": 0.5307167235494881,
71
- "acc_stderr,none": 0.014583792546304037,
72
- "acc_norm,none": 0.5563139931740614,
73
- "acc_norm_stderr,none": 0.014518421825670444,
74
- "alias": "arc_challenge"
75
- },
76
- "harness|belebele_ita_Latn|5": {
77
- "acc,none": 0.7577777777777778,
78
- "acc_stderr,none": 0.014288876375699817,
79
- "acc_norm,none": 0.7577777777777778,
80
- "acc_norm_stderr,none": 0.014288876375699817,
81
- "alias": "belebele_ita_Latn"
82
- },
83
- "harness|arc_challenge_m_de|25": {
84
- "acc,none": 0.4627887082976903,
85
- "acc_stderr,none": 0.014589571001051863,
86
- "acc_norm,none": 0.5106928999144568,
87
- "acc_norm_stderr,none": 0.014626797451054007,
88
- "alias": "arc_challenge_m_de"
89
- },
90
- "harness|belebele_fra_Latn|5": {
91
- "acc,none": 0.7788888888888889,
92
- "acc_stderr,none": 0.013840863699859542,
93
- "acc_norm,none": 0.7788888888888889,
94
- "acc_norm_stderr,none": 0.013840863699859542,
95
- "alias": "belebele_fra_Latn"
96
- },
97
- "harness|hendrycksTest|5": {
98
- "acc,none": 0.608033043725965,
99
- "acc_stderr,none": 0.13197695278486477,
100
- "alias": "mmlu"
101
- },
102
- "harness|hendrycksTest-humanities|5": {
103
- "acc,none": 0.608033043725965,
104
- "acc_stderr,none": 0.13197695278486477,
105
- "alias": "mmlu"
106
- },
107
- "harness|hendrycksTest-formal_logic|5": {
108
- "acc,none": 0.608033043725965,
109
- "acc_stderr,none": 0.13197695278486477,
110
- "alias": "mmlu"
111
- },
112
- "harness|hendrycksTest-high_school_european_history|5": {
113
- "acc,none": 0.608033043725965,
114
- "acc_stderr,none": 0.13197695278486477,
115
- "alias": "mmlu"
116
- },
117
- "harness|hendrycksTest-high_school_us_history|5": {
118
- "acc,none": 0.608033043725965,
119
- "acc_stderr,none": 0.13197695278486477,
120
- "alias": "mmlu"
121
- },
122
- "harness|hendrycksTest-high_school_world_history|5": {
123
- "acc,none": 0.608033043725965,
124
- "acc_stderr,none": 0.13197695278486477,
125
- "alias": "mmlu"
126
- },
127
- "harness|hendrycksTest-international_law|5": {
128
- "acc,none": 0.608033043725965,
129
- "acc_stderr,none": 0.13197695278486477,
130
- "alias": "mmlu"
131
- },
132
- "harness|hendrycksTest-jurisprudence|5": {
133
- "acc,none": 0.608033043725965,
134
- "acc_stderr,none": 0.13197695278486477,
135
- "alias": "mmlu"
136
- },
137
- "harness|hendrycksTest-logical_fallacies|5": {
138
- "acc,none": 0.608033043725965,
139
- "acc_stderr,none": 0.13197695278486477,
140
- "alias": "mmlu"
141
- },
142
- "harness|hendrycksTest-moral_disputes|5": {
143
- "acc,none": 0.608033043725965,
144
- "acc_stderr,none": 0.13197695278486477,
145
- "alias": "mmlu"
146
- },
147
- "harness|hendrycksTest-moral_scenarios|5": {
148
- "acc,none": 0.608033043725965,
149
- "acc_stderr,none": 0.13197695278486477,
150
- "alias": "mmlu"
151
- },
152
- "harness|hendrycksTest-philosophy|5": {
153
- "acc,none": 0.608033043725965,
154
- "acc_stderr,none": 0.13197695278486477,
155
- "alias": "mmlu"
156
- },
157
- "harness|hendrycksTest-prehistory|5": {
158
- "acc,none": 0.608033043725965,
159
- "acc_stderr,none": 0.13197695278486477,
160
- "alias": "mmlu"
161
- },
162
- "harness|hendrycksTest-professional_law|5": {
163
- "acc,none": 0.608033043725965,
164
- "acc_stderr,none": 0.13197695278486477,
165
- "alias": "mmlu"
166
- },
167
- "harness|hendrycksTest-world_religions|5": {
168
- "acc,none": 0.608033043725965,
169
- "acc_stderr,none": 0.13197695278486477,
170
- "alias": "mmlu"
171
- },
172
- "harness|hendrycksTest-other|5": {
173
- "acc,none": 0.608033043725965,
174
- "acc_stderr,none": 0.13197695278486477,
175
- "alias": "mmlu"
176
- },
177
- "harness|hendrycksTest-business_ethics|5": {
178
- "acc,none": 0.608033043725965,
179
- "acc_stderr,none": 0.13197695278486477,
180
- "alias": "mmlu"
181
- },
182
- "harness|hendrycksTest-clinical_knowledge|5": {
183
- "acc,none": 0.608033043725965,
184
- "acc_stderr,none": 0.13197695278486477,
185
- "alias": "mmlu"
186
- },
187
- "harness|hendrycksTest-college_medicine|5": {
188
- "acc,none": 0.608033043725965,
189
- "acc_stderr,none": 0.13197695278486477,
190
- "alias": "mmlu"
191
- },
192
- "harness|hendrycksTest-global_facts|5": {
193
- "acc,none": 0.608033043725965,
194
- "acc_stderr,none": 0.13197695278486477,
195
- "alias": "mmlu"
196
- },
197
- "harness|hendrycksTest-human_aging|5": {
198
- "acc,none": 0.608033043725965,
199
- "acc_stderr,none": 0.13197695278486477,
200
- "alias": "mmlu"
201
- },
202
- "harness|hendrycksTest-management|5": {
203
- "acc,none": 0.608033043725965,
204
- "acc_stderr,none": 0.13197695278486477,
205
- "alias": "mmlu"
206
- },
207
- "harness|hendrycksTest-marketing|5": {
208
- "acc,none": 0.608033043725965,
209
- "acc_stderr,none": 0.13197695278486477,
210
- "alias": "mmlu"
211
- },
212
- "harness|hendrycksTest-medical_genetics|5": {
213
- "acc,none": 0.608033043725965,
214
- "acc_stderr,none": 0.13197695278486477,
215
- "alias": "mmlu"
216
- },
217
- "harness|hendrycksTest-miscellaneous|5": {
218
- "acc,none": 0.608033043725965,
219
- "acc_stderr,none": 0.13197695278486477,
220
- "alias": "mmlu"
221
- },
222
- "harness|hendrycksTest-nutrition|5": {
223
- "acc,none": 0.608033043725965,
224
- "acc_stderr,none": 0.13197695278486477,
225
- "alias": "mmlu"
226
- },
227
- "harness|hendrycksTest-professional_accounting|5": {
228
- "acc,none": 0.608033043725965,
229
- "acc_stderr,none": 0.13197695278486477,
230
- "alias": "mmlu"
231
- },
232
- "harness|hendrycksTest-professional_medicine|5": {
233
- "acc,none": 0.608033043725965,
234
- "acc_stderr,none": 0.13197695278486477,
235
- "alias": "mmlu"
236
- },
237
- "harness|hendrycksTest-virology|5": {
238
- "acc,none": 0.608033043725965,
239
- "acc_stderr,none": 0.13197695278486477,
240
- "alias": "mmlu"
241
- },
242
- "harness|hendrycksTest-social_sciences|5": {
243
- "acc,none": 0.608033043725965,
244
- "acc_stderr,none": 0.13197695278486477,
245
- "alias": "mmlu"
246
- },
247
- "harness|hendrycksTest-econometrics|5": {
248
- "acc,none": 0.608033043725965,
249
- "acc_stderr,none": 0.13197695278486477,
250
- "alias": "mmlu"
251
- },
252
- "harness|hendrycksTest-high_school_geography|5": {
253
- "acc,none": 0.608033043725965,
254
- "acc_stderr,none": 0.13197695278486477,
255
- "alias": "mmlu"
256
- },
257
- "harness|hendrycksTest-high_school_government_and_politics|5": {
258
- "acc,none": 0.608033043725965,
259
- "acc_stderr,none": 0.13197695278486477,
260
- "alias": "mmlu"
261
- },
262
- "harness|hendrycksTest-high_school_macroeconomics|5": {
263
- "acc,none": 0.608033043725965,
264
- "acc_stderr,none": 0.13197695278486477,
265
- "alias": "mmlu"
266
- },
267
- "harness|hendrycksTest-high_school_microeconomics|5": {
268
- "acc,none": 0.608033043725965,
269
- "acc_stderr,none": 0.13197695278486477,
270
- "alias": "mmlu"
271
- },
272
- "harness|hendrycksTest-high_school_psychology|5": {
273
- "acc,none": 0.608033043725965,
274
- "acc_stderr,none": 0.13197695278486477,
275
- "alias": "mmlu"
276
- },
277
- "harness|hendrycksTest-human_sexuality|5": {
278
- "acc,none": 0.608033043725965,
279
- "acc_stderr,none": 0.13197695278486477,
280
- "alias": "mmlu"
281
- },
282
- "harness|hendrycksTest-professional_psychology|5": {
283
- "acc,none": 0.608033043725965,
284
- "acc_stderr,none": 0.13197695278486477,
285
- "alias": "mmlu"
286
- },
287
- "harness|hendrycksTest-public_relations|5": {
288
- "acc,none": 0.608033043725965,
289
- "acc_stderr,none": 0.13197695278486477,
290
- "alias": "mmlu"
291
- },
292
- "harness|hendrycksTest-security_studies|5": {
293
- "acc,none": 0.608033043725965,
294
- "acc_stderr,none": 0.13197695278486477,
295
- "alias": "mmlu"
296
- },
297
- "harness|hendrycksTest-sociology|5": {
298
- "acc,none": 0.608033043725965,
299
- "acc_stderr,none": 0.13197695278486477,
300
- "alias": "mmlu"
301
- },
302
- "harness|hendrycksTest-us_foreign_policy|5": {
303
- "acc,none": 0.608033043725965,
304
- "acc_stderr,none": 0.13197695278486477,
305
- "alias": "mmlu"
306
- },
307
- "harness|hendrycksTest-stem|5": {
308
- "acc,none": 0.608033043725965,
309
- "acc_stderr,none": 0.13197695278486477,
310
- "alias": "mmlu"
311
- },
312
- "harness|hendrycksTest-abstract_algebra|5": {
313
- "acc,none": 0.608033043725965,
314
- "acc_stderr,none": 0.13197695278486477,
315
- "alias": "mmlu"
316
- },
317
- "harness|hendrycksTest-anatomy|5": {
318
- "acc,none": 0.608033043725965,
319
- "acc_stderr,none": 0.13197695278486477,
320
- "alias": "mmlu"
321
- },
322
- "harness|hendrycksTest-astronomy|5": {
323
- "acc,none": 0.608033043725965,
324
- "acc_stderr,none": 0.13197695278486477,
325
- "alias": "mmlu"
326
- },
327
- "harness|hendrycksTest-college_biology|5": {
328
- "acc,none": 0.608033043725965,
329
- "acc_stderr,none": 0.13197695278486477,
330
- "alias": "mmlu"
331
- },
332
- "harness|hendrycksTest-college_chemistry|5": {
333
- "acc,none": 0.608033043725965,
334
- "acc_stderr,none": 0.13197695278486477,
335
- "alias": "mmlu"
336
- },
337
- "harness|hendrycksTest-college_computer_science|5": {
338
- "acc,none": 0.608033043725965,
339
- "acc_stderr,none": 0.13197695278486477,
340
- "alias": "mmlu"
341
- },
342
- "harness|hendrycksTest-college_mathematics|5": {
343
- "acc,none": 0.608033043725965,
344
- "acc_stderr,none": 0.13197695278486477,
345
- "alias": "mmlu"
346
- },
347
- "harness|hendrycksTest-college_physics|5": {
348
- "acc,none": 0.608033043725965,
349
- "acc_stderr,none": 0.13197695278486477,
350
- "alias": "mmlu"
351
- },
352
- "harness|hendrycksTest-computer_security|5": {
353
- "acc,none": 0.608033043725965,
354
- "acc_stderr,none": 0.13197695278486477,
355
- "alias": "mmlu"
356
- },
357
- "harness|hendrycksTest-conceptual_physics|5": {
358
- "acc,none": 0.608033043725965,
359
- "acc_stderr,none": 0.13197695278486477,
360
- "alias": "mmlu"
361
- },
362
- "harness|hendrycksTest-electrical_engineering|5": {
363
- "acc,none": 0.608033043725965,
364
- "acc_stderr,none": 0.13197695278486477,
365
- "alias": "mmlu"
366
- },
367
- "harness|hendrycksTest-elementary_mathematics|5": {
368
- "acc,none": 0.608033043725965,
369
- "acc_stderr,none": 0.13197695278486477,
370
- "alias": "mmlu"
371
- },
372
- "harness|hendrycksTest-high_school_biology|5": {
373
- "acc,none": 0.608033043725965,
374
- "acc_stderr,none": 0.13197695278486477,
375
- "alias": "mmlu"
376
- },
377
- "harness|hendrycksTest-high_school_chemistry|5": {
378
- "acc,none": 0.608033043725965,
379
- "acc_stderr,none": 0.13197695278486477,
380
- "alias": "mmlu"
381
- },
382
- "harness|hendrycksTest-high_school_computer_science|5": {
383
- "acc,none": 0.608033043725965,
384
- "acc_stderr,none": 0.13197695278486477,
385
- "alias": "mmlu"
386
- },
387
- "harness|hendrycksTest-high_school_mathematics|5": {
388
- "acc,none": 0.608033043725965,
389
- "acc_stderr,none": 0.13197695278486477,
390
- "alias": "mmlu"
391
- },
392
- "harness|hendrycksTest-high_school_physics|5": {
393
- "acc,none": 0.608033043725965,
394
- "acc_stderr,none": 0.13197695278486477,
395
- "alias": "mmlu"
396
- },
397
- "harness|hendrycksTest-high_school_statistics|5": {
398
- "acc,none": 0.608033043725965,
399
- "acc_stderr,none": 0.13197695278486477,
400
- "alias": "mmlu"
401
- },
402
- "harness|hendrycksTest-machine_learning|5": {
403
- "acc,none": 0.608033043725965,
404
- "acc_stderr,none": 0.13197695278486477,
405
- "alias": "mmlu"
406
- },
407
- "harness|gsm8k|5": {
408
- "exact_match,get-answer": 0.38059135708870356,
409
- "exact_match_stderr,get-answer": 0.01337397127772982,
410
- "alias": "gsm8k"
411
- },
412
- "harness|mmlu_m_de|5": {
413
- "acc,none": 0.547971036355408,
414
- "acc_stderr,none": 0.00432254283159349,
415
- "alias": "mmlu_m_de"
416
- },
417
- "harness|mmlu_m_it|5": {
418
- "acc,none": 0.5183198609956939,
419
- "acc_stderr,none": 0.004343100866371246,
420
- "alias": "mmlu_m_it"
421
- },
422
- "harness|truthfulqa_mc2|0": {
423
- "acc,none": 0.44852440609276806,
424
- "acc_stderr,none": 0.014462041584884736,
425
- "alias": "truthfulqa_mc2"
426
- },
427
- "harness|truthfulqa_mc2_m_de|0": {
428
- "acc,none": 0.22588832487309646,
429
- "acc_stderr,none": 0.01490600748558298,
430
- "alias": "truthfulqa_mc2_m_de"
431
- },
432
- "harness|belebele_eng_Latn|5": {
433
- "acc,none": 0.8355555555555556,
434
- "acc_stderr,none": 0.01236281648813283,
435
- "acc_norm,none": 0.8355555555555556,
436
- "acc_norm_stderr,none": 0.01236281648813283,
437
- "alias": "belebele_eng_Latn"
438
- },
439
- "harness|arc_challenge_m_it|25": {
440
- "acc,none": 0.43370402053036783,
441
- "acc_stderr,none": 0.014500969591642427,
442
- "acc_norm,none": 0.4679213002566296,
443
- "acc_norm_stderr,none": 0.014600002258361335,
444
- "alias": "arc_challenge_m_it"
445
- },
446
- "harness|hellaswag_it|10": {
447
- "acc,none": 0.4543674534972261,
448
- "acc_stderr,none": 0.005193363646828413,
449
- "acc_norm,none": 0.617861416295007,
450
- "acc_norm_stderr,none": 0.005068167676974343,
451
- "alias": "hellaswag_it"
452
- },
453
- "harness|mmlu_m_fr|5": {
454
- "acc,none": 0.5196700022916507,
455
- "acc_stderr,none": 0.004366805572708734,
456
- "alias": "mmlu_m_fr"
457
- },
458
- "harness|truthfulqa_mc2_m_fr|0": {
459
- "acc,none": 0.2363405336721728,
460
- "acc_stderr,none": 0.015153316796783202,
461
- "alias": "truthfulqa_mc2_m_fr"
462
- },
463
- "harness|mmlu_m_es|5": {
464
- "acc,none": 0.527973601319934,
465
- "acc_stderr,none": 0.004323398907877741,
466
- "alias": "mmlu_m_es"
467
- },
468
- "harness|truthfulqa_mc2_m_it|0": {
469
- "acc,none": 0.26947637292464877,
470
- "acc_stderr,none": 0.01586624307321505,
471
- "alias": "truthfulqa_mc2_m_it"
472
- },
473
- "harness|arc_challenge_m_fr|25": {
474
- "acc,none": 0.41573994867408043,
475
- "acc_stderr,none": 0.014420906263396443,
476
- "acc_norm,none": 0.47476475620188197,
477
- "acc_norm_stderr,none": 0.014611498054709703,
478
- "alias": "arc_challenge_m_fr"
479
- }
480
- },
481
- "versions": {
482
- "harness|hellaswag|10": 1.0,
483
- "harness|truthfulqa_mc2_m_es|0": "Yaml",
484
- "harness|hellaswag_es|10": 1.0,
485
- "harness|arc_challenge_m_es|25": 1.0,
486
- "harness|hellaswag_de|10": 1.0,
487
- "harness|hellaswag_fr|10": 1.0,
488
- "harness|belebele_spa_Latn|5": 0.0,
489
- "harness|belebele_deu_Latn|5": 0.0,
490
- "harness|arc_challenge|25": 1.0,
491
- "harness|belebele_ita_Latn|5": 0.0,
492
- "harness|arc_challenge_m_de|25": 1.0,
493
- "harness|belebele_fra_Latn|5": 0.0,
494
- "harness|hendrycksTest|5": "N/A",
495
- "harness|hendrycksTest-humanities|5": "N/A",
496
- "harness|hendrycksTest-formal_logic|5": "N/A",
497
- "harness|hendrycksTest-high_school_european_history|5": "N/A",
498
- "harness|hendrycksTest-high_school_us_history|5": "N/A",
499
- "harness|hendrycksTest-high_school_world_history|5": "N/A",
500
- "harness|hendrycksTest-international_law|5": "N/A",
501
- "harness|hendrycksTest-jurisprudence|5": "N/A",
502
- "harness|hendrycksTest-logical_fallacies|5": "N/A",
503
- "harness|hendrycksTest-moral_disputes|5": "N/A",
504
- "harness|hendrycksTest-moral_scenarios|5": "N/A",
505
- "harness|hendrycksTest-philosophy|5": "N/A",
506
- "harness|hendrycksTest-prehistory|5": "N/A",
507
- "harness|hendrycksTest-professional_law|5": "N/A",
508
- "harness|hendrycksTest-world_religions|5": "N/A",
509
- "harness|hendrycksTest-other|5": "N/A",
510
- "harness|hendrycksTest-business_ethics|5": "N/A",
511
- "harness|hendrycksTest-clinical_knowledge|5": "N/A",
512
- "harness|hendrycksTest-college_medicine|5": "N/A",
513
- "harness|hendrycksTest-global_facts|5": "N/A",
514
- "harness|hendrycksTest-human_aging|5": "N/A",
515
- "harness|hendrycksTest-management|5": "N/A",
516
- "harness|hendrycksTest-marketing|5": "N/A",
517
- "harness|hendrycksTest-medical_genetics|5": "N/A",
518
- "harness|hendrycksTest-miscellaneous|5": "N/A",
519
- "harness|hendrycksTest-nutrition|5": "N/A",
520
- "harness|hendrycksTest-professional_accounting|5": "N/A",
521
- "harness|hendrycksTest-professional_medicine|5": "N/A",
522
- "harness|hendrycksTest-virology|5": "N/A",
523
- "harness|hendrycksTest-social_sciences|5": "N/A",
524
- "harness|hendrycksTest-econometrics|5": "N/A",
525
- "harness|hendrycksTest-high_school_geography|5": "N/A",
526
- "harness|hendrycksTest-high_school_government_and_politics|5": "N/A",
527
- "harness|hendrycksTest-high_school_macroeconomics|5": "N/A",
528
- "harness|hendrycksTest-high_school_microeconomics|5": "N/A",
529
- "harness|hendrycksTest-high_school_psychology|5": "N/A",
530
- "harness|hendrycksTest-human_sexuality|5": "N/A",
531
- "harness|hendrycksTest-professional_psychology|5": "N/A",
532
- "harness|hendrycksTest-public_relations|5": "N/A",
533
- "harness|hendrycksTest-security_studies|5": "N/A",
534
- "harness|hendrycksTest-sociology|5": "N/A",
535
- "harness|hendrycksTest-us_foreign_policy|5": "N/A",
536
- "harness|hendrycksTest-stem|5": "N/A",
537
- "harness|hendrycksTest-abstract_algebra|5": "N/A",
538
- "harness|hendrycksTest-anatomy|5": "N/A",
539
- "harness|hendrycksTest-astronomy|5": "N/A",
540
- "harness|hendrycksTest-college_biology|5": "N/A",
541
- "harness|hendrycksTest-college_chemistry|5": "N/A",
542
- "harness|hendrycksTest-college_computer_science|5": "N/A",
543
- "harness|hendrycksTest-college_mathematics|5": "N/A",
544
- "harness|hendrycksTest-college_physics|5": "N/A",
545
- "harness|hendrycksTest-computer_security|5": "N/A",
546
- "harness|hendrycksTest-conceptual_physics|5": "N/A",
547
- "harness|hendrycksTest-electrical_engineering|5": "N/A",
548
- "harness|hendrycksTest-elementary_mathematics|5": "N/A",
549
- "harness|hendrycksTest-high_school_biology|5": "N/A",
550
- "harness|hendrycksTest-high_school_chemistry|5": "N/A",
551
- "harness|hendrycksTest-high_school_computer_science|5": "N/A",
552
- "harness|hendrycksTest-high_school_mathematics|5": "N/A",
553
- "harness|hendrycksTest-high_school_physics|5": "N/A",
554
- "harness|hendrycksTest-high_school_statistics|5": "N/A",
555
- "harness|hendrycksTest-machine_learning|5": "N/A",
556
- "harness|gsm8k|5": 2.0,
557
- "harness|mmlu_m_de|5": "Yaml",
558
- "harness|mmlu_m_it|5": "Yaml",
559
- "harness|truthfulqa_mc2|0": 2.0,
560
- "harness|truthfulqa_mc2_m_de|0": "Yaml",
561
- "harness|belebele_eng_Latn|5": 0.0,
562
- "harness|arc_challenge_m_it|25": 1.0,
563
- "harness|hellaswag_it|10": 1.0,
564
- "harness|mmlu_m_fr|5": "Yaml",
565
- "harness|truthfulqa_mc2_m_fr|0": "Yaml",
566
- "harness|mmlu_m_es|5": "Yaml",
567
- "harness|truthfulqa_mc2_m_it|0": "Yaml",
568
- "harness|arc_challenge_m_fr|25": 1.0
569
- }
570
- }