lvkaokao commited on
Commit
124d34b
1 Parent(s): 7be779d

add Qwen2.

Browse files
Intel/results_2024-06-06-00-18-35.json ADDED
@@ -0,0 +1,598 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config_general": {
3
+ "lighteval_sha": "1.4",
4
+ "num_few_shot_default": null,
5
+ "num_fewshot_seeds": null,
6
+ "override_batch_size": null,
7
+ "max_samples": null,
8
+ "job_id": -1,
9
+ "start_time": null,
10
+ "end_time": "2024-06-06-00-18-35",
11
+ "total_evaluation_time_secondes": "",
12
+ "model_name": "Intel/Qwen2-7B-int4-inc",
13
+ "model_sha": "",
14
+ "model_dtype": "4bit",
15
+ "model_size": 5.2,
16
+ "model_params": 7,
17
+ "quant_type": "AutoRound",
18
+ "precision": "4bit"
19
+ },
20
+ "results": {
21
+ "harness|truthfulqa:mc2|0": {
22
+ "acc,none": 0.5428341738734119,
23
+ "acc_stderr,none": 0.014866609073167824,
24
+ "alias": "truthfulqa_mc2"
25
+ },
26
+ "harness|arc:easy|0": {
27
+ "acc,none": 0.7638888888888888,
28
+ "acc_stderr,none": 0.008714480491711292,
29
+ "acc_norm,none": 0.7117003367003367,
30
+ "acc_norm_stderr,none": 0.00929477425202962,
31
+ "alias": "arc_easy"
32
+ },
33
+ "harness|lambada:openai|0": {
34
+ "perplexity,none": 3.5641183746208984,
35
+ "perplexity_stderr,none": 0.07387640175498134,
36
+ "acc,none": 0.7252086163399961,
37
+ "acc_stderr,none": 0.006219351548299035,
38
+ "alias": "lambada_openai"
39
+ },
40
+ "harness|arc:challenge|0": {
41
+ "acc,none": 0.4616040955631399,
42
+ "acc_stderr,none": 0.01456824555029636,
43
+ "acc_norm,none": 0.4761092150170648,
44
+ "acc_norm_stderr,none": 0.014594701798071654,
45
+ "alias": "arc_challenge"
46
+ },
47
+ "harness|boolq|0": {
48
+ "acc,none": 0.8159021406727829,
49
+ "acc_stderr,none": 0.006778536599685005,
50
+ "alias": "boolq"
51
+ },
52
+ "harness|truthfulqa:mc1|0": {
53
+ "acc,none": 0.36107711138310894,
54
+ "acc_stderr,none": 0.016814312844836886,
55
+ "alias": "truthfulqa_mc1"
56
+ },
57
+ "harness|openbookqa|0": {
58
+ "acc,none": 0.34,
59
+ "acc_stderr,none": 0.021206117013673066,
60
+ "acc_norm,none": 0.426,
61
+ "acc_norm_stderr,none": 0.022136577335085637,
62
+ "alias": "openbookqa"
63
+ },
64
+ "harness|winogrande|0": {
65
+ "acc,none": 0.7261247040252565,
66
+ "acc_stderr,none": 0.012533292732620296,
67
+ "alias": "winogrande"
68
+ },
69
+ "harness|hellaswag|0": {
70
+ "acc,none": 0.5762796255725952,
71
+ "acc_stderr,none": 0.004931372657129809,
72
+ "acc_norm,none": 0.7721569408484366,
73
+ "acc_norm_stderr,none": 0.0041858359906645996,
74
+ "alias": "hellaswag"
75
+ },
76
+ "harness|piqa|0": {
77
+ "acc,none": 0.7916213275299239,
78
+ "acc_stderr,none": 0.009476125383049452,
79
+ "acc_norm,none": 0.8025027203482046,
80
+ "acc_norm_stderr,none": 0.009288578108523267,
81
+ "alias": "piqa"
82
+ },
83
+ "harness|mmlu|0": {
84
+ "acc,none": 0.6652186298248113,
85
+ "acc_stderr,none": 0.003715571902177435,
86
+ "alias": "mmlu"
87
+ },
88
+ "harness|mmlu_humanities|0": {
89
+ "alias": " - humanities",
90
+ "acc,none": 0.5846971307120085,
91
+ "acc_stderr,none": 0.006515055624404063
92
+ },
93
+ "harness|mmlu_formal_logic|0": {
94
+ "alias": " - formal_logic",
95
+ "acc,none": 0.5555555555555556,
96
+ "acc_stderr,none": 0.04444444444444449
97
+ },
98
+ "harness|mmlu_high_school_european_history|0": {
99
+ "alias": " - high_school_european_history",
100
+ "acc,none": 0.7757575757575758,
101
+ "acc_stderr,none": 0.03256866661681102
102
+ },
103
+ "harness|mmlu_high_school_us_history|0": {
104
+ "alias": " - high_school_us_history",
105
+ "acc,none": 0.8431372549019608,
106
+ "acc_stderr,none": 0.025524722324553318
107
+ },
108
+ "harness|mmlu_high_school_world_history|0": {
109
+ "alias": " - high_school_world_history",
110
+ "acc,none": 0.810126582278481,
111
+ "acc_stderr,none": 0.025530100460233494
112
+ },
113
+ "harness|mmlu_international_law|0": {
114
+ "alias": " - international_law",
115
+ "acc,none": 0.8347107438016529,
116
+ "acc_stderr,none": 0.03390780612972776
117
+ },
118
+ "harness|mmlu_jurisprudence|0": {
119
+ "alias": " - jurisprudence",
120
+ "acc,none": 0.7962962962962963,
121
+ "acc_stderr,none": 0.03893542518824849
122
+ },
123
+ "harness|mmlu_logical_fallacies|0": {
124
+ "alias": " - logical_fallacies",
125
+ "acc,none": 0.7975460122699386,
126
+ "acc_stderr,none": 0.03157065078911899
127
+ },
128
+ "harness|mmlu_moral_disputes|0": {
129
+ "alias": " - moral_disputes",
130
+ "acc,none": 0.7658959537572254,
131
+ "acc_stderr,none": 0.02279711027807113
132
+ },
133
+ "harness|mmlu_moral_scenarios|0": {
134
+ "alias": " - moral_scenarios",
135
+ "acc,none": 0.2424581005586592,
136
+ "acc_stderr,none": 0.014333522059217887
137
+ },
138
+ "harness|mmlu_philosophy|0": {
139
+ "alias": " - philosophy",
140
+ "acc,none": 0.7395498392282959,
141
+ "acc_stderr,none": 0.02492672322484555
142
+ },
143
+ "harness|mmlu_prehistory|0": {
144
+ "alias": " - prehistory",
145
+ "acc,none": 0.7623456790123457,
146
+ "acc_stderr,none": 0.023683591837008557
147
+ },
148
+ "harness|mmlu_professional_law|0": {
149
+ "alias": " - professional_law",
150
+ "acc,none": 0.5032594524119948,
151
+ "acc_stderr,none": 0.012769964760343314
152
+ },
153
+ "harness|mmlu_world_religions|0": {
154
+ "alias": " - world_religions",
155
+ "acc,none": 0.8245614035087719,
156
+ "acc_stderr,none": 0.02917088550072766
157
+ },
158
+ "harness|mmlu_other|0": {
159
+ "alias": " - other",
160
+ "acc,none": 0.7347924042484711,
161
+ "acc_stderr,none": 0.007602922006730982
162
+ },
163
+ "harness|mmlu_business_ethics|0": {
164
+ "alias": " - business_ethics",
165
+ "acc,none": 0.75,
166
+ "acc_stderr,none": 0.04351941398892446
167
+ },
168
+ "harness|mmlu_clinical_knowledge|0": {
169
+ "alias": " - clinical_knowledge",
170
+ "acc,none": 0.7584905660377359,
171
+ "acc_stderr,none": 0.026341480371118352
172
+ },
173
+ "harness|mmlu_college_medicine|0": {
174
+ "alias": " - college_medicine",
175
+ "acc,none": 0.6936416184971098,
176
+ "acc_stderr,none": 0.03514942551267438
177
+ },
178
+ "harness|mmlu_global_facts|0": {
179
+ "alias": " - global_facts",
180
+ "acc,none": 0.37,
181
+ "acc_stderr,none": 0.04852365870939099
182
+ },
183
+ "harness|mmlu_human_aging|0": {
184
+ "alias": " - human_aging",
185
+ "acc,none": 0.7219730941704036,
186
+ "acc_stderr,none": 0.030069584874494033
187
+ },
188
+ "harness|mmlu_management|0": {
189
+ "alias": " - management",
190
+ "acc,none": 0.8543689320388349,
191
+ "acc_stderr,none": 0.0349260647662379
192
+ },
193
+ "harness|mmlu_marketing|0": {
194
+ "alias": " - marketing",
195
+ "acc,none": 0.9273504273504274,
196
+ "acc_stderr,none": 0.017004368568132356
197
+ },
198
+ "harness|mmlu_medical_genetics|0": {
199
+ "alias": " - medical_genetics",
200
+ "acc,none": 0.78,
201
+ "acc_stderr,none": 0.04163331998932263
202
+ },
203
+ "harness|mmlu_miscellaneous|0": {
204
+ "alias": " - miscellaneous",
205
+ "acc,none": 0.8250319284802043,
206
+ "acc_stderr,none": 0.013586619219903341
207
+ },
208
+ "harness|mmlu_nutrition|0": {
209
+ "alias": " - nutrition",
210
+ "acc,none": 0.7581699346405228,
211
+ "acc_stderr,none": 0.024518195641879334
212
+ },
213
+ "harness|mmlu_professional_accounting|0": {
214
+ "alias": " - professional_accounting",
215
+ "acc,none": 0.5460992907801419,
216
+ "acc_stderr,none": 0.02970045324729148
217
+ },
218
+ "harness|mmlu_professional_medicine|0": {
219
+ "alias": " - professional_medicine",
220
+ "acc,none": 0.6985294117647058,
221
+ "acc_stderr,none": 0.027875982114273168
222
+ },
223
+ "harness|mmlu_virology|0": {
224
+ "alias": " - virology",
225
+ "acc,none": 0.5060240963855421,
226
+ "acc_stderr,none": 0.03892212195333045
227
+ },
228
+ "harness|mmlu_social_sciences|0": {
229
+ "alias": " - social_sciences",
230
+ "acc,none": 0.7747806304842378,
231
+ "acc_stderr,none": 0.007405038078278177
232
+ },
233
+ "harness|mmlu_econometrics|0": {
234
+ "alias": " - econometrics",
235
+ "acc,none": 0.5263157894736842,
236
+ "acc_stderr,none": 0.046970851366478626
237
+ },
238
+ "harness|mmlu_high_school_geography|0": {
239
+ "alias": " - high_school_geography",
240
+ "acc,none": 0.8636363636363636,
241
+ "acc_stderr,none": 0.024450155973189835
242
+ },
243
+ "harness|mmlu_high_school_government_and_politics|0": {
244
+ "alias": " - high_school_government_and_politics",
245
+ "acc,none": 0.8860103626943006,
246
+ "acc_stderr,none": 0.022935144053919426
247
+ },
248
+ "harness|mmlu_high_school_macroeconomics|0": {
249
+ "alias": " - high_school_macroeconomics",
250
+ "acc,none": 0.7230769230769231,
251
+ "acc_stderr,none": 0.022688042352424994
252
+ },
253
+ "harness|mmlu_high_school_microeconomics|0": {
254
+ "alias": " - high_school_microeconomics",
255
+ "acc,none": 0.7857142857142857,
256
+ "acc_stderr,none": 0.026653531596715473
257
+ },
258
+ "harness|mmlu_high_school_psychology|0": {
259
+ "alias": " - high_school_psychology",
260
+ "acc,none": 0.8532110091743119,
261
+ "acc_stderr,none": 0.01517314184512626
262
+ },
263
+ "harness|mmlu_human_sexuality|0": {
264
+ "alias": " - human_sexuality",
265
+ "acc,none": 0.7862595419847328,
266
+ "acc_stderr,none": 0.0359546161177469
267
+ },
268
+ "harness|mmlu_professional_psychology|0": {
269
+ "alias": " - professional_psychology",
270
+ "acc,none": 0.704248366013072,
271
+ "acc_stderr,none": 0.01846315413263282
272
+ },
273
+ "harness|mmlu_public_relations|0": {
274
+ "alias": " - public_relations",
275
+ "acc,none": 0.7,
276
+ "acc_stderr,none": 0.04389311454644287
277
+ },
278
+ "harness|mmlu_security_studies|0": {
279
+ "alias": " - security_studies",
280
+ "acc,none": 0.7755102040816326,
281
+ "acc_stderr,none": 0.02671143055553841
282
+ },
283
+ "harness|mmlu_sociology|0": {
284
+ "alias": " - sociology",
285
+ "acc,none": 0.8109452736318408,
286
+ "acc_stderr,none": 0.027686913588013024
287
+ },
288
+ "harness|mmlu_us_foreign_policy|0": {
289
+ "alias": " - us_foreign_policy",
290
+ "acc,none": 0.84,
291
+ "acc_stderr,none": 0.03684529491774708
292
+ },
293
+ "harness|mmlu_stem|0": {
294
+ "alias": " - stem",
295
+ "acc,none": 0.609895337773549,
296
+ "acc_stderr,none": 0.008423044034949694
297
+ },
298
+ "harness|mmlu_abstract_algebra|0": {
299
+ "alias": " - abstract_algebra",
300
+ "acc,none": 0.39,
301
+ "acc_stderr,none": 0.04902071300001975
302
+ },
303
+ "harness|mmlu_anatomy|0": {
304
+ "alias": " - anatomy",
305
+ "acc,none": 0.6,
306
+ "acc_stderr,none": 0.04232073695151589
307
+ },
308
+ "harness|mmlu_astronomy|0": {
309
+ "alias": " - astronomy",
310
+ "acc,none": 0.7302631578947368,
311
+ "acc_stderr,none": 0.03611780560284898
312
+ },
313
+ "harness|mmlu_college_biology|0": {
314
+ "alias": " - college_biology",
315
+ "acc,none": 0.7569444444444444,
316
+ "acc_stderr,none": 0.035868792800803406
317
+ },
318
+ "harness|mmlu_college_chemistry|0": {
319
+ "alias": " - college_chemistry",
320
+ "acc,none": 0.47,
321
+ "acc_stderr,none": 0.05016135580465919
322
+ },
323
+ "harness|mmlu_college_computer_science|0": {
324
+ "alias": " - college_computer_science",
325
+ "acc,none": 0.58,
326
+ "acc_stderr,none": 0.04960449637488583
327
+ },
328
+ "harness|mmlu_college_mathematics|0": {
329
+ "alias": " - college_mathematics",
330
+ "acc,none": 0.37,
331
+ "acc_stderr,none": 0.04852365870939099
332
+ },
333
+ "harness|mmlu_college_physics|0": {
334
+ "alias": " - college_physics",
335
+ "acc,none": 0.4411764705882353,
336
+ "acc_stderr,none": 0.049406356306056595
337
+ },
338
+ "harness|mmlu_computer_security|0": {
339
+ "alias": " - computer_security",
340
+ "acc,none": 0.69,
341
+ "acc_stderr,none": 0.04648231987117316
342
+ },
343
+ "harness|mmlu_conceptual_physics|0": {
344
+ "alias": " - conceptual_physics",
345
+ "acc,none": 0.6382978723404256,
346
+ "acc_stderr,none": 0.0314108219759624
347
+ },
348
+ "harness|mmlu_electrical_engineering|0": {
349
+ "alias": " - electrical_engineering",
350
+ "acc,none": 0.6482758620689655,
351
+ "acc_stderr,none": 0.0397923663749741
352
+ },
353
+ "harness|mmlu_elementary_mathematics|0": {
354
+ "alias": " - elementary_mathematics",
355
+ "acc,none": 0.5978835978835979,
356
+ "acc_stderr,none": 0.025253032554997692
357
+ },
358
+ "harness|mmlu_high_school_biology|0": {
359
+ "alias": " - high_school_biology",
360
+ "acc,none": 0.8387096774193549,
361
+ "acc_stderr,none": 0.020923327006423294
362
+ },
363
+ "harness|mmlu_high_school_chemistry|0": {
364
+ "alias": " - high_school_chemistry",
365
+ "acc,none": 0.5665024630541872,
366
+ "acc_stderr,none": 0.034867317274198714
367
+ },
368
+ "harness|mmlu_high_school_computer_science|0": {
369
+ "alias": " - high_school_computer_science",
370
+ "acc,none": 0.79,
371
+ "acc_stderr,none": 0.040936018074033256
372
+ },
373
+ "harness|mmlu_high_school_mathematics|0": {
374
+ "alias": " - high_school_mathematics",
375
+ "acc,none": 0.5,
376
+ "acc_stderr,none": 0.030485538042484616
377
+ },
378
+ "harness|mmlu_high_school_physics|0": {
379
+ "alias": " - high_school_physics",
380
+ "acc,none": 0.45695364238410596,
381
+ "acc_stderr,none": 0.04067325174247443
382
+ },
383
+ "harness|mmlu_high_school_statistics|0": {
384
+ "alias": " - high_school_statistics",
385
+ "acc,none": 0.6342592592592593,
386
+ "acc_stderr,none": 0.032847388576472056
387
+ },
388
+ "harness|mmlu_machine_learning|0": {
389
+ "alias": " - machine_learning",
390
+ "acc,none": 0.5535714285714286,
391
+ "acc_stderr,none": 0.04718471485219587
392
+ }
393
+ },
394
+ "task_info": {
395
+ "model": "Intel/Qwen2-7B-int4-inc",
396
+ "local": true,
397
+ "revision": "main",
398
+ "private": false,
399
+ "params": 7,
400
+ "architectures": "QwenForCausalLM",
401
+ "quant_type": "AutoRound",
402
+ "precision": "4bit",
403
+ "model_params": 7,
404
+ "model_size": 5.2,
405
+ "weight_dtype": "int4",
406
+ "compute_dtype": "float16",
407
+ "gguf_ftype": "*Q4_0.gguf",
408
+ "hardware": "gpu",
409
+ "status": "Finished",
410
+ "submitted_time": "2024-04-23T15:44:22Z",
411
+ "model_type": "quantization",
412
+ "job_id": -1,
413
+ "job_start_time": null,
414
+ "scripts": "ITREX"
415
+ },
416
+ "quantization_config": {
417
+ "amp": true,
418
+ "autoround_version": "0.2.1.dev",
419
+ "backend": "autoround:exllamav2",
420
+ "bits": 4,
421
+ "data_type": "int",
422
+ "dataset": "NeelNanda/pile-10k",
423
+ "enable_minmax_tuning": true,
424
+ "enable_quanted_input": true,
425
+ "gradient_accumulate_steps": 1,
426
+ "group_size": 128,
427
+ "iters": 1000,
428
+ "low_gpu_mem_usage": false,
429
+ "lr": 0.001,
430
+ "minmax_lr": 0.001,
431
+ "n_samples": 512,
432
+ "quant_method": "intel/auto-round",
433
+ "scale_dtype": "torch.float16",
434
+ "seqlen": 2048,
435
+ "sym": false,
436
+ "train_bs": 8
437
+ },
438
+ "versions": {
439
+ "harness|truthfulqa:mc2|0": 2.0,
440
+ "harness|arc:easy|0": 1.0,
441
+ "harness|lambada:openai|0": 1.0,
442
+ "harness|arc:challenge|0": 1.0,
443
+ "harness|boolq|0": 2.0,
444
+ "harness|truthfulqa:mc1|0": 2.0,
445
+ "harness|openbookqa|0": 1.0,
446
+ "harness|winogrande|0": 1.0,
447
+ "harness|hellaswag|0": 1.0,
448
+ "harness|piqa|0": 1.0,
449
+ "harness|mmlu|0": null,
450
+ "harness|mmlu_humanities|0": null,
451
+ "harness|mmlu_formal_logic|0": 0.0,
452
+ "harness|mmlu_high_school_european_history|0": 0.0,
453
+ "harness|mmlu_high_school_us_history|0": 0.0,
454
+ "harness|mmlu_high_school_world_history|0": 0.0,
455
+ "harness|mmlu_international_law|0": 0.0,
456
+ "harness|mmlu_jurisprudence|0": 0.0,
457
+ "harness|mmlu_logical_fallacies|0": 0.0,
458
+ "harness|mmlu_moral_disputes|0": 0.0,
459
+ "harness|mmlu_moral_scenarios|0": 0.0,
460
+ "harness|mmlu_philosophy|0": 0.0,
461
+ "harness|mmlu_prehistory|0": 0.0,
462
+ "harness|mmlu_professional_law|0": 0.0,
463
+ "harness|mmlu_world_religions|0": 0.0,
464
+ "harness|mmlu_other|0": null,
465
+ "harness|mmlu_business_ethics|0": 0.0,
466
+ "harness|mmlu_clinical_knowledge|0": 0.0,
467
+ "harness|mmlu_college_medicine|0": 0.0,
468
+ "harness|mmlu_global_facts|0": 0.0,
469
+ "harness|mmlu_human_aging|0": 0.0,
470
+ "harness|mmlu_management|0": 0.0,
471
+ "harness|mmlu_marketing|0": 0.0,
472
+ "harness|mmlu_medical_genetics|0": 0.0,
473
+ "harness|mmlu_miscellaneous|0": 0.0,
474
+ "harness|mmlu_nutrition|0": 0.0,
475
+ "harness|mmlu_professional_accounting|0": 0.0,
476
+ "harness|mmlu_professional_medicine|0": 0.0,
477
+ "harness|mmlu_virology|0": 0.0,
478
+ "harness|mmlu_social_sciences|0": null,
479
+ "harness|mmlu_econometrics|0": 0.0,
480
+ "harness|mmlu_high_school_geography|0": 0.0,
481
+ "harness|mmlu_high_school_government_and_politics|0": 0.0,
482
+ "harness|mmlu_high_school_macroeconomics|0": 0.0,
483
+ "harness|mmlu_high_school_microeconomics|0": 0.0,
484
+ "harness|mmlu_high_school_psychology|0": 0.0,
485
+ "harness|mmlu_human_sexuality|0": 0.0,
486
+ "harness|mmlu_professional_psychology|0": 0.0,
487
+ "harness|mmlu_public_relations|0": 0.0,
488
+ "harness|mmlu_security_studies|0": 0.0,
489
+ "harness|mmlu_sociology|0": 0.0,
490
+ "harness|mmlu_us_foreign_policy|0": 0.0,
491
+ "harness|mmlu_stem|0": null,
492
+ "harness|mmlu_abstract_algebra|0": 0.0,
493
+ "harness|mmlu_anatomy|0": 0.0,
494
+ "harness|mmlu_astronomy|0": 0.0,
495
+ "harness|mmlu_college_biology|0": 0.0,
496
+ "harness|mmlu_college_chemistry|0": 0.0,
497
+ "harness|mmlu_college_computer_science|0": 0.0,
498
+ "harness|mmlu_college_mathematics|0": 0.0,
499
+ "harness|mmlu_college_physics|0": 0.0,
500
+ "harness|mmlu_computer_security|0": 0.0,
501
+ "harness|mmlu_conceptual_physics|0": 0.0,
502
+ "harness|mmlu_electrical_engineering|0": 0.0,
503
+ "harness|mmlu_elementary_mathematics|0": 0.0,
504
+ "harness|mmlu_high_school_biology|0": 0.0,
505
+ "harness|mmlu_high_school_chemistry|0": 0.0,
506
+ "harness|mmlu_high_school_computer_science|0": 0.0,
507
+ "harness|mmlu_high_school_mathematics|0": 0.0,
508
+ "harness|mmlu_high_school_physics|0": 0.0,
509
+ "harness|mmlu_high_school_statistics|0": 0.0,
510
+ "harness|mmlu_machine_learning|0": 0.0
511
+ },
512
+ "n-shot": {
513
+ "arc_challenge": 0,
514
+ "arc_easy": 0,
515
+ "boolq": 0,
516
+ "hellaswag": 0,
517
+ "lambada_openai": 0,
518
+ "mmlu": 0,
519
+ "mmlu_abstract_algebra": 0,
520
+ "mmlu_anatomy": 0,
521
+ "mmlu_astronomy": 0,
522
+ "mmlu_business_ethics": 0,
523
+ "mmlu_clinical_knowledge": 0,
524
+ "mmlu_college_biology": 0,
525
+ "mmlu_college_chemistry": 0,
526
+ "mmlu_college_computer_science": 0,
527
+ "mmlu_college_mathematics": 0,
528
+ "mmlu_college_medicine": 0,
529
+ "mmlu_college_physics": 0,
530
+ "mmlu_computer_security": 0,
531
+ "mmlu_conceptual_physics": 0,
532
+ "mmlu_econometrics": 0,
533
+ "mmlu_electrical_engineering": 0,
534
+ "mmlu_elementary_mathematics": 0,
535
+ "mmlu_formal_logic": 0,
536
+ "mmlu_global_facts": 0,
537
+ "mmlu_high_school_biology": 0,
538
+ "mmlu_high_school_chemistry": 0,
539
+ "mmlu_high_school_computer_science": 0,
540
+ "mmlu_high_school_european_history": 0,
541
+ "mmlu_high_school_geography": 0,
542
+ "mmlu_high_school_government_and_politics": 0,
543
+ "mmlu_high_school_macroeconomics": 0,
544
+ "mmlu_high_school_mathematics": 0,
545
+ "mmlu_high_school_microeconomics": 0,
546
+ "mmlu_high_school_physics": 0,
547
+ "mmlu_high_school_psychology": 0,
548
+ "mmlu_high_school_statistics": 0,
549
+ "mmlu_high_school_us_history": 0,
550
+ "mmlu_high_school_world_history": 0,
551
+ "mmlu_human_aging": 0,
552
+ "mmlu_human_sexuality": 0,
553
+ "mmlu_humanities": 0,
554
+ "mmlu_international_law": 0,
555
+ "mmlu_jurisprudence": 0,
556
+ "mmlu_logical_fallacies": 0,
557
+ "mmlu_machine_learning": 0,
558
+ "mmlu_management": 0,
559
+ "mmlu_marketing": 0,
560
+ "mmlu_medical_genetics": 0,
561
+ "mmlu_miscellaneous": 0,
562
+ "mmlu_moral_disputes": 0,
563
+ "mmlu_moral_scenarios": 0,
564
+ "mmlu_nutrition": 0,
565
+ "mmlu_other": 0,
566
+ "mmlu_philosophy": 0,
567
+ "mmlu_prehistory": 0,
568
+ "mmlu_professional_accounting": 0,
569
+ "mmlu_professional_law": 0,
570
+ "mmlu_professional_medicine": 0,
571
+ "mmlu_professional_psychology": 0,
572
+ "mmlu_public_relations": 0,
573
+ "mmlu_security_studies": 0,
574
+ "mmlu_social_sciences": 0,
575
+ "mmlu_sociology": 0,
576
+ "mmlu_stem": 0,
577
+ "mmlu_us_foreign_policy": 0,
578
+ "mmlu_virology": 0,
579
+ "mmlu_world_religions": 0,
580
+ "openbookqa": 0,
581
+ "piqa": 0,
582
+ "truthfulqa_mc1": 0,
583
+ "truthfulqa_mc2": 0,
584
+ "winogrande": 0
585
+ },
586
+ "date": 1717599011.729154,
587
+ "config": {
588
+ "model": "hf",
589
+ "model_args": "pretrained=Intel/Qwen2-7B-int4-inc,trust_remote_code=True,dtype=float16,_commit_hash=main",
590
+ "batch_size": 2,
591
+ "batch_sizes": [],
592
+ "device": "cuda",
593
+ "use_cache": null,
594
+ "limit": null,
595
+ "bootstrap_iters": 100000,
596
+ "gen_kwargs": null
597
+ }
598
+ }
Intel/results_2024-06-06-09-32-36.json ADDED
@@ -0,0 +1,596 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config_general": {
3
+ "lighteval_sha": "1.4",
4
+ "num_few_shot_default": null,
5
+ "num_fewshot_seeds": null,
6
+ "override_batch_size": null,
7
+ "max_samples": null,
8
+ "job_id": -1,
9
+ "start_time": null,
10
+ "end_time": "2024-06-06-09-32-36",
11
+ "total_evaluation_time_secondes": "",
12
+ "model_name": "Intel/Qwen2-1.5B-Instuct-int4-inc",
13
+ "model_sha": "",
14
+ "model_dtype": "4bit",
15
+ "model_size": 1.6,
16
+ "model_params": 1.5,
17
+ "quant_type": "AutoRound",
18
+ "precision": "4bit"
19
+ },
20
+ "results": {
21
+ "harness|lambada:openai|0": {
22
+ "perplexity,none": 5.285488721585025,
23
+ "perplexity_stderr,none": 0.1356474096089869,
24
+ "acc,none": 0.6411798952066757,
25
+ "acc_stderr,none": 0.006682522124089319,
26
+ "alias": "lambada_openai"
27
+ },
28
+ "harness|openbookqa|0": {
29
+ "acc,none": 0.272,
30
+ "acc_stderr,none": 0.01992048320956607,
31
+ "acc_norm,none": 0.36,
32
+ "acc_norm_stderr,none": 0.02148775108972052,
33
+ "alias": "openbookqa"
34
+ },
35
+ "harness|mmlu|0": {
36
+ "acc,none": 0.5502777382139297,
37
+ "acc_stderr,none": 0.004026718834956645,
38
+ "alias": "mmlu"
39
+ },
40
+ "harness|mmlu_humanities|0": {
41
+ "alias": " - humanities",
42
+ "acc,none": 0.5030818278427205,
43
+ "acc_stderr,none": 0.006947472831518005
44
+ },
45
+ "harness|mmlu_formal_logic|0": {
46
+ "alias": " - formal_logic",
47
+ "acc,none": 0.3888888888888889,
48
+ "acc_stderr,none": 0.04360314860077459
49
+ },
50
+ "harness|mmlu_high_school_european_history|0": {
51
+ "alias": " - high_school_european_history",
52
+ "acc,none": 0.6848484848484848,
53
+ "acc_stderr,none": 0.0362773057502241
54
+ },
55
+ "harness|mmlu_high_school_us_history|0": {
56
+ "alias": " - high_school_us_history",
57
+ "acc,none": 0.6715686274509803,
58
+ "acc_stderr,none": 0.032962451101722294
59
+ },
60
+ "harness|mmlu_high_school_world_history|0": {
61
+ "alias": " - high_school_world_history",
62
+ "acc,none": 0.7341772151898734,
63
+ "acc_stderr,none": 0.02875679962965834
64
+ },
65
+ "harness|mmlu_international_law|0": {
66
+ "alias": " - international_law",
67
+ "acc,none": 0.7272727272727273,
68
+ "acc_stderr,none": 0.04065578140908705
69
+ },
70
+ "harness|mmlu_jurisprudence|0": {
71
+ "alias": " - jurisprudence",
72
+ "acc,none": 0.6481481481481481,
73
+ "acc_stderr,none": 0.046166311118017146
74
+ },
75
+ "harness|mmlu_logical_fallacies|0": {
76
+ "alias": " - logical_fallacies",
77
+ "acc,none": 0.6993865030674846,
78
+ "acc_stderr,none": 0.03602511318806771
79
+ },
80
+ "harness|mmlu_moral_disputes|0": {
81
+ "alias": " - moral_disputes",
82
+ "acc,none": 0.6242774566473989,
83
+ "acc_stderr,none": 0.026074314851657083
84
+ },
85
+ "harness|mmlu_moral_scenarios|0": {
86
+ "alias": " - moral_scenarios",
87
+ "acc,none": 0.2837988826815642,
88
+ "acc_stderr,none": 0.01507835897075178
89
+ },
90
+ "harness|mmlu_philosophy|0": {
91
+ "alias": " - philosophy",
92
+ "acc,none": 0.6334405144694534,
93
+ "acc_stderr,none": 0.027368078243971628
94
+ },
95
+ "harness|mmlu_prehistory|0": {
96
+ "alias": " - prehistory",
97
+ "acc,none": 0.5555555555555556,
98
+ "acc_stderr,none": 0.02764847787741332
99
+ },
100
+ "harness|mmlu_professional_law|0": {
101
+ "alias": " - professional_law",
102
+ "acc,none": 0.42633637548891784,
103
+ "acc_stderr,none": 0.012630884771599698
104
+ },
105
+ "harness|mmlu_world_religions|0": {
106
+ "alias": " - world_religions",
107
+ "acc,none": 0.7076023391812866,
108
+ "acc_stderr,none": 0.034886477134579215
109
+ },
110
+ "harness|mmlu_other|0": {
111
+ "alias": " - other",
112
+ "acc,none": 0.6195687158030254,
113
+ "acc_stderr,none": 0.00842476546352716
114
+ },
115
+ "harness|mmlu_business_ethics|0": {
116
+ "alias": " - business_ethics",
117
+ "acc,none": 0.65,
118
+ "acc_stderr,none": 0.047937248544110196
119
+ },
120
+ "harness|mmlu_clinical_knowledge|0": {
121
+ "alias": " - clinical_knowledge",
122
+ "acc,none": 0.6075471698113207,
123
+ "acc_stderr,none": 0.03005258057955784
124
+ },
125
+ "harness|mmlu_college_medicine|0": {
126
+ "alias": " - college_medicine",
127
+ "acc,none": 0.5606936416184971,
128
+ "acc_stderr,none": 0.03784271932887467
129
+ },
130
+ "harness|mmlu_global_facts|0": {
131
+ "alias": " - global_facts",
132
+ "acc,none": 0.23,
133
+ "acc_stderr,none": 0.04229525846816506
134
+ },
135
+ "harness|mmlu_human_aging|0": {
136
+ "alias": " - human_aging",
137
+ "acc,none": 0.6143497757847534,
138
+ "acc_stderr,none": 0.03266842214289202
139
+ },
140
+ "harness|mmlu_management|0": {
141
+ "alias": " - management",
142
+ "acc,none": 0.7864077669902912,
143
+ "acc_stderr,none": 0.040580420156460344
144
+ },
145
+ "harness|mmlu_marketing|0": {
146
+ "alias": " - marketing",
147
+ "acc,none": 0.811965811965812,
148
+ "acc_stderr,none": 0.025598193686652244
149
+ },
150
+ "harness|mmlu_medical_genetics|0": {
151
+ "alias": " - medical_genetics",
152
+ "acc,none": 0.62,
153
+ "acc_stderr,none": 0.048783173121456316
154
+ },
155
+ "harness|mmlu_miscellaneous|0": {
156
+ "alias": " - miscellaneous",
157
+ "acc,none": 0.7075351213282248,
158
+ "acc_stderr,none": 0.016267000684598642
159
+ },
160
+ "harness|mmlu_nutrition|0": {
161
+ "alias": " - nutrition",
162
+ "acc,none": 0.6830065359477124,
163
+ "acc_stderr,none": 0.026643278474508755
164
+ },
165
+ "harness|mmlu_professional_accounting|0": {
166
+ "alias": " - professional_accounting",
167
+ "acc,none": 0.44680851063829785,
168
+ "acc_stderr,none": 0.02965823509766691
169
+ },
170
+ "harness|mmlu_professional_medicine|0": {
171
+ "alias": " - professional_medicine",
172
+ "acc,none": 0.5367647058823529,
173
+ "acc_stderr,none": 0.030290619180485697
174
+ },
175
+ "harness|mmlu_virology|0": {
176
+ "alias": " - virology",
177
+ "acc,none": 0.4457831325301205,
178
+ "acc_stderr,none": 0.03869543323472101
179
+ },
180
+ "harness|mmlu_social_sciences|0": {
181
+ "alias": " - social_sciences",
182
+ "acc,none": 0.6379590510237244,
183
+ "acc_stderr,none": 0.008481043207876597
184
+ },
185
+ "harness|mmlu_econometrics|0": {
186
+ "alias": " - econometrics",
187
+ "acc,none": 0.3508771929824561,
188
+ "acc_stderr,none": 0.04489539350270701
189
+ },
190
+ "harness|mmlu_high_school_geography|0": {
191
+ "alias": " - high_school_geography",
192
+ "acc,none": 0.7121212121212122,
193
+ "acc_stderr,none": 0.03225883512300992
194
+ },
195
+ "harness|mmlu_high_school_government_and_politics|0": {
196
+ "alias": " - high_school_government_and_politics",
197
+ "acc,none": 0.7512953367875648,
198
+ "acc_stderr,none": 0.031195840877700293
199
+ },
200
+ "harness|mmlu_high_school_macroeconomics|0": {
201
+ "alias": " - high_school_macroeconomics",
202
+ "acc,none": 0.5743589743589743,
203
+ "acc_stderr,none": 0.025069094387296535
204
+ },
205
+ "harness|mmlu_high_school_microeconomics|0": {
206
+ "alias": " - high_school_microeconomics",
207
+ "acc,none": 0.5798319327731093,
208
+ "acc_stderr,none": 0.03206183783236152
209
+ },
210
+ "harness|mmlu_high_school_psychology|0": {
211
+ "alias": " - high_school_psychology",
212
+ "acc,none": 0.7339449541284404,
213
+ "acc_stderr,none": 0.018946022322225593
214
+ },
215
+ "harness|mmlu_human_sexuality|0": {
216
+ "alias": " - human_sexuality",
217
+ "acc,none": 0.6412213740458015,
218
+ "acc_stderr,none": 0.04206739313864908
219
+ },
220
+ "harness|mmlu_professional_psychology|0": {
221
+ "alias": " - professional_psychology",
222
+ "acc,none": 0.5359477124183006,
223
+ "acc_stderr,none": 0.020175488765484043
224
+ },
225
+ "harness|mmlu_public_relations|0": {
226
+ "alias": " - public_relations",
227
+ "acc,none": 0.5909090909090909,
228
+ "acc_stderr,none": 0.04709306978661896
229
+ },
230
+ "harness|mmlu_security_studies|0": {
231
+ "alias": " - security_studies",
232
+ "acc,none": 0.6857142857142857,
233
+ "acc_stderr,none": 0.029719329422417468
234
+ },
235
+ "harness|mmlu_sociology|0": {
236
+ "alias": " - sociology",
237
+ "acc,none": 0.7661691542288557,
238
+ "acc_stderr,none": 0.029929415408348387
239
+ },
240
+ "harness|mmlu_us_foreign_policy|0": {
241
+ "alias": " - us_foreign_policy",
242
+ "acc,none": 0.76,
243
+ "acc_stderr,none": 0.04292346959909282
244
+ },
245
+ "harness|mmlu_stem|0": {
246
+ "alias": " - stem",
247
+ "acc,none": 0.4668569616238503,
248
+ "acc_stderr,none": 0.008757565781117154
249
+ },
250
+ "harness|mmlu_abstract_algebra|0": {
251
+ "alias": " - abstract_algebra",
252
+ "acc,none": 0.35,
253
+ "acc_stderr,none": 0.0479372485441102
254
+ },
255
+ "harness|mmlu_anatomy|0": {
256
+ "alias": " - anatomy",
257
+ "acc,none": 0.48148148148148145,
258
+ "acc_stderr,none": 0.043163785995113245
259
+ },
260
+ "harness|mmlu_astronomy|0": {
261
+ "alias": " - astronomy",
262
+ "acc,none": 0.5789473684210527,
263
+ "acc_stderr,none": 0.040179012759817494
264
+ },
265
+ "harness|mmlu_college_biology|0": {
266
+ "alias": " - college_biology",
267
+ "acc,none": 0.4930555555555556,
268
+ "acc_stderr,none": 0.04180806750294938
269
+ },
270
+ "harness|mmlu_college_chemistry|0": {
271
+ "alias": " - college_chemistry",
272
+ "acc,none": 0.42,
273
+ "acc_stderr,none": 0.049604496374885836
274
+ },
275
+ "harness|mmlu_college_computer_science|0": {
276
+ "alias": " - college_computer_science",
277
+ "acc,none": 0.48,
278
+ "acc_stderr,none": 0.050211673156867795
279
+ },
280
+ "harness|mmlu_college_mathematics|0": {
281
+ "alias": " - college_mathematics",
282
+ "acc,none": 0.31,
283
+ "acc_stderr,none": 0.04648231987117316
284
+ },
285
+ "harness|mmlu_college_physics|0": {
286
+ "alias": " - college_physics",
287
+ "acc,none": 0.3137254901960784,
288
+ "acc_stderr,none": 0.04617034827006718
289
+ },
290
+ "harness|mmlu_computer_security|0": {
291
+ "alias": " - computer_security",
292
+ "acc,none": 0.65,
293
+ "acc_stderr,none": 0.0479372485441102
294
+ },
295
+ "harness|mmlu_conceptual_physics|0": {
296
+ "alias": " - conceptual_physics",
297
+ "acc,none": 0.4553191489361702,
298
+ "acc_stderr,none": 0.03255525359340355
299
+ },
300
+ "harness|mmlu_electrical_engineering|0": {
301
+ "alias": " - electrical_engineering",
302
+ "acc,none": 0.5655172413793104,
303
+ "acc_stderr,none": 0.04130740879555498
304
+ },
305
+ "harness|mmlu_elementary_mathematics|0": {
306
+ "alias": " - elementary_mathematics",
307
+ "acc,none": 0.4444444444444444,
308
+ "acc_stderr,none": 0.025591857761382182
309
+ },
310
+ "harness|mmlu_high_school_biology|0": {
311
+ "alias": " - high_school_biology",
312
+ "acc,none": 0.6096774193548387,
313
+ "acc_stderr,none": 0.027751256636969576
314
+ },
315
+ "harness|mmlu_high_school_chemistry|0": {
316
+ "alias": " - high_school_chemistry",
317
+ "acc,none": 0.5073891625615764,
318
+ "acc_stderr,none": 0.035176035403610105
319
+ },
320
+ "harness|mmlu_high_school_computer_science|0": {
321
+ "alias": " - high_school_computer_science",
322
+ "acc,none": 0.57,
323
+ "acc_stderr,none": 0.049756985195624284
324
+ },
325
+ "harness|mmlu_high_school_mathematics|0": {
326
+ "alias": " - high_school_mathematics",
327
+ "acc,none": 0.37777777777777777,
328
+ "acc_stderr,none": 0.029560707392465718
329
+ },
330
+ "harness|mmlu_high_school_physics|0": {
331
+ "alias": " - high_school_physics",
332
+ "acc,none": 0.33112582781456956,
333
+ "acc_stderr,none": 0.038425817186598696
334
+ },
335
+ "harness|mmlu_high_school_statistics|0": {
336
+ "alias": " - high_school_statistics",
337
+ "acc,none": 0.4212962962962963,
338
+ "acc_stderr,none": 0.03367462138896078
339
+ },
340
+ "harness|mmlu_machine_learning|0": {
341
+ "alias": " - machine_learning",
342
+ "acc,none": 0.4107142857142857,
343
+ "acc_stderr,none": 0.04669510663875191
344
+ },
345
+ "harness|arc:challenge|0": {
346
+ "acc,none": 0.36945392491467577,
347
+ "acc_stderr,none": 0.01410457836649189,
348
+ "acc_norm,none": 0.39761092150170646,
349
+ "acc_norm_stderr,none": 0.014301752223279535,
350
+ "alias": "arc_challenge"
351
+ },
352
+ "harness|winogrande|0": {
353
+ "acc,none": 0.6606156274664562,
354
+ "acc_stderr,none": 0.01330771492894175,
355
+ "alias": "winogrande"
356
+ },
357
+ "harness|boolq|0": {
358
+ "acc,none": 0.7629969418960245,
359
+ "acc_stderr,none": 0.007437567381277126,
360
+ "alias": "boolq"
361
+ },
362
+ "harness|arc:easy|0": {
363
+ "acc,none": 0.6957070707070707,
364
+ "acc_stderr,none": 0.009441202922359185,
365
+ "acc_norm,none": 0.6620370370370371,
366
+ "acc_norm_stderr,none": 0.00970608053863286,
367
+ "alias": "arc_easy"
368
+ },
369
+ "harness|hellaswag|0": {
370
+ "acc,none": 0.4856602270464051,
371
+ "acc_stderr,none": 0.004987728900897589,
372
+ "acc_norm,none": 0.6493726349332802,
373
+ "acc_norm_stderr,none": 0.004761912511707522,
374
+ "alias": "hellaswag"
375
+ },
376
+ "harness|truthfulqa:mc1|0": {
377
+ "acc,none": 0.27539779681762544,
378
+ "acc_stderr,none": 0.01563813566777552,
379
+ "alias": "truthfulqa_mc1"
380
+ },
381
+ "harness|truthfulqa:mc2|0": {
382
+ "acc,none": 0.4316038703653941,
383
+ "acc_stderr,none": 0.014556182158523336,
384
+ "alias": "truthfulqa_mc2"
385
+ },
386
+ "harness|piqa|0": {
387
+ "acc,none": 0.7562568008705114,
388
+ "acc_stderr,none": 0.010017199471500612,
389
+ "acc_norm,none": 0.7535364526659413,
390
+ "acc_norm_stderr,none": 0.01005481078967182,
391
+ "alias": "piqa"
392
+ }
393
+ },
394
+ "task_info": {
395
+ "model": "Intel/Qwen2-1.5B-Instuct-int4-inc",
396
+ "local": true,
397
+ "revision": "main",
398
+ "private": false,
399
+ "params": 1.5,
400
+ "architectures": "QwenForCausalLM",
401
+ "quant_type": "AutoRound",
402
+ "precision": "4bit",
403
+ "model_params": 1.5,
404
+ "model_size": 1.6,
405
+ "weight_dtype": "int4",
406
+ "compute_dtype": "float16",
407
+ "gguf_ftype": "*Q4_0.gguf",
408
+ "hardware": "gpu",
409
+ "status": "Finished",
410
+ "submitted_time": "2024-04-23T15:44:22Z",
411
+ "model_type": "quantization",
412
+ "job_id": -1,
413
+ "job_start_time": null,
414
+ "scripts": "ITREX"
415
+ },
416
+ "quantization_config": {
417
+ "autoround_version": "0.2.1.dev",
418
+ "bits": 4,
419
+ "damp_percent": 0.01,
420
+ "desc_act": false,
421
+ "enable_minmax_tuning": true,
422
+ "enable_quanted_input": true,
423
+ "group_size": 32,
424
+ "is_marlin_format": false,
425
+ "iters": 1000,
426
+ "lr": 0.001,
427
+ "minmax_lr": 0.001,
428
+ "model_file_base_name": "model",
429
+ "model_name_or_path": null,
430
+ "quant_method": "gptq",
431
+ "scale_dtype": "float16",
432
+ "static_groups": false,
433
+ "sym": true,
434
+ "true_sequential": false
435
+ },
436
+ "versions": {
437
+ "harness|lambada:openai|0": 1.0,
438
+ "harness|openbookqa|0": 1.0,
439
+ "harness|mmlu|0": null,
440
+ "harness|mmlu_humanities|0": null,
441
+ "harness|mmlu_formal_logic|0": 0.0,
442
+ "harness|mmlu_high_school_european_history|0": 0.0,
443
+ "harness|mmlu_high_school_us_history|0": 0.0,
444
+ "harness|mmlu_high_school_world_history|0": 0.0,
445
+ "harness|mmlu_international_law|0": 0.0,
446
+ "harness|mmlu_jurisprudence|0": 0.0,
447
+ "harness|mmlu_logical_fallacies|0": 0.0,
448
+ "harness|mmlu_moral_disputes|0": 0.0,
449
+ "harness|mmlu_moral_scenarios|0": 0.0,
450
+ "harness|mmlu_philosophy|0": 0.0,
451
+ "harness|mmlu_prehistory|0": 0.0,
452
+ "harness|mmlu_professional_law|0": 0.0,
453
+ "harness|mmlu_world_religions|0": 0.0,
454
+ "harness|mmlu_other|0": null,
455
+ "harness|mmlu_business_ethics|0": 0.0,
456
+ "harness|mmlu_clinical_knowledge|0": 0.0,
457
+ "harness|mmlu_college_medicine|0": 0.0,
458
+ "harness|mmlu_global_facts|0": 0.0,
459
+ "harness|mmlu_human_aging|0": 0.0,
460
+ "harness|mmlu_management|0": 0.0,
461
+ "harness|mmlu_marketing|0": 0.0,
462
+ "harness|mmlu_medical_genetics|0": 0.0,
463
+ "harness|mmlu_miscellaneous|0": 0.0,
464
+ "harness|mmlu_nutrition|0": 0.0,
465
+ "harness|mmlu_professional_accounting|0": 0.0,
466
+ "harness|mmlu_professional_medicine|0": 0.0,
467
+ "harness|mmlu_virology|0": 0.0,
468
+ "harness|mmlu_social_sciences|0": null,
469
+ "harness|mmlu_econometrics|0": 0.0,
470
+ "harness|mmlu_high_school_geography|0": 0.0,
471
+ "harness|mmlu_high_school_government_and_politics|0": 0.0,
472
+ "harness|mmlu_high_school_macroeconomics|0": 0.0,
473
+ "harness|mmlu_high_school_microeconomics|0": 0.0,
474
+ "harness|mmlu_high_school_psychology|0": 0.0,
475
+ "harness|mmlu_human_sexuality|0": 0.0,
476
+ "harness|mmlu_professional_psychology|0": 0.0,
477
+ "harness|mmlu_public_relations|0": 0.0,
478
+ "harness|mmlu_security_studies|0": 0.0,
479
+ "harness|mmlu_sociology|0": 0.0,
480
+ "harness|mmlu_us_foreign_policy|0": 0.0,
481
+ "harness|mmlu_stem|0": null,
482
+ "harness|mmlu_abstract_algebra|0": 0.0,
483
+ "harness|mmlu_anatomy|0": 0.0,
484
+ "harness|mmlu_astronomy|0": 0.0,
485
+ "harness|mmlu_college_biology|0": 0.0,
486
+ "harness|mmlu_college_chemistry|0": 0.0,
487
+ "harness|mmlu_college_computer_science|0": 0.0,
488
+ "harness|mmlu_college_mathematics|0": 0.0,
489
+ "harness|mmlu_college_physics|0": 0.0,
490
+ "harness|mmlu_computer_security|0": 0.0,
491
+ "harness|mmlu_conceptual_physics|0": 0.0,
492
+ "harness|mmlu_electrical_engineering|0": 0.0,
493
+ "harness|mmlu_elementary_mathematics|0": 0.0,
494
+ "harness|mmlu_high_school_biology|0": 0.0,
495
+ "harness|mmlu_high_school_chemistry|0": 0.0,
496
+ "harness|mmlu_high_school_computer_science|0": 0.0,
497
+ "harness|mmlu_high_school_mathematics|0": 0.0,
498
+ "harness|mmlu_high_school_physics|0": 0.0,
499
+ "harness|mmlu_high_school_statistics|0": 0.0,
500
+ "harness|mmlu_machine_learning|0": 0.0,
501
+ "harness|arc:challenge|0": 1.0,
502
+ "harness|winogrande|0": 1.0,
503
+ "harness|boolq|0": 2.0,
504
+ "harness|arc:easy|0": 1.0,
505
+ "harness|hellaswag|0": 1.0,
506
+ "harness|truthfulqa:mc1|0": 2.0,
507
+ "harness|truthfulqa:mc2|0": 2.0,
508
+ "harness|piqa|0": 1.0
509
+ },
510
+ "n-shot": {
511
+ "arc_challenge": 0,
512
+ "arc_easy": 0,
513
+ "boolq": 0,
514
+ "hellaswag": 0,
515
+ "lambada_openai": 0,
516
+ "mmlu": 0,
517
+ "mmlu_abstract_algebra": 0,
518
+ "mmlu_anatomy": 0,
519
+ "mmlu_astronomy": 0,
520
+ "mmlu_business_ethics": 0,
521
+ "mmlu_clinical_knowledge": 0,
522
+ "mmlu_college_biology": 0,
523
+ "mmlu_college_chemistry": 0,
524
+ "mmlu_college_computer_science": 0,
525
+ "mmlu_college_mathematics": 0,
526
+ "mmlu_college_medicine": 0,
527
+ "mmlu_college_physics": 0,
528
+ "mmlu_computer_security": 0,
529
+ "mmlu_conceptual_physics": 0,
530
+ "mmlu_econometrics": 0,
531
+ "mmlu_electrical_engineering": 0,
532
+ "mmlu_elementary_mathematics": 0,
533
+ "mmlu_formal_logic": 0,
534
+ "mmlu_global_facts": 0,
535
+ "mmlu_high_school_biology": 0,
536
+ "mmlu_high_school_chemistry": 0,
537
+ "mmlu_high_school_computer_science": 0,
538
+ "mmlu_high_school_european_history": 0,
539
+ "mmlu_high_school_geography": 0,
540
+ "mmlu_high_school_government_and_politics": 0,
541
+ "mmlu_high_school_macroeconomics": 0,
542
+ "mmlu_high_school_mathematics": 0,
543
+ "mmlu_high_school_microeconomics": 0,
544
+ "mmlu_high_school_physics": 0,
545
+ "mmlu_high_school_psychology": 0,
546
+ "mmlu_high_school_statistics": 0,
547
+ "mmlu_high_school_us_history": 0,
548
+ "mmlu_high_school_world_history": 0,
549
+ "mmlu_human_aging": 0,
550
+ "mmlu_human_sexuality": 0,
551
+ "mmlu_humanities": 0,
552
+ "mmlu_international_law": 0,
553
+ "mmlu_jurisprudence": 0,
554
+ "mmlu_logical_fallacies": 0,
555
+ "mmlu_machine_learning": 0,
556
+ "mmlu_management": 0,
557
+ "mmlu_marketing": 0,
558
+ "mmlu_medical_genetics": 0,
559
+ "mmlu_miscellaneous": 0,
560
+ "mmlu_moral_disputes": 0,
561
+ "mmlu_moral_scenarios": 0,
562
+ "mmlu_nutrition": 0,
563
+ "mmlu_other": 0,
564
+ "mmlu_philosophy": 0,
565
+ "mmlu_prehistory": 0,
566
+ "mmlu_professional_accounting": 0,
567
+ "mmlu_professional_law": 0,
568
+ "mmlu_professional_medicine": 0,
569
+ "mmlu_professional_psychology": 0,
570
+ "mmlu_public_relations": 0,
571
+ "mmlu_security_studies": 0,
572
+ "mmlu_social_sciences": 0,
573
+ "mmlu_sociology": 0,
574
+ "mmlu_stem": 0,
575
+ "mmlu_us_foreign_policy": 0,
576
+ "mmlu_virology": 0,
577
+ "mmlu_world_religions": 0,
578
+ "openbookqa": 0,
579
+ "piqa": 0,
580
+ "truthfulqa_mc1": 0,
581
+ "truthfulqa_mc2": 0,
582
+ "winogrande": 0
583
+ },
584
+ "date": 1717635926.568774,
585
+ "config": {
586
+ "model": "hf",
587
+ "model_args": "pretrained=Intel/Qwen2-1.5B-Instuct-int4-inc,trust_remote_code=True,dtype=float16,_commit_hash=main",
588
+ "batch_size": 4,
589
+ "batch_sizes": [],
590
+ "device": "cuda",
591
+ "use_cache": null,
592
+ "limit": null,
593
+ "bootstrap_iters": 100000,
594
+ "gen_kwargs": null
595
+ }
596
+ }
Intel/results_2024-06-06-10-10-10.json ADDED
@@ -0,0 +1,596 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config_general": {
3
+ "lighteval_sha": "1.4",
4
+ "num_few_shot_default": null,
5
+ "num_fewshot_seeds": null,
6
+ "override_batch_size": null,
7
+ "max_samples": null,
8
+ "job_id": -1,
9
+ "start_time": null,
10
+ "end_time": "2024-06-06-10-10-10",
11
+ "total_evaluation_time_secondes": "",
12
+ "model_name": "Intel/Qwen2-0.5B-Instuct-int4-inc",
13
+ "model_sha": "",
14
+ "model_dtype": "4bit",
15
+ "model_size": 0.719,
16
+ "model_params": 0.5,
17
+ "quant_type": "AutoRound",
18
+ "precision": "4bit"
19
+ },
20
+ "results": {
21
+ "harness|hellaswag|0": {
22
+ "acc,none": 0.3846843258315077,
23
+ "acc_stderr,none": 0.004855262903270792,
24
+ "acc_norm,none": 0.4843656642103167,
25
+ "acc_norm_stderr,none": 0.00498734148585666,
26
+ "alias": "hellaswag"
27
+ },
28
+ "harness|arc:easy|0": {
29
+ "acc,none": 0.5862794612794613,
30
+ "acc_stderr,none": 0.01010587853023814,
31
+ "acc_norm,none": 0.5395622895622896,
32
+ "acc_norm_stderr,none": 0.010227616386289008,
33
+ "alias": "arc_easy"
34
+ },
35
+ "harness|truthfulqa:mc1|0": {
36
+ "acc,none": 0.24724602203182375,
37
+ "acc_stderr,none": 0.01510240479735965,
38
+ "alias": "truthfulqa_mc1"
39
+ },
40
+ "harness|mmlu|0": {
41
+ "acc,none": 0.4223045150263495,
42
+ "acc_stderr,none": 0.004088092420326436,
43
+ "alias": "mmlu"
44
+ },
45
+ "harness|mmlu_humanities|0": {
46
+ "alias": " - humanities",
47
+ "acc,none": 0.3980871413390011,
48
+ "acc_stderr,none": 0.006959278348932478
49
+ },
50
+ "harness|mmlu_formal_logic|0": {
51
+ "alias": " - formal_logic",
52
+ "acc,none": 0.2619047619047619,
53
+ "acc_stderr,none": 0.039325376803928704
54
+ },
55
+ "harness|mmlu_high_school_european_history|0": {
56
+ "alias": " - high_school_european_history",
57
+ "acc,none": 0.5393939393939394,
58
+ "acc_stderr,none": 0.03892207016552013
59
+ },
60
+ "harness|mmlu_high_school_us_history|0": {
61
+ "alias": " - high_school_us_history",
62
+ "acc,none": 0.5392156862745098,
63
+ "acc_stderr,none": 0.03498501649369527
64
+ },
65
+ "harness|mmlu_high_school_world_history|0": {
66
+ "alias": " - high_school_world_history",
67
+ "acc,none": 0.5611814345991561,
68
+ "acc_stderr,none": 0.032302649315470375
69
+ },
70
+ "harness|mmlu_international_law|0": {
71
+ "alias": " - international_law",
72
+ "acc,none": 0.6528925619834711,
73
+ "acc_stderr,none": 0.04345724570292534
74
+ },
75
+ "harness|mmlu_jurisprudence|0": {
76
+ "alias": " - jurisprudence",
77
+ "acc,none": 0.5370370370370371,
78
+ "acc_stderr,none": 0.04820403072760628
79
+ },
80
+ "harness|mmlu_logical_fallacies|0": {
81
+ "alias": " - logical_fallacies",
82
+ "acc,none": 0.44785276073619634,
83
+ "acc_stderr,none": 0.03906947479456601
84
+ },
85
+ "harness|mmlu_moral_disputes|0": {
86
+ "alias": " - moral_disputes",
87
+ "acc,none": 0.5028901734104047,
88
+ "acc_stderr,none": 0.026918645383239004
89
+ },
90
+ "harness|mmlu_moral_scenarios|0": {
91
+ "alias": " - moral_scenarios",
92
+ "acc,none": 0.2424581005586592,
93
+ "acc_stderr,none": 0.014333522059217887
94
+ },
95
+ "harness|mmlu_philosophy|0": {
96
+ "alias": " - philosophy",
97
+ "acc,none": 0.4694533762057878,
98
+ "acc_stderr,none": 0.028345045864840684
99
+ },
100
+ "harness|mmlu_prehistory|0": {
101
+ "alias": " - prehistory",
102
+ "acc,none": 0.42592592592592593,
103
+ "acc_stderr,none": 0.027513747284379428
104
+ },
105
+ "harness|mmlu_professional_law|0": {
106
+ "alias": " - professional_law",
107
+ "acc,none": 0.3520208604954368,
108
+ "acc_stderr,none": 0.012198140605353609
109
+ },
110
+ "harness|mmlu_world_religions|0": {
111
+ "alias": " - world_religions",
112
+ "acc,none": 0.4853801169590643,
113
+ "acc_stderr,none": 0.038331852752130205
114
+ },
115
+ "harness|mmlu_other|0": {
116
+ "alias": " - other",
117
+ "acc,none": 0.467653685226907,
118
+ "acc_stderr,none": 0.008853315632851513
119
+ },
120
+ "harness|mmlu_business_ethics|0": {
121
+ "alias": " - business_ethics",
122
+ "acc,none": 0.54,
123
+ "acc_stderr,none": 0.05009082659620333
124
+ },
125
+ "harness|mmlu_clinical_knowledge|0": {
126
+ "alias": " - clinical_knowledge",
127
+ "acc,none": 0.4716981132075472,
128
+ "acc_stderr,none": 0.030723535249006107
129
+ },
130
+ "harness|mmlu_college_medicine|0": {
131
+ "alias": " - college_medicine",
132
+ "acc,none": 0.43352601156069365,
133
+ "acc_stderr,none": 0.03778621079092056
134
+ },
135
+ "harness|mmlu_global_facts|0": {
136
+ "alias": " - global_facts",
137
+ "acc,none": 0.36,
138
+ "acc_stderr,none": 0.04824181513244218
139
+ },
140
+ "harness|mmlu_human_aging|0": {
141
+ "alias": " - human_aging",
142
+ "acc,none": 0.4125560538116592,
143
+ "acc_stderr,none": 0.03304062175449296
144
+ },
145
+ "harness|mmlu_management|0": {
146
+ "alias": " - management",
147
+ "acc,none": 0.5825242718446602,
148
+ "acc_stderr,none": 0.048828405482122375
149
+ },
150
+ "harness|mmlu_marketing|0": {
151
+ "alias": " - marketing",
152
+ "acc,none": 0.6367521367521367,
153
+ "acc_stderr,none": 0.03150712523091264
154
+ },
155
+ "harness|mmlu_medical_genetics|0": {
156
+ "alias": " - medical_genetics",
157
+ "acc,none": 0.44,
158
+ "acc_stderr,none": 0.04988876515698589
159
+ },
160
+ "harness|mmlu_miscellaneous|0": {
161
+ "alias": " - miscellaneous",
162
+ "acc,none": 0.5019157088122606,
163
+ "acc_stderr,none": 0.017879832259026673
164
+ },
165
+ "harness|mmlu_nutrition|0": {
166
+ "alias": " - nutrition",
167
+ "acc,none": 0.5261437908496732,
168
+ "acc_stderr,none": 0.028590752958852394
169
+ },
170
+ "harness|mmlu_professional_accounting|0": {
171
+ "alias": " - professional_accounting",
172
+ "acc,none": 0.3723404255319149,
173
+ "acc_stderr,none": 0.02883892147125146
174
+ },
175
+ "harness|mmlu_professional_medicine|0": {
176
+ "alias": " - professional_medicine",
177
+ "acc,none": 0.3639705882352941,
178
+ "acc_stderr,none": 0.02922719246003203
179
+ },
180
+ "harness|mmlu_virology|0": {
181
+ "alias": " - virology",
182
+ "acc,none": 0.3614457831325301,
183
+ "acc_stderr,none": 0.037400593820293204
184
+ },
185
+ "harness|mmlu_social_sciences|0": {
186
+ "alias": " - social_sciences",
187
+ "acc,none": 0.48033799155021123,
188
+ "acc_stderr,none": 0.008913902203243567
189
+ },
190
+ "harness|mmlu_econometrics|0": {
191
+ "alias": " - econometrics",
192
+ "acc,none": 0.34210526315789475,
193
+ "acc_stderr,none": 0.04462917535336936
194
+ },
195
+ "harness|mmlu_high_school_geography|0": {
196
+ "alias": " - high_school_geography",
197
+ "acc,none": 0.5,
198
+ "acc_stderr,none": 0.035623524993954825
199
+ },
200
+ "harness|mmlu_high_school_government_and_politics|0": {
201
+ "alias": " - high_school_government_and_politics",
202
+ "acc,none": 0.5440414507772021,
203
+ "acc_stderr,none": 0.03594413711272435
204
+ },
205
+ "harness|mmlu_high_school_macroeconomics|0": {
206
+ "alias": " - high_school_macroeconomics",
207
+ "acc,none": 0.43333333333333335,
208
+ "acc_stderr,none": 0.025124653525885124
209
+ },
210
+ "harness|mmlu_high_school_microeconomics|0": {
211
+ "alias": " - high_school_microeconomics",
212
+ "acc,none": 0.42436974789915966,
213
+ "acc_stderr,none": 0.032104790510157764
214
+ },
215
+ "harness|mmlu_high_school_psychology|0": {
216
+ "alias": " - high_school_psychology",
217
+ "acc,none": 0.5467889908256881,
218
+ "acc_stderr,none": 0.021343255165546034
219
+ },
220
+ "harness|mmlu_human_sexuality|0": {
221
+ "alias": " - human_sexuality",
222
+ "acc,none": 0.4732824427480916,
223
+ "acc_stderr,none": 0.04379024936553894
224
+ },
225
+ "harness|mmlu_professional_psychology|0": {
226
+ "alias": " - professional_psychology",
227
+ "acc,none": 0.39052287581699346,
228
+ "acc_stderr,none": 0.019737008998094604
229
+ },
230
+ "harness|mmlu_public_relations|0": {
231
+ "alias": " - public_relations",
232
+ "acc,none": 0.5363636363636364,
233
+ "acc_stderr,none": 0.04776449162396197
234
+ },
235
+ "harness|mmlu_security_studies|0": {
236
+ "alias": " - security_studies",
237
+ "acc,none": 0.49795918367346936,
238
+ "acc_stderr,none": 0.0320089533497105
239
+ },
240
+ "harness|mmlu_sociology|0": {
241
+ "alias": " - sociology",
242
+ "acc,none": 0.5771144278606966,
243
+ "acc_stderr,none": 0.034932317774212816
244
+ },
245
+ "harness|mmlu_us_foreign_policy|0": {
246
+ "alias": " - us_foreign_policy",
247
+ "acc,none": 0.69,
248
+ "acc_stderr,none": 0.04648231987117316
249
+ },
250
+ "harness|mmlu_stem|0": {
251
+ "alias": " - stem",
252
+ "acc,none": 0.35712020298128766,
253
+ "acc_stderr,none": 0.008476247679652571
254
+ },
255
+ "harness|mmlu_abstract_algebra|0": {
256
+ "alias": " - abstract_algebra",
257
+ "acc,none": 0.29,
258
+ "acc_stderr,none": 0.04560480215720683
259
+ },
260
+ "harness|mmlu_anatomy|0": {
261
+ "alias": " - anatomy",
262
+ "acc,none": 0.4222222222222222,
263
+ "acc_stderr,none": 0.04266763404099582
264
+ },
265
+ "harness|mmlu_astronomy|0": {
266
+ "alias": " - astronomy",
267
+ "acc,none": 0.39473684210526316,
268
+ "acc_stderr,none": 0.039777499346220734
269
+ },
270
+ "harness|mmlu_college_biology|0": {
271
+ "alias": " - college_biology",
272
+ "acc,none": 0.3819444444444444,
273
+ "acc_stderr,none": 0.040629907841466674
274
+ },
275
+ "harness|mmlu_college_chemistry|0": {
276
+ "alias": " - college_chemistry",
277
+ "acc,none": 0.27,
278
+ "acc_stderr,none": 0.0446196043338474
279
+ },
280
+ "harness|mmlu_college_computer_science|0": {
281
+ "alias": " - college_computer_science",
282
+ "acc,none": 0.32,
283
+ "acc_stderr,none": 0.046882617226215034
284
+ },
285
+ "harness|mmlu_college_mathematics|0": {
286
+ "alias": " - college_mathematics",
287
+ "acc,none": 0.33,
288
+ "acc_stderr,none": 0.04725815626252604
289
+ },
290
+ "harness|mmlu_college_physics|0": {
291
+ "alias": " - college_physics",
292
+ "acc,none": 0.29411764705882354,
293
+ "acc_stderr,none": 0.045338381959297736
294
+ },
295
+ "harness|mmlu_computer_security|0": {
296
+ "alias": " - computer_security",
297
+ "acc,none": 0.41,
298
+ "acc_stderr,none": 0.049431107042371025
299
+ },
300
+ "harness|mmlu_conceptual_physics|0": {
301
+ "alias": " - conceptual_physics",
302
+ "acc,none": 0.40425531914893614,
303
+ "acc_stderr,none": 0.03208115750788684
304
+ },
305
+ "harness|mmlu_electrical_engineering|0": {
306
+ "alias": " - electrical_engineering",
307
+ "acc,none": 0.43448275862068964,
308
+ "acc_stderr,none": 0.04130740879555497
309
+ },
310
+ "harness|mmlu_elementary_mathematics|0": {
311
+ "alias": " - elementary_mathematics",
312
+ "acc,none": 0.3253968253968254,
313
+ "acc_stderr,none": 0.024130158299762602
314
+ },
315
+ "harness|mmlu_high_school_biology|0": {
316
+ "alias": " - high_school_biology",
317
+ "acc,none": 0.4612903225806452,
318
+ "acc_stderr,none": 0.028358634859836935
319
+ },
320
+ "harness|mmlu_high_school_chemistry|0": {
321
+ "alias": " - high_school_chemistry",
322
+ "acc,none": 0.3645320197044335,
323
+ "acc_stderr,none": 0.0338640574606209
324
+ },
325
+ "harness|mmlu_high_school_computer_science|0": {
326
+ "alias": " - high_school_computer_science",
327
+ "acc,none": 0.49,
328
+ "acc_stderr,none": 0.05024183937956911
329
+ },
330
+ "harness|mmlu_high_school_mathematics|0": {
331
+ "alias": " - high_school_mathematics",
332
+ "acc,none": 0.2740740740740741,
333
+ "acc_stderr,none": 0.027195934804085626
334
+ },
335
+ "harness|mmlu_high_school_physics|0": {
336
+ "alias": " - high_school_physics",
337
+ "acc,none": 0.2582781456953642,
338
+ "acc_stderr,none": 0.035737053147634576
339
+ },
340
+ "harness|mmlu_high_school_statistics|0": {
341
+ "alias": " - high_school_statistics",
342
+ "acc,none": 0.3101851851851852,
343
+ "acc_stderr,none": 0.03154696285656629
344
+ },
345
+ "harness|mmlu_machine_learning|0": {
346
+ "alias": " - machine_learning",
347
+ "acc,none": 0.3125,
348
+ "acc_stderr,none": 0.043994650575715215
349
+ },
350
+ "harness|piqa|0": {
351
+ "acc,none": 0.6855277475516867,
352
+ "acc_stderr,none": 0.010833009065106569,
353
+ "acc_norm,none": 0.6833514689880305,
354
+ "acc_norm_stderr,none": 0.010853160531978483,
355
+ "alias": "piqa"
356
+ },
357
+ "harness|boolq|0": {
358
+ "acc,none": 0.6281345565749236,
359
+ "acc_stderr,none": 0.00845301800735403,
360
+ "alias": "boolq"
361
+ },
362
+ "harness|openbookqa|0": {
363
+ "acc,none": 0.25,
364
+ "acc_stderr,none": 0.019384310743640384,
365
+ "acc_norm,none": 0.334,
366
+ "acc_norm_stderr,none": 0.021113492347743734,
367
+ "alias": "openbookqa"
368
+ },
369
+ "harness|arc:challenge|0": {
370
+ "acc,none": 0.2619453924914676,
371
+ "acc_stderr,none": 0.012849054826858108,
372
+ "acc_norm,none": 0.2883959044368601,
373
+ "acc_norm_stderr,none": 0.01323839442242817,
374
+ "alias": "arc_challenge"
375
+ },
376
+ "harness|truthfulqa:mc2|0": {
377
+ "acc,none": 0.39131629337724455,
378
+ "acc_stderr,none": 0.013943043500103466,
379
+ "alias": "truthfulqa_mc2"
380
+ },
381
+ "harness|winogrande|0": {
382
+ "acc,none": 0.5651144435674822,
383
+ "acc_stderr,none": 0.013932814110418027,
384
+ "alias": "winogrande"
385
+ },
386
+ "harness|lambada:openai|0": {
387
+ "perplexity,none": 12.554918626547046,
388
+ "perplexity_stderr,none": 0.4194570823476025,
389
+ "acc,none": 0.4867067727537357,
390
+ "acc_stderr,none": 0.006963515307693607,
391
+ "alias": "lambada_openai"
392
+ }
393
+ },
394
+ "task_info": {
395
+ "model": "Intel/Qwen2-0.5B-Instuct-int4-inc",
396
+ "local": true,
397
+ "revision": "main",
398
+ "private": false,
399
+ "params": 0.5,
400
+ "architectures": "QwenForCausalLM",
401
+ "quant_type": "AutoRound",
402
+ "precision": "4bit",
403
+ "model_params": 0.5,
404
+ "model_size": 0.719,
405
+ "weight_dtype": "int4",
406
+ "compute_dtype": "float16",
407
+ "gguf_ftype": "*Q4_0.gguf",
408
+ "hardware": "gpu",
409
+ "status": "Finished",
410
+ "submitted_time": "2024-04-23T15:44:22Z",
411
+ "model_type": "quantization",
412
+ "job_id": -1,
413
+ "job_start_time": null,
414
+ "scripts": "ITREX"
415
+ },
416
+ "quantization_config": {
417
+ "autoround_version": "0.2.1.dev",
418
+ "bits": 4,
419
+ "damp_percent": 0.01,
420
+ "desc_act": false,
421
+ "enable_minmax_tuning": true,
422
+ "enable_quanted_input": true,
423
+ "group_size": 32,
424
+ "is_marlin_format": false,
425
+ "iters": 1000,
426
+ "lr": 0.001,
427
+ "minmax_lr": 0.002,
428
+ "model_file_base_name": "model",
429
+ "model_name_or_path": null,
430
+ "quant_method": "gptq",
431
+ "scale_dtype": "float16",
432
+ "static_groups": false,
433
+ "sym": true,
434
+ "true_sequential": false
435
+ },
436
+ "versions": {
437
+ "harness|hellaswag|0": 1.0,
438
+ "harness|arc:easy|0": 1.0,
439
+ "harness|truthfulqa:mc1|0": 2.0,
440
+ "harness|mmlu|0": null,
441
+ "harness|mmlu_humanities|0": null,
442
+ "harness|mmlu_formal_logic|0": 0.0,
443
+ "harness|mmlu_high_school_european_history|0": 0.0,
444
+ "harness|mmlu_high_school_us_history|0": 0.0,
445
+ "harness|mmlu_high_school_world_history|0": 0.0,
446
+ "harness|mmlu_international_law|0": 0.0,
447
+ "harness|mmlu_jurisprudence|0": 0.0,
448
+ "harness|mmlu_logical_fallacies|0": 0.0,
449
+ "harness|mmlu_moral_disputes|0": 0.0,
450
+ "harness|mmlu_moral_scenarios|0": 0.0,
451
+ "harness|mmlu_philosophy|0": 0.0,
452
+ "harness|mmlu_prehistory|0": 0.0,
453
+ "harness|mmlu_professional_law|0": 0.0,
454
+ "harness|mmlu_world_religions|0": 0.0,
455
+ "harness|mmlu_other|0": null,
456
+ "harness|mmlu_business_ethics|0": 0.0,
457
+ "harness|mmlu_clinical_knowledge|0": 0.0,
458
+ "harness|mmlu_college_medicine|0": 0.0,
459
+ "harness|mmlu_global_facts|0": 0.0,
460
+ "harness|mmlu_human_aging|0": 0.0,
461
+ "harness|mmlu_management|0": 0.0,
462
+ "harness|mmlu_marketing|0": 0.0,
463
+ "harness|mmlu_medical_genetics|0": 0.0,
464
+ "harness|mmlu_miscellaneous|0": 0.0,
465
+ "harness|mmlu_nutrition|0": 0.0,
466
+ "harness|mmlu_professional_accounting|0": 0.0,
467
+ "harness|mmlu_professional_medicine|0": 0.0,
468
+ "harness|mmlu_virology|0": 0.0,
469
+ "harness|mmlu_social_sciences|0": null,
470
+ "harness|mmlu_econometrics|0": 0.0,
471
+ "harness|mmlu_high_school_geography|0": 0.0,
472
+ "harness|mmlu_high_school_government_and_politics|0": 0.0,
473
+ "harness|mmlu_high_school_macroeconomics|0": 0.0,
474
+ "harness|mmlu_high_school_microeconomics|0": 0.0,
475
+ "harness|mmlu_high_school_psychology|0": 0.0,
476
+ "harness|mmlu_human_sexuality|0": 0.0,
477
+ "harness|mmlu_professional_psychology|0": 0.0,
478
+ "harness|mmlu_public_relations|0": 0.0,
479
+ "harness|mmlu_security_studies|0": 0.0,
480
+ "harness|mmlu_sociology|0": 0.0,
481
+ "harness|mmlu_us_foreign_policy|0": 0.0,
482
+ "harness|mmlu_stem|0": null,
483
+ "harness|mmlu_abstract_algebra|0": 0.0,
484
+ "harness|mmlu_anatomy|0": 0.0,
485
+ "harness|mmlu_astronomy|0": 0.0,
486
+ "harness|mmlu_college_biology|0": 0.0,
487
+ "harness|mmlu_college_chemistry|0": 0.0,
488
+ "harness|mmlu_college_computer_science|0": 0.0,
489
+ "harness|mmlu_college_mathematics|0": 0.0,
490
+ "harness|mmlu_college_physics|0": 0.0,
491
+ "harness|mmlu_computer_security|0": 0.0,
492
+ "harness|mmlu_conceptual_physics|0": 0.0,
493
+ "harness|mmlu_electrical_engineering|0": 0.0,
494
+ "harness|mmlu_elementary_mathematics|0": 0.0,
495
+ "harness|mmlu_high_school_biology|0": 0.0,
496
+ "harness|mmlu_high_school_chemistry|0": 0.0,
497
+ "harness|mmlu_high_school_computer_science|0": 0.0,
498
+ "harness|mmlu_high_school_mathematics|0": 0.0,
499
+ "harness|mmlu_high_school_physics|0": 0.0,
500
+ "harness|mmlu_high_school_statistics|0": 0.0,
501
+ "harness|mmlu_machine_learning|0": 0.0,
502
+ "harness|piqa|0": 1.0,
503
+ "harness|boolq|0": 2.0,
504
+ "harness|openbookqa|0": 1.0,
505
+ "harness|arc:challenge|0": 1.0,
506
+ "harness|truthfulqa:mc2|0": 2.0,
507
+ "harness|winogrande|0": 1.0,
508
+ "harness|lambada:openai|0": 1.0
509
+ },
510
+ "n-shot": {
511
+ "arc_challenge": 0,
512
+ "arc_easy": 0,
513
+ "boolq": 0,
514
+ "hellaswag": 0,
515
+ "lambada_openai": 0,
516
+ "mmlu": 0,
517
+ "mmlu_abstract_algebra": 0,
518
+ "mmlu_anatomy": 0,
519
+ "mmlu_astronomy": 0,
520
+ "mmlu_business_ethics": 0,
521
+ "mmlu_clinical_knowledge": 0,
522
+ "mmlu_college_biology": 0,
523
+ "mmlu_college_chemistry": 0,
524
+ "mmlu_college_computer_science": 0,
525
+ "mmlu_college_mathematics": 0,
526
+ "mmlu_college_medicine": 0,
527
+ "mmlu_college_physics": 0,
528
+ "mmlu_computer_security": 0,
529
+ "mmlu_conceptual_physics": 0,
530
+ "mmlu_econometrics": 0,
531
+ "mmlu_electrical_engineering": 0,
532
+ "mmlu_elementary_mathematics": 0,
533
+ "mmlu_formal_logic": 0,
534
+ "mmlu_global_facts": 0,
535
+ "mmlu_high_school_biology": 0,
536
+ "mmlu_high_school_chemistry": 0,
537
+ "mmlu_high_school_computer_science": 0,
538
+ "mmlu_high_school_european_history": 0,
539
+ "mmlu_high_school_geography": 0,
540
+ "mmlu_high_school_government_and_politics": 0,
541
+ "mmlu_high_school_macroeconomics": 0,
542
+ "mmlu_high_school_mathematics": 0,
543
+ "mmlu_high_school_microeconomics": 0,
544
+ "mmlu_high_school_physics": 0,
545
+ "mmlu_high_school_psychology": 0,
546
+ "mmlu_high_school_statistics": 0,
547
+ "mmlu_high_school_us_history": 0,
548
+ "mmlu_high_school_world_history": 0,
549
+ "mmlu_human_aging": 0,
550
+ "mmlu_human_sexuality": 0,
551
+ "mmlu_humanities": 0,
552
+ "mmlu_international_law": 0,
553
+ "mmlu_jurisprudence": 0,
554
+ "mmlu_logical_fallacies": 0,
555
+ "mmlu_machine_learning": 0,
556
+ "mmlu_management": 0,
557
+ "mmlu_marketing": 0,
558
+ "mmlu_medical_genetics": 0,
559
+ "mmlu_miscellaneous": 0,
560
+ "mmlu_moral_disputes": 0,
561
+ "mmlu_moral_scenarios": 0,
562
+ "mmlu_nutrition": 0,
563
+ "mmlu_other": 0,
564
+ "mmlu_philosophy": 0,
565
+ "mmlu_prehistory": 0,
566
+ "mmlu_professional_accounting": 0,
567
+ "mmlu_professional_law": 0,
568
+ "mmlu_professional_medicine": 0,
569
+ "mmlu_professional_psychology": 0,
570
+ "mmlu_public_relations": 0,
571
+ "mmlu_security_studies": 0,
572
+ "mmlu_social_sciences": 0,
573
+ "mmlu_sociology": 0,
574
+ "mmlu_stem": 0,
575
+ "mmlu_us_foreign_policy": 0,
576
+ "mmlu_virology": 0,
577
+ "mmlu_world_religions": 0,
578
+ "openbookqa": 0,
579
+ "piqa": 0,
580
+ "truthfulqa_mc1": 0,
581
+ "truthfulqa_mc2": 0,
582
+ "winogrande": 0
583
+ },
584
+ "date": 1717638292.2679868,
585
+ "config": {
586
+ "model": "hf",
587
+ "model_args": "pretrained=Intel/Qwen2-0.5B-Instuct-int4-inc,trust_remote_code=True,dtype=float16,_commit_hash=main",
588
+ "batch_size": 4,
589
+ "batch_sizes": [],
590
+ "device": "cuda",
591
+ "use_cache": null,
592
+ "limit": null,
593
+ "bootstrap_iters": 100000,
594
+ "gen_kwargs": null
595
+ }
596
+ }