results / 4yo1 /llama3-eng-ko-8b-sl6_1 /result_2024-07-16 04:26:59.json
choco9966
add backup results
70a679f
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.19965870307167236,
"acc_stderr": 0.01168162575688869,
"acc_norm": 0.26109215017064846,
"acc_norm_stderr": 0.012835523909473843
},
"harness|ko_hellaswag|10": {
"acc": 0.262796255725951,
"acc_stderr": 0.004392531344297395,
"acc_norm": 0.28809002190798644,
"acc_norm_stderr": 0.004519476835646786
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.42105263157894735,
"acc_stderr": 0.037867207062342145,
"acc_norm": 0.42105263157894735,
"acc_norm_stderr": 0.037867207062342145
},
"harness|ko_mmlu_management|5": {
"acc": 0.2815533980582524,
"acc_stderr": 0.04453254836326466,
"acc_norm": 0.2815533980582524,
"acc_norm_stderr": 0.04453254836326466
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.35759897828863346,
"acc_stderr": 0.017139488998803302,
"acc_norm": 0.35759897828863346,
"acc_norm_stderr": 0.017139488998803302
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.32592592592592595,
"acc_stderr": 0.04049122041702506,
"acc_norm": 0.32592592592592595,
"acc_norm_stderr": 0.04049122041702506
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.24,
"acc_stderr": 0.04292346959909284,
"acc_norm": 0.24,
"acc_norm_stderr": 0.04292346959909284
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.3574468085106383,
"acc_stderr": 0.03132941789476425,
"acc_norm": 0.3574468085106383,
"acc_norm_stderr": 0.03132941789476425
},
"harness|ko_mmlu_virology|5": {
"acc": 0.35542168674698793,
"acc_stderr": 0.03726214354322415,
"acc_norm": 0.35542168674698793,
"acc_norm_stderr": 0.03726214354322415
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.33440514469453375,
"acc_stderr": 0.026795422327893937,
"acc_norm": 0.33440514469453375,
"acc_norm_stderr": 0.026795422327893937
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.27802690582959644,
"acc_stderr": 0.030069584874494047,
"acc_norm": 0.27802690582959644,
"acc_norm_stderr": 0.030069584874494047
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.3435114503816794,
"acc_stderr": 0.041649760719448786,
"acc_norm": 0.3435114503816794,
"acc_norm_stderr": 0.041649760719448786
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.3838383838383838,
"acc_stderr": 0.034648816750163375,
"acc_norm": 0.3838383838383838,
"acc_norm_stderr": 0.034648816750163375
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.43448275862068964,
"acc_stderr": 0.041307408795554966,
"acc_norm": 0.43448275862068964,
"acc_norm_stderr": 0.041307408795554966
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.2647058823529412,
"acc_stderr": 0.04389869956808777,
"acc_norm": 0.2647058823529412,
"acc_norm_stderr": 0.04389869956808777
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.40756302521008403,
"acc_stderr": 0.03191863374478466,
"acc_norm": 0.40756302521008403,
"acc_norm_stderr": 0.03191863374478466
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.02390115797940254,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.02390115797940254
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.47,
"acc_stderr": 0.050161355804659205,
"acc_norm": 0.47,
"acc_norm_stderr": 0.050161355804659205
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.37962962962962965,
"acc_stderr": 0.04691521224077742,
"acc_norm": 0.37962962962962965,
"acc_norm_stderr": 0.04691521224077742
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.35960591133004927,
"acc_stderr": 0.03376458246509567,
"acc_norm": 0.35960591133004927,
"acc_norm_stderr": 0.03376458246509567
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.36129032258064514,
"acc_stderr": 0.027327548447957532,
"acc_norm": 0.36129032258064514,
"acc_norm_stderr": 0.027327548447957532
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.36752136752136755,
"acc_stderr": 0.03158539157745636,
"acc_norm": 0.36752136752136755,
"acc_norm_stderr": 0.03158539157745636
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.3283018867924528,
"acc_stderr": 0.028901593612411784,
"acc_norm": 0.3283018867924528,
"acc_norm_stderr": 0.028901593612411784
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.32727272727272727,
"acc_stderr": 0.04494290866252088,
"acc_norm": 0.32727272727272727,
"acc_norm_stderr": 0.04494290866252088
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.3148148148148148,
"acc_stderr": 0.028317533496066465,
"acc_norm": 0.3148148148148148,
"acc_norm_stderr": 0.028317533496066465
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.2781456953642384,
"acc_stderr": 0.03658603262763743,
"acc_norm": 0.2781456953642384,
"acc_norm_stderr": 0.03658603262763743
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.39800995024875624,
"acc_stderr": 0.03461199429040014,
"acc_norm": 0.39800995024875624,
"acc_norm_stderr": 0.03461199429040014
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.31213872832369943,
"acc_stderr": 0.035331333893236574,
"acc_norm": 0.31213872832369943,
"acc_norm_stderr": 0.035331333893236574
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.328042328042328,
"acc_stderr": 0.0241804971643769,
"acc_norm": 0.328042328042328,
"acc_norm_stderr": 0.0241804971643769
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.2222222222222222,
"acc_stderr": 0.03476590104304134,
"acc_norm": 0.2222222222222222,
"acc_norm_stderr": 0.03476590104304134
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.44,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.44,
"acc_norm_stderr": 0.04988876515698589
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.32947976878612717,
"acc_stderr": 0.025305258131879706,
"acc_norm": 0.32947976878612717,
"acc_norm_stderr": 0.025305258131879706
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.3006134969325153,
"acc_stderr": 0.03602511318806771,
"acc_norm": 0.3006134969325153,
"acc_norm_stderr": 0.03602511318806771
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.2962962962962963,
"acc_stderr": 0.02540719779889017,
"acc_norm": 0.2962962962962963,
"acc_norm_stderr": 0.02540719779889017
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252605,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252605
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.29015544041450775,
"acc_stderr": 0.03275264467791516,
"acc_norm": 0.29015544041450775,
"acc_norm_stderr": 0.03275264467791516
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.2631578947368421,
"acc_stderr": 0.0414243971948936,
"acc_norm": 0.2631578947368421,
"acc_norm_stderr": 0.0414243971948936
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.3339449541284404,
"acc_stderr": 0.020220554196736403,
"acc_norm": 0.3339449541284404,
"acc_norm_stderr": 0.020220554196736403
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.2857142857142857,
"acc_stderr": 0.0404061017820884,
"acc_norm": 0.2857142857142857,
"acc_norm_stderr": 0.0404061017820884
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.38235294117647056,
"acc_stderr": 0.027826109307283686,
"acc_norm": 0.38235294117647056,
"acc_norm_stderr": 0.027826109307283686
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.32,
"acc_stderr": 0.046882617226215034,
"acc_norm": 0.32,
"acc_norm_stderr": 0.046882617226215034
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.5041322314049587,
"acc_stderr": 0.045641987674327526,
"acc_norm": 0.5041322314049587,
"acc_norm_stderr": 0.045641987674327526
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.42105263157894735,
"acc_stderr": 0.04017901275981748,
"acc_norm": 0.42105263157894735,
"acc_norm_stderr": 0.04017901275981748
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.29411764705882354,
"acc_stderr": 0.018433427649401896,
"acc_norm": 0.29411764705882354,
"acc_norm_stderr": 0.018433427649401896
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.3262411347517731,
"acc_stderr": 0.027968453043563168,
"acc_norm": 0.3262411347517731,
"acc_norm_stderr": 0.027968453043563168
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.3125,
"acc_stderr": 0.043994650575715215,
"acc_norm": 0.3125,
"acc_norm_stderr": 0.043994650575715215
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.39351851851851855,
"acc_stderr": 0.03331747876370312,
"acc_norm": 0.39351851851851855,
"acc_norm_stderr": 0.03331747876370312
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2636871508379888,
"acc_stderr": 0.014736926383761983,
"acc_norm": 0.2636871508379888,
"acc_norm_stderr": 0.014736926383761983
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.36,
"acc_stderr": 0.04824181513244218,
"acc_norm": 0.36,
"acc_norm_stderr": 0.04824181513244218
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.46,
"acc_stderr": 0.05009082659620332,
"acc_norm": 0.46,
"acc_norm_stderr": 0.05009082659620332
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.3713235294117647,
"acc_stderr": 0.02934980313976587,
"acc_norm": 0.3713235294117647,
"acc_norm_stderr": 0.02934980313976587
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.4448979591836735,
"acc_stderr": 0.031814251181977865,
"acc_norm": 0.4448979591836735,
"acc_norm_stderr": 0.031814251181977865
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.33755274261603374,
"acc_stderr": 0.03078154910202622,
"acc_norm": 0.33755274261603374,
"acc_norm_stderr": 0.03078154910202622
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.303129074315515,
"acc_stderr": 0.011738669951254298,
"acc_norm": 0.303129074315515,
"acc_norm_stderr": 0.011738669951254298
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.3480392156862745,
"acc_stderr": 0.03343311240488418,
"acc_norm": 0.3480392156862745,
"acc_norm_stderr": 0.03343311240488418
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.3515151515151515,
"acc_stderr": 0.037282069986826503,
"acc_norm": 0.3515151515151515,
"acc_norm_stderr": 0.037282069986826503
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.2484700122399021,
"mc1_stderr": 0.015127427096520688,
"mc2": 0.4886424856035836,
"mc2_stderr": 0.016537537410601783
},
"harness|ko_commongen_v2|2": {
"acc": 0.12160566706021252,
"acc_stderr": 0.011236640546845988,
"acc_norm": 0.2644628099173554,
"acc_norm_stderr": 0.015163499477892412
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "4yo1/llama3-eng-ko-8b-sl6_1",
"model_sha": "82835f842c382a868f4c6f5fac4baa1d80c1cfab",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}