{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.44368600682593856, "acc_stderr": 0.014518421825670447, "acc_norm": 0.4906143344709898, "acc_norm_stderr": 0.014608816322065 }, "harness|ko_hellaswag|10": { "acc": 0.4576777534355706, "acc_stderr": 0.0049718741597776965, "acc_norm": 0.6170085640310695, "acc_norm_stderr": 0.0048512275270708935 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.6900584795321637, "acc_stderr": 0.035469769593931624, "acc_norm": 0.6900584795321637, "acc_norm_stderr": 0.035469769593931624 }, "harness|ko_mmlu_management|5": { "acc": 0.6990291262135923, "acc_stderr": 0.045416094465039476, "acc_norm": 0.6990291262135923, "acc_norm_stderr": 0.045416094465039476 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.70242656449553, "acc_stderr": 0.016349111912909435, "acc_norm": 0.70242656449553, "acc_norm_stderr": 0.016349111912909435 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4444444444444444, "acc_stderr": 0.042925967182569816, "acc_norm": 0.4444444444444444, "acc_norm_stderr": 0.042925967182569816 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.36, "acc_stderr": 0.048241815132442176, "acc_norm": 0.36, "acc_norm_stderr": 0.048241815132442176 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.4553191489361702, "acc_stderr": 0.03255525359340355, "acc_norm": 0.4553191489361702, "acc_norm_stderr": 0.03255525359340355 }, "harness|ko_mmlu_virology|5": { "acc": 0.5120481927710844, "acc_stderr": 0.038913644958358175, "acc_norm": 0.5120481927710844, "acc_norm_stderr": 0.038913644958358175 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.6012861736334405, "acc_stderr": 0.0278093225857745, "acc_norm": 0.6012861736334405, "acc_norm_stderr": 0.0278093225857745 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.5964125560538116, "acc_stderr": 0.032928028193303135, "acc_norm": 0.5964125560538116, "acc_norm_stderr": 0.032928028193303135 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.5419847328244275, "acc_stderr": 0.04369802690578757, "acc_norm": 0.5419847328244275, "acc_norm_stderr": 0.04369802690578757 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.49, "acc_stderr": 0.05024183937956913, "acc_norm": 0.49, "acc_norm_stderr": 0.05024183937956913 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.696969696969697, "acc_stderr": 0.032742879140268674, "acc_norm": 0.696969696969697, "acc_norm_stderr": 0.032742879140268674 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.47586206896551725, "acc_stderr": 0.041618085035015295, "acc_norm": 0.47586206896551725, "acc_norm_stderr": 0.041618085035015295 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.28431372549019607, "acc_stderr": 0.04488482852329017, "acc_norm": 0.28431372549019607, "acc_norm_stderr": 0.04488482852329017 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.5756302521008403, "acc_stderr": 0.03210479051015776, "acc_norm": 0.5756302521008403, "acc_norm_stderr": 0.03210479051015776 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.5384615384615384, "acc_stderr": 0.025275892070240648, "acc_norm": 0.5384615384615384, "acc_norm_stderr": 0.025275892070240648 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.58, "acc_stderr": 0.04960449637488583, "acc_norm": 0.58, "acc_norm_stderr": 0.04960449637488583 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.6944444444444444, "acc_stderr": 0.044531975073749834, "acc_norm": 0.6944444444444444, "acc_norm_stderr": 0.044531975073749834 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.4088669950738916, "acc_stderr": 0.034590588158832314, "acc_norm": 0.4088669950738916, "acc_norm_stderr": 0.034590588158832314 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.6064516129032258, "acc_stderr": 0.027791878753132267, "acc_norm": 0.6064516129032258, "acc_norm_stderr": 0.027791878753132267 }, "harness|ko_mmlu_marketing|5": { "acc": 0.7692307692307693, "acc_stderr": 0.02760192138141762, "acc_norm": 0.7692307692307693, "acc_norm_stderr": 0.02760192138141762 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.5547169811320755, "acc_stderr": 0.030588052974270658, "acc_norm": 0.5547169811320755, "acc_norm_stderr": 0.030588052974270658 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5272727272727272, "acc_stderr": 0.04782001791380061, "acc_norm": 0.5272727272727272, "acc_norm_stderr": 0.04782001791380061 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.32592592592592595, "acc_stderr": 0.028578348365473072, "acc_norm": 0.32592592592592595, "acc_norm_stderr": 0.028578348365473072 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2980132450331126, "acc_stderr": 0.03734535676787198, "acc_norm": 0.2980132450331126, "acc_norm_stderr": 0.03734535676787198 }, "harness|ko_mmlu_sociology|5": { "acc": 0.6567164179104478, "acc_stderr": 0.03357379665433431, "acc_norm": 0.6567164179104478, "acc_norm_stderr": 0.03357379665433431 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.48554913294797686, "acc_stderr": 0.03810871630454764, "acc_norm": 0.48554913294797686, "acc_norm_stderr": 0.03810871630454764 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.3333333333333333, "acc_stderr": 0.024278568024307695, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.024278568024307695 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.4583333333333333, "acc_stderr": 0.04166666666666665, "acc_norm": 0.4583333333333333, "acc_norm_stderr": 0.04166666666666665 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.81, "acc_stderr": 0.03942772444036623, "acc_norm": 0.81, "acc_norm_stderr": 0.03942772444036623 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5838150289017341, "acc_stderr": 0.026538189104705477, "acc_norm": 0.5838150289017341, "acc_norm_stderr": 0.026538189104705477 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.5337423312883436, "acc_stderr": 0.039194155450484096, "acc_norm": 0.5337423312883436, "acc_norm_stderr": 0.039194155450484096 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.5895061728395061, "acc_stderr": 0.027371350925124768, "acc_norm": 0.5895061728395061, "acc_norm_stderr": 0.027371350925124768 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.27, "acc_stderr": 0.0446196043338474, "acc_norm": 0.27, "acc_norm_stderr": 0.0446196043338474 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.6994818652849741, "acc_stderr": 0.0330881859441575, "acc_norm": 0.6994818652849741, "acc_norm_stderr": 0.0330881859441575 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.32456140350877194, "acc_stderr": 0.04404556157374768, "acc_norm": 0.32456140350877194, "acc_norm_stderr": 0.04404556157374768 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.6256880733944954, "acc_stderr": 0.020748959408988327, "acc_norm": 0.6256880733944954, "acc_norm_stderr": 0.020748959408988327 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.31746031746031744, "acc_stderr": 0.04163453031302859, "acc_norm": 0.31746031746031744, "acc_norm_stderr": 0.04163453031302859 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5849673202614379, "acc_stderr": 0.028213504177824093, "acc_norm": 0.5849673202614379, "acc_norm_stderr": 0.028213504177824093 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.54, "acc_stderr": 0.05009082659620333, "acc_norm": 0.54, "acc_norm_stderr": 0.05009082659620333 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6446280991735537, "acc_stderr": 0.04369236326573981, "acc_norm": 0.6446280991735537, "acc_norm_stderr": 0.04369236326573981 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.6052631578947368, "acc_stderr": 0.039777499346220734, "acc_norm": 0.6052631578947368, "acc_norm_stderr": 0.039777499346220734 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.4852941176470588, "acc_stderr": 0.020219083895133924, "acc_norm": 0.4852941176470588, "acc_norm_stderr": 0.020219083895133924 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.36879432624113473, "acc_stderr": 0.028782227561347247, "acc_norm": 0.36879432624113473, "acc_norm_stderr": 0.028782227561347247 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.4017857142857143, "acc_stderr": 0.04653333146973646, "acc_norm": 0.4017857142857143, "acc_norm_stderr": 0.04653333146973646 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4583333333333333, "acc_stderr": 0.03398110890294636, "acc_norm": 0.4583333333333333, "acc_norm_stderr": 0.03398110890294636 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.23798882681564246, "acc_stderr": 0.014242630070574894, "acc_norm": 0.23798882681564246, "acc_norm_stderr": 0.014242630070574894 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.42, "acc_stderr": 0.049604496374885836, "acc_norm": 0.42, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.65, "acc_stderr": 0.0479372485441102, "acc_norm": 0.65, "acc_norm_stderr": 0.0479372485441102 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4889705882352941, "acc_stderr": 0.030365446477275675, "acc_norm": 0.4889705882352941, "acc_norm_stderr": 0.030365446477275675 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.6448979591836734, "acc_stderr": 0.030635655150387634, "acc_norm": 0.6448979591836734, "acc_norm_stderr": 0.030635655150387634 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.70042194092827, "acc_stderr": 0.0298180247497531, "acc_norm": 0.70042194092827, "acc_norm_stderr": 0.0298180247497531 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3650586701434159, "acc_stderr": 0.012296373743443475, "acc_norm": 0.3650586701434159, "acc_norm_stderr": 0.012296373743443475 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.6715686274509803, "acc_stderr": 0.03296245110172229, "acc_norm": 0.6715686274509803, "acc_norm_stderr": 0.03296245110172229 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.6909090909090909, "acc_stderr": 0.036085410115739666, "acc_norm": 0.6909090909090909, "acc_norm_stderr": 0.036085410115739666 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2607099143206854, "mc1_stderr": 0.015368841620766372, "mc2": 0.4167656798454404, "mc2_stderr": 0.015096503185058233 }, "harness|ko_commongen_v2|2": { "acc": 0.48760330578512395, "acc_stderr": 0.01718506973267653, "acc_norm": 0.526564344746163, "acc_norm_stderr": 0.017166075717577747 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "ENERGY-DRINK-LOVE/eeve_leaderboard_inst_v1.5", "model_sha": "78505257163f6e45ed327b68f7bddb7c9ff509fd", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }