{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.33276450511945393, "acc_stderr": 0.0137698630461923, "acc_norm": 0.39334470989761094, "acc_norm_stderr": 0.014275101465693028 }, "harness|ko_hellaswag|10": { "acc": 0.36825333598884685, "acc_stderr": 0.004813448615404438, "acc_norm": 0.4763991236805417, "acc_norm_stderr": 0.004984219681732663 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.42105263157894735, "acc_stderr": 0.03786720706234215, "acc_norm": 0.42105263157894735, "acc_norm_stderr": 0.03786720706234215 }, "harness|ko_mmlu_management|5": { "acc": 0.42718446601941745, "acc_stderr": 0.04897957737781168, "acc_norm": 0.42718446601941745, "acc_norm_stderr": 0.04897957737781168 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.4929757343550447, "acc_stderr": 0.017878199003432214, "acc_norm": 0.4929757343550447, "acc_norm_stderr": 0.017878199003432214 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.3925925925925926, "acc_stderr": 0.04218506215368879, "acc_norm": 0.3925925925925926, "acc_norm_stderr": 0.04218506215368879 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.32, "acc_stderr": 0.04688261722621504, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621504 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.3659574468085106, "acc_stderr": 0.0314895582974553, "acc_norm": 0.3659574468085106, "acc_norm_stderr": 0.0314895582974553 }, "harness|ko_mmlu_virology|5": { "acc": 0.35542168674698793, "acc_stderr": 0.03726214354322415, "acc_norm": 0.35542168674698793, "acc_norm_stderr": 0.03726214354322415 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.4662379421221865, "acc_stderr": 0.02833327710956279, "acc_norm": 0.4662379421221865, "acc_norm_stderr": 0.02833327710956279 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.4260089686098655, "acc_stderr": 0.0331883328621728, "acc_norm": 0.4260089686098655, "acc_norm_stderr": 0.0331883328621728 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.45038167938931295, "acc_stderr": 0.04363643698524779, "acc_norm": 0.45038167938931295, "acc_norm_stderr": 0.04363643698524779 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.32, "acc_stderr": 0.04688261722621504, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621504 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.45454545454545453, "acc_stderr": 0.03547601494006938, "acc_norm": 0.45454545454545453, "acc_norm_stderr": 0.03547601494006938 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.4206896551724138, "acc_stderr": 0.0411391498118926, "acc_norm": 0.4206896551724138, "acc_norm_stderr": 0.0411391498118926 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.20588235294117646, "acc_stderr": 0.04023382273617747, "acc_norm": 0.20588235294117646, "acc_norm_stderr": 0.04023382273617747 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.35714285714285715, "acc_stderr": 0.031124619309328177, "acc_norm": 0.35714285714285715, "acc_norm_stderr": 0.031124619309328177 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.358974358974359, "acc_stderr": 0.02432173848460237, "acc_norm": 0.358974358974359, "acc_norm_stderr": 0.02432173848460237 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.46, "acc_stderr": 0.05009082659620332, "acc_norm": 0.46, "acc_norm_stderr": 0.05009082659620332 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5277777777777778, "acc_stderr": 0.048262172941398944, "acc_norm": 0.5277777777777778, "acc_norm_stderr": 0.048262172941398944 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.4088669950738916, "acc_stderr": 0.034590588158832314, "acc_norm": 0.4088669950738916, "acc_norm_stderr": 0.034590588158832314 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.43870967741935485, "acc_stderr": 0.02822949732031722, "acc_norm": 0.43870967741935485, "acc_norm_stderr": 0.02822949732031722 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6153846153846154, "acc_stderr": 0.03187195347942466, "acc_norm": 0.6153846153846154, "acc_norm_stderr": 0.03187195347942466 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.4037735849056604, "acc_stderr": 0.03019761160019795, "acc_norm": 0.4037735849056604, "acc_norm_stderr": 0.03019761160019795 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.42727272727272725, "acc_stderr": 0.04738198703545483, "acc_norm": 0.42727272727272725, "acc_norm_stderr": 0.04738198703545483 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.24814814814814815, "acc_stderr": 0.0263357394040558, "acc_norm": 0.24814814814814815, "acc_norm_stderr": 0.0263357394040558 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.23178807947019867, "acc_stderr": 0.03445406271987054, "acc_norm": 0.23178807947019867, "acc_norm_stderr": 0.03445406271987054 }, "harness|ko_mmlu_sociology|5": { "acc": 0.5223880597014925, "acc_stderr": 0.03531987930208731, "acc_norm": 0.5223880597014925, "acc_norm_stderr": 0.03531987930208731 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.3179190751445087, "acc_stderr": 0.03550683989165582, "acc_norm": 0.3179190751445087, "acc_norm_stderr": 0.03550683989165582 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.26455026455026454, "acc_stderr": 0.022717467897708617, "acc_norm": 0.26455026455026454, "acc_norm_stderr": 0.022717467897708617 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3263888888888889, "acc_stderr": 0.03921067198982266, "acc_norm": 0.3263888888888889, "acc_norm_stderr": 0.03921067198982266 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.24, "acc_stderr": 0.042923469599092816, "acc_norm": 0.24, "acc_norm_stderr": 0.042923469599092816 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.53, "acc_stderr": 0.050161355804659205, "acc_norm": 0.53, "acc_norm_stderr": 0.050161355804659205 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.43641618497109824, "acc_stderr": 0.026700545424943684, "acc_norm": 0.43641618497109824, "acc_norm_stderr": 0.026700545424943684 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.3987730061349693, "acc_stderr": 0.03847021420456024, "acc_norm": 0.3987730061349693, "acc_norm_stderr": 0.03847021420456024 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.42592592592592593, "acc_stderr": 0.027513747284379424, "acc_norm": 0.42592592592592593, "acc_norm_stderr": 0.027513747284379424 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.23, "acc_stderr": 0.04229525846816505, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816505 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.45595854922279794, "acc_stderr": 0.035944137112724366, "acc_norm": 0.45595854922279794, "acc_norm_stderr": 0.035944137112724366 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2982456140350877, "acc_stderr": 0.043036840335373173, "acc_norm": 0.2982456140350877, "acc_norm_stderr": 0.043036840335373173 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.42201834862385323, "acc_stderr": 0.02117499140776317, "acc_norm": 0.42201834862385323, "acc_norm_stderr": 0.02117499140776317 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.24603174603174602, "acc_stderr": 0.03852273364924316, "acc_norm": 0.24603174603174602, "acc_norm_stderr": 0.03852273364924316 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.39869281045751637, "acc_stderr": 0.028036092273891765, "acc_norm": 0.39869281045751637, "acc_norm_stderr": 0.028036092273891765 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.36, "acc_stderr": 0.04824181513244218, "acc_norm": 0.36, "acc_norm_stderr": 0.04824181513244218 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6115702479338843, "acc_stderr": 0.04449270350068383, "acc_norm": 0.6115702479338843, "acc_norm_stderr": 0.04449270350068383 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.4605263157894737, "acc_stderr": 0.04056242252249033, "acc_norm": 0.4605263157894737, "acc_norm_stderr": 0.04056242252249033 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.33169934640522875, "acc_stderr": 0.01904748523936038, "acc_norm": 0.33169934640522875, "acc_norm_stderr": 0.01904748523936038 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.31560283687943264, "acc_stderr": 0.027724989449509317, "acc_norm": 0.31560283687943264, "acc_norm_stderr": 0.027724989449509317 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.25, "acc_stderr": 0.04109974682633932, "acc_norm": 0.25, "acc_norm_stderr": 0.04109974682633932 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.2638888888888889, "acc_stderr": 0.030058202704309846, "acc_norm": 0.2638888888888889, "acc_norm_stderr": 0.030058202704309846 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2424581005586592, "acc_stderr": 0.01433352205921789, "acc_norm": 0.2424581005586592, "acc_norm_stderr": 0.01433352205921789 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.26, "acc_stderr": 0.0440844002276808, "acc_norm": 0.26, "acc_norm_stderr": 0.0440844002276808 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.2757352941176471, "acc_stderr": 0.027146271936625162, "acc_norm": 0.2757352941176471, "acc_norm_stderr": 0.027146271936625162 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.3183673469387755, "acc_stderr": 0.029822533793982052, "acc_norm": 0.3183673469387755, "acc_norm_stderr": 0.029822533793982052 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.45147679324894513, "acc_stderr": 0.032393600173974704, "acc_norm": 0.45147679324894513, "acc_norm_stderr": 0.032393600173974704 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.2926988265971317, "acc_stderr": 0.01162094919584953, "acc_norm": 0.2926988265971317, "acc_norm_stderr": 0.01162094919584953 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.35784313725490197, "acc_stderr": 0.033644872860882996, "acc_norm": 0.35784313725490197, "acc_norm_stderr": 0.033644872860882996 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.4303030303030303, "acc_stderr": 0.03866225962879076, "acc_norm": 0.4303030303030303, "acc_norm_stderr": 0.03866225962879076 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.3084455324357405, "mc1_stderr": 0.01616803938315687, "mc2": 0.4796326606446058, "mc2_stderr": 0.015522483900178127 }, "harness|ko_commongen_v2|2": { "acc": 0.46162927981109797, "acc_stderr": 0.017139660221845564, "acc_norm": 0.5608028335301063, "acc_norm_stderr": 0.017062775744780705 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "AIdenU/LLAMA-2-13b-ko-Y24-DPO_v2.0", "model_sha": "64a7cfb337e43c31e60f8a3710e55db2b76bc09a", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }