{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.41552901023890787, "acc_stderr": 0.014401366641216386, "acc_norm": 0.492320819112628, "acc_norm_stderr": 0.014609667440892567 }, "harness|ko_hellaswag|10": { "acc": 0.44214299940250945, "acc_stderr": 0.004956262919324406, "acc_norm": 0.6040629356701852, "acc_norm_stderr": 0.0048805154313231605 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.5497076023391813, "acc_stderr": 0.038158273659132366, "acc_norm": 0.5497076023391813, "acc_norm_stderr": 0.038158273659132366 }, "harness|ko_mmlu_management|5": { "acc": 0.5436893203883495, "acc_stderr": 0.049318019942204146, "acc_norm": 0.5436893203883495, "acc_norm_stderr": 0.049318019942204146 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.5683269476372924, "acc_stderr": 0.017712228939299798, "acc_norm": 0.5683269476372924, "acc_norm_stderr": 0.017712228939299798 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4666666666666667, "acc_stderr": 0.043097329010363554, "acc_norm": 0.4666666666666667, "acc_norm_stderr": 0.043097329010363554 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.29, "acc_stderr": 0.045604802157206824, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206824 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.37872340425531914, "acc_stderr": 0.03170995606040655, "acc_norm": 0.37872340425531914, "acc_norm_stderr": 0.03170995606040655 }, "harness|ko_mmlu_virology|5": { "acc": 0.39156626506024095, "acc_stderr": 0.03799857454479637, "acc_norm": 0.39156626506024095, "acc_norm_stderr": 0.03799857454479637 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5080385852090032, "acc_stderr": 0.028394421370984538, "acc_norm": 0.5080385852090032, "acc_norm_stderr": 0.028394421370984538 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.4977578475336323, "acc_stderr": 0.03355746535223263, "acc_norm": 0.4977578475336323, "acc_norm_stderr": 0.03355746535223263 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.48091603053435117, "acc_stderr": 0.04382094705550988, "acc_norm": 0.48091603053435117, "acc_norm_stderr": 0.04382094705550988 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.44, "acc_stderr": 0.0498887651569859, "acc_norm": 0.44, "acc_norm_stderr": 0.0498887651569859 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5757575757575758, "acc_stderr": 0.03521224908841586, "acc_norm": 0.5757575757575758, "acc_norm_stderr": 0.03521224908841586 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.4413793103448276, "acc_stderr": 0.04137931034482758, "acc_norm": 0.4413793103448276, "acc_norm_stderr": 0.04137931034482758 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.21568627450980393, "acc_stderr": 0.040925639582376556, "acc_norm": 0.21568627450980393, "acc_norm_stderr": 0.040925639582376556 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.48739495798319327, "acc_stderr": 0.032468167657521745, "acc_norm": 0.48739495798319327, "acc_norm_stderr": 0.032468167657521745 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.4307692307692308, "acc_stderr": 0.02510682066053975, "acc_norm": 0.4307692307692308, "acc_norm_stderr": 0.02510682066053975 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.51, "acc_stderr": 0.05024183937956912, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956912 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.39, "acc_stderr": 0.04902071300001975, "acc_norm": 0.39, "acc_norm_stderr": 0.04902071300001975 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5, "acc_stderr": 0.04833682445228318, "acc_norm": 0.5, "acc_norm_stderr": 0.04833682445228318 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.4088669950738916, "acc_stderr": 0.03459058815883233, "acc_norm": 0.4088669950738916, "acc_norm_stderr": 0.03459058815883233 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.43870967741935485, "acc_stderr": 0.02822949732031722, "acc_norm": 0.43870967741935485, "acc_norm_stderr": 0.02822949732031722 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6410256410256411, "acc_stderr": 0.031426169937919246, "acc_norm": 0.6410256410256411, "acc_norm_stderr": 0.031426169937919246 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.4339622641509434, "acc_stderr": 0.030503292013342596, "acc_norm": 0.4339622641509434, "acc_norm_stderr": 0.030503292013342596 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5181818181818182, "acc_stderr": 0.04785964010794916, "acc_norm": 0.5181818181818182, "acc_norm_stderr": 0.04785964010794916 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.25555555555555554, "acc_stderr": 0.026593939101844065, "acc_norm": 0.25555555555555554, "acc_norm_stderr": 0.026593939101844065 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.304635761589404, "acc_stderr": 0.03757949922943342, "acc_norm": 0.304635761589404, "acc_norm_stderr": 0.03757949922943342 }, "harness|ko_mmlu_sociology|5": { "acc": 0.5920398009950248, "acc_stderr": 0.03475116365194092, "acc_norm": 0.5920398009950248, "acc_norm_stderr": 0.03475116365194092 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.43352601156069365, "acc_stderr": 0.03778621079092055, "acc_norm": 0.43352601156069365, "acc_norm_stderr": 0.03778621079092055 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.28835978835978837, "acc_stderr": 0.0233306540545359, "acc_norm": 0.28835978835978837, "acc_norm_stderr": 0.0233306540545359 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3472222222222222, "acc_stderr": 0.039812405437178615, "acc_norm": 0.3472222222222222, "acc_norm_stderr": 0.039812405437178615 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.27, "acc_stderr": 0.044619604333847394, "acc_norm": 0.27, "acc_norm_stderr": 0.044619604333847394 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.63, "acc_stderr": 0.048523658709391, "acc_norm": 0.63, "acc_norm_stderr": 0.048523658709391 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5317919075144508, "acc_stderr": 0.02686462436675665, "acc_norm": 0.5317919075144508, "acc_norm_stderr": 0.02686462436675665 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.5644171779141104, "acc_stderr": 0.03895632464138937, "acc_norm": 0.5644171779141104, "acc_norm_stderr": 0.03895632464138937 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.5216049382716049, "acc_stderr": 0.02779476010500873, "acc_norm": 0.5216049382716049, "acc_norm_stderr": 0.02779476010500873 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.33, "acc_stderr": 0.047258156262526045, "acc_norm": 0.33, "acc_norm_stderr": 0.047258156262526045 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.6269430051813472, "acc_stderr": 0.03490205592048573, "acc_norm": 0.6269430051813472, "acc_norm_stderr": 0.03490205592048573 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2543859649122807, "acc_stderr": 0.0409698513984367, "acc_norm": 0.2543859649122807, "acc_norm_stderr": 0.0409698513984367 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.5688073394495413, "acc_stderr": 0.021233365030319567, "acc_norm": 0.5688073394495413, "acc_norm_stderr": 0.021233365030319567 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.30158730158730157, "acc_stderr": 0.04104947269903394, "acc_norm": 0.30158730158730157, "acc_norm_stderr": 0.04104947269903394 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.42810457516339867, "acc_stderr": 0.028332397483664274, "acc_norm": 0.42810457516339867, "acc_norm_stderr": 0.028332397483664274 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.42, "acc_stderr": 0.049604496374885836, "acc_norm": 0.42, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6776859504132231, "acc_stderr": 0.042664163633521685, "acc_norm": 0.6776859504132231, "acc_norm_stderr": 0.042664163633521685 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.4144736842105263, "acc_stderr": 0.04008973785779206, "acc_norm": 0.4144736842105263, "acc_norm_stderr": 0.04008973785779206 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.4264705882352941, "acc_stderr": 0.02000791273935935, "acc_norm": 0.4264705882352941, "acc_norm_stderr": 0.02000791273935935 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3475177304964539, "acc_stderr": 0.02840662780959095, "acc_norm": 0.3475177304964539, "acc_norm_stderr": 0.02840662780959095 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.33035714285714285, "acc_stderr": 0.044642857142857116, "acc_norm": 0.33035714285714285, "acc_norm_stderr": 0.044642857142857116 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.33796296296296297, "acc_stderr": 0.03225941352631295, "acc_norm": 0.33796296296296297, "acc_norm_stderr": 0.03225941352631295 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.264804469273743, "acc_stderr": 0.014756906483260659, "acc_norm": 0.264804469273743, "acc_norm_stderr": 0.014756906483260659 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.4, "acc_stderr": 0.049236596391733084, "acc_norm": 0.4, "acc_norm_stderr": 0.049236596391733084 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.47, "acc_stderr": 0.050161355804659205, "acc_norm": 0.47, "acc_norm_stderr": 0.050161355804659205 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.35661764705882354, "acc_stderr": 0.02909720956841197, "acc_norm": 0.35661764705882354, "acc_norm_stderr": 0.02909720956841197 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.4857142857142857, "acc_stderr": 0.03199615232806287, "acc_norm": 0.4857142857142857, "acc_norm_stderr": 0.03199615232806287 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.6666666666666666, "acc_stderr": 0.03068582059661081, "acc_norm": 0.6666666666666666, "acc_norm_stderr": 0.03068582059661081 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3520208604954368, "acc_stderr": 0.01219814060535359, "acc_norm": 0.3520208604954368, "acc_norm_stderr": 0.01219814060535359 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.5343137254901961, "acc_stderr": 0.03501038327635897, "acc_norm": 0.5343137254901961, "acc_norm_stderr": 0.03501038327635897 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.6121212121212121, "acc_stderr": 0.0380491365397101, "acc_norm": 0.6121212121212121, "acc_norm_stderr": 0.0380491365397101 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.34761321909424725, "mc1_stderr": 0.016670769188897306, "mc2": 0.5104539931249092, "mc2_stderr": 0.01608799028808744 }, "harness|ko_commongen_v2|2": { "acc": 0.46989374262101535, "acc_stderr": 0.017159163590170216, "acc_norm": 0.5159386068476978, "acc_norm_stderr": 0.017181617837190192 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "LDCC/LDCC-Instruct-Llama-2-ko-16B", "model_sha": "f7a3f41bb36b1e9b9d894512aa266fd30d4b5298", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }