{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.4718430034129693, "acc_stderr": 0.0145882041051022, "acc_norm": 0.5332764505119454, "acc_norm_stderr": 0.01457899585960581 }, "harness|ko_hellaswag|10": { "acc": 0.4615614419438359, "acc_stderr": 0.004975014529648634, "acc_norm": 0.6256721768571998, "acc_norm_stderr": 0.004829598101635782 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.6608187134502924, "acc_stderr": 0.03631053496488905, "acc_norm": 0.6608187134502924, "acc_norm_stderr": 0.03631053496488905 }, "harness|ko_mmlu_management|5": { "acc": 0.6796116504854369, "acc_stderr": 0.04620284082280041, "acc_norm": 0.6796116504854369, "acc_norm_stderr": 0.04620284082280041 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.7011494252873564, "acc_stderr": 0.016369256815093113, "acc_norm": 0.7011494252873564, "acc_norm_stderr": 0.016369256815093113 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4444444444444444, "acc_stderr": 0.042925967182569816, "acc_norm": 0.4444444444444444, "acc_norm_stderr": 0.042925967182569816 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.4553191489361702, "acc_stderr": 0.03255525359340355, "acc_norm": 0.4553191489361702, "acc_norm_stderr": 0.03255525359340355 }, "harness|ko_mmlu_virology|5": { "acc": 0.42168674698795183, "acc_stderr": 0.03844453181770917, "acc_norm": 0.42168674698795183, "acc_norm_stderr": 0.03844453181770917 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5627009646302251, "acc_stderr": 0.028173917761762906, "acc_norm": 0.5627009646302251, "acc_norm_stderr": 0.028173917761762906 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.5291479820627802, "acc_stderr": 0.03350073248773404, "acc_norm": 0.5291479820627802, "acc_norm_stderr": 0.03350073248773404 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.6030534351145038, "acc_stderr": 0.04291135671009224, "acc_norm": 0.6030534351145038, "acc_norm_stderr": 0.04291135671009224 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.48, "acc_stderr": 0.05021167315686779, "acc_norm": 0.48, "acc_norm_stderr": 0.05021167315686779 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.6666666666666666, "acc_stderr": 0.03358618145732524, "acc_norm": 0.6666666666666666, "acc_norm_stderr": 0.03358618145732524 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.4896551724137931, "acc_stderr": 0.041657747757287644, "acc_norm": 0.4896551724137931, "acc_norm_stderr": 0.041657747757287644 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.2549019607843137, "acc_stderr": 0.043364327079931785, "acc_norm": 0.2549019607843137, "acc_norm_stderr": 0.043364327079931785 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.49159663865546216, "acc_stderr": 0.0324739027656967, "acc_norm": 0.49159663865546216, "acc_norm_stderr": 0.0324739027656967 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.5205128205128206, "acc_stderr": 0.02532966316348994, "acc_norm": 0.5205128205128206, "acc_norm_stderr": 0.02532966316348994 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.62, "acc_stderr": 0.048783173121456316, "acc_norm": 0.62, "acc_norm_stderr": 0.048783173121456316 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.33, "acc_stderr": 0.047258156262526045, "acc_norm": 0.33, "acc_norm_stderr": 0.047258156262526045 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.6018518518518519, "acc_stderr": 0.047323326159788126, "acc_norm": 0.6018518518518519, "acc_norm_stderr": 0.047323326159788126 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3694581280788177, "acc_stderr": 0.03395970381998574, "acc_norm": 0.3694581280788177, "acc_norm_stderr": 0.03395970381998574 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.5870967741935483, "acc_stderr": 0.028009138125400387, "acc_norm": 0.5870967741935483, "acc_norm_stderr": 0.028009138125400387 }, "harness|ko_mmlu_marketing|5": { "acc": 0.7478632478632479, "acc_stderr": 0.028447965476231022, "acc_norm": 0.7478632478632479, "acc_norm_stderr": 0.028447965476231022 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.5433962264150943, "acc_stderr": 0.030656748696739438, "acc_norm": 0.5433962264150943, "acc_norm_stderr": 0.030656748696739438 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5636363636363636, "acc_stderr": 0.04750185058907296, "acc_norm": 0.5636363636363636, "acc_norm_stderr": 0.04750185058907296 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.34444444444444444, "acc_stderr": 0.02897264888484427, "acc_norm": 0.34444444444444444, "acc_norm_stderr": 0.02897264888484427 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.3443708609271523, "acc_stderr": 0.038796870240733264, "acc_norm": 0.3443708609271523, "acc_norm_stderr": 0.038796870240733264 }, "harness|ko_mmlu_sociology|5": { "acc": 0.6218905472636815, "acc_stderr": 0.03428867848778658, "acc_norm": 0.6218905472636815, "acc_norm_stderr": 0.03428867848778658 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.43352601156069365, "acc_stderr": 0.03778621079092055, "acc_norm": 0.43352601156069365, "acc_norm_stderr": 0.03778621079092055 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.35978835978835977, "acc_stderr": 0.024718075944129277, "acc_norm": 0.35978835978835977, "acc_norm_stderr": 0.024718075944129277 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.4652777777777778, "acc_stderr": 0.041711158581816184, "acc_norm": 0.4652777777777778, "acc_norm_stderr": 0.041711158581816184 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.78, "acc_stderr": 0.04163331998932262, "acc_norm": 0.78, "acc_norm_stderr": 0.04163331998932262 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.476878612716763, "acc_stderr": 0.026890297881303125, "acc_norm": 0.476878612716763, "acc_norm_stderr": 0.026890297881303125 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.49079754601226994, "acc_stderr": 0.039277056007874414, "acc_norm": 0.49079754601226994, "acc_norm_stderr": 0.039277056007874414 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.5401234567901234, "acc_stderr": 0.02773102275353928, "acc_norm": 0.5401234567901234, "acc_norm_stderr": 0.02773102275353928 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.26, "acc_stderr": 0.04408440022768079, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768079 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.6424870466321243, "acc_stderr": 0.03458816042181012, "acc_norm": 0.6424870466321243, "acc_norm_stderr": 0.03458816042181012 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.35964912280701755, "acc_stderr": 0.04514496132873633, "acc_norm": 0.35964912280701755, "acc_norm_stderr": 0.04514496132873633 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.6293577981651376, "acc_stderr": 0.020707458164352984, "acc_norm": 0.6293577981651376, "acc_norm_stderr": 0.020707458164352984 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.3253968253968254, "acc_stderr": 0.041905964388711366, "acc_norm": 0.3253968253968254, "acc_norm_stderr": 0.041905964388711366 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5326797385620915, "acc_stderr": 0.02856869975222588, "acc_norm": 0.5326797385620915, "acc_norm_stderr": 0.02856869975222588 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.48, "acc_stderr": 0.050211673156867795, "acc_norm": 0.48, "acc_norm_stderr": 0.050211673156867795 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6859504132231405, "acc_stderr": 0.0423696475304102, "acc_norm": 0.6859504132231405, "acc_norm_stderr": 0.0423696475304102 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.5526315789473685, "acc_stderr": 0.0404633688397825, "acc_norm": 0.5526315789473685, "acc_norm_stderr": 0.0404633688397825 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.4591503267973856, "acc_stderr": 0.020160213617222516, "acc_norm": 0.4591503267973856, "acc_norm_stderr": 0.020160213617222516 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3475177304964539, "acc_stderr": 0.02840662780959095, "acc_norm": 0.3475177304964539, "acc_norm_stderr": 0.02840662780959095 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.30357142857142855, "acc_stderr": 0.04364226155841044, "acc_norm": 0.30357142857142855, "acc_norm_stderr": 0.04364226155841044 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.37037037037037035, "acc_stderr": 0.03293377139415191, "acc_norm": 0.37037037037037035, "acc_norm_stderr": 0.03293377139415191 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.22793296089385476, "acc_stderr": 0.014030149950805097, "acc_norm": 0.22793296089385476, "acc_norm_stderr": 0.014030149950805097 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.48, "acc_stderr": 0.050211673156867795, "acc_norm": 0.48, "acc_norm_stderr": 0.050211673156867795 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.6, "acc_stderr": 0.04923659639173309, "acc_norm": 0.6, "acc_norm_stderr": 0.04923659639173309 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.40441176470588236, "acc_stderr": 0.029812630701569746, "acc_norm": 0.40441176470588236, "acc_norm_stderr": 0.029812630701569746 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.5714285714285714, "acc_stderr": 0.03168091161233881, "acc_norm": 0.5714285714285714, "acc_norm_stderr": 0.03168091161233881 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.6624472573839663, "acc_stderr": 0.030781549102026212, "acc_norm": 0.6624472573839663, "acc_norm_stderr": 0.030781549102026212 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3513689700130378, "acc_stderr": 0.012192969457484035, "acc_norm": 0.3513689700130378, "acc_norm_stderr": 0.012192969457484035 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.6274509803921569, "acc_stderr": 0.033933885849584046, "acc_norm": 0.6274509803921569, "acc_norm_stderr": 0.033933885849584046 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.5757575757575758, "acc_stderr": 0.03859268142070262, "acc_norm": 0.5757575757575758, "acc_norm_stderr": 0.03859268142070262 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.3108935128518972, "mc1_stderr": 0.016203316673559696, "mc2": 0.4900739935604153, "mc2_stderr": 0.01572393085880814 }, "harness|ko_commongen_v2|2": { "acc": 0.602125147579693, "acc_stderr": 0.016827959054733395, "acc_norm": 0.615112160566706, "acc_norm_stderr": 0.016728579701498644 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Edentns/DataVortexS-10.7B-dpo-v1.9", "model_sha": "b9a06af3584f2e115f613e7086549ef6e67b5471", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }