{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.3447098976109215, "acc_stderr": 0.01388881628678211, "acc_norm": 0.41552901023890787, "acc_norm_stderr": 0.01440136664121639 }, "harness|ko_hellaswag|10": { "acc": 0.40300736904999, "acc_stderr": 0.004894997736719054, "acc_norm": 0.5375423222465644, "acc_norm_stderr": 0.00497569607624085 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.5029239766081871, "acc_stderr": 0.03834759370936839, "acc_norm": 0.5029239766081871, "acc_norm_stderr": 0.03834759370936839 }, "harness|ko_mmlu_management|5": { "acc": 0.6019417475728155, "acc_stderr": 0.048467482539772386, "acc_norm": 0.6019417475728155, "acc_norm_stderr": 0.048467482539772386 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.5542784163473818, "acc_stderr": 0.01777429728247951, "acc_norm": 0.5542784163473818, "acc_norm_stderr": 0.01777429728247951 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4740740740740741, "acc_stderr": 0.04313531696750573, "acc_norm": 0.4740740740740741, "acc_norm_stderr": 0.04313531696750573 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.26, "acc_stderr": 0.04408440022768078, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768078 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.39148936170212767, "acc_stderr": 0.03190701242326812, "acc_norm": 0.39148936170212767, "acc_norm_stderr": 0.03190701242326812 }, "harness|ko_mmlu_virology|5": { "acc": 0.3855421686746988, "acc_stderr": 0.037891344246115496, "acc_norm": 0.3855421686746988, "acc_norm_stderr": 0.037891344246115496 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5337620578778135, "acc_stderr": 0.0283332771095628, "acc_norm": 0.5337620578778135, "acc_norm_stderr": 0.0283332771095628 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.4663677130044843, "acc_stderr": 0.033481800170603065, "acc_norm": 0.4663677130044843, "acc_norm_stderr": 0.033481800170603065 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.5572519083969466, "acc_stderr": 0.04356447202665069, "acc_norm": 0.5572519083969466, "acc_norm_stderr": 0.04356447202665069 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.36, "acc_stderr": 0.04824181513244218, "acc_norm": 0.36, "acc_norm_stderr": 0.04824181513244218 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.6363636363636364, "acc_stderr": 0.034273086529999365, "acc_norm": 0.6363636363636364, "acc_norm_stderr": 0.034273086529999365 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.5103448275862069, "acc_stderr": 0.04165774775728763, "acc_norm": 0.5103448275862069, "acc_norm_stderr": 0.04165774775728763 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.19607843137254902, "acc_stderr": 0.03950581861179964, "acc_norm": 0.19607843137254902, "acc_norm_stderr": 0.03950581861179964 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.4579831932773109, "acc_stderr": 0.032363611119519416, "acc_norm": 0.4579831932773109, "acc_norm_stderr": 0.032363611119519416 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.4564102564102564, "acc_stderr": 0.02525448542479961, "acc_norm": 0.4564102564102564, "acc_norm_stderr": 0.02525448542479961 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.68, "acc_stderr": 0.046882617226215034, "acc_norm": 0.68, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.3, "acc_stderr": 0.04605661864718381, "acc_norm": 0.3, "acc_norm_stderr": 0.04605661864718381 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.48148148148148145, "acc_stderr": 0.04830366024635331, "acc_norm": 0.48148148148148145, "acc_norm_stderr": 0.04830366024635331 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3842364532019704, "acc_stderr": 0.0342239856565755, "acc_norm": 0.3842364532019704, "acc_norm_stderr": 0.0342239856565755 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.43548387096774194, "acc_stderr": 0.028206225591502734, "acc_norm": 0.43548387096774194, "acc_norm_stderr": 0.028206225591502734 }, "harness|ko_mmlu_marketing|5": { "acc": 0.717948717948718, "acc_stderr": 0.029480360549541194, "acc_norm": 0.717948717948718, "acc_norm_stderr": 0.029480360549541194 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.4528301886792453, "acc_stderr": 0.030635627957961816, "acc_norm": 0.4528301886792453, "acc_norm_stderr": 0.030635627957961816 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.4727272727272727, "acc_stderr": 0.04782001791380063, "acc_norm": 0.4727272727272727, "acc_norm_stderr": 0.04782001791380063 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.3074074074074074, "acc_stderr": 0.02813325257881564, "acc_norm": 0.3074074074074074, "acc_norm_stderr": 0.02813325257881564 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.304635761589404, "acc_stderr": 0.03757949922943342, "acc_norm": 0.304635761589404, "acc_norm_stderr": 0.03757949922943342 }, "harness|ko_mmlu_sociology|5": { "acc": 0.582089552238806, "acc_stderr": 0.034875586404620636, "acc_norm": 0.582089552238806, "acc_norm_stderr": 0.034875586404620636 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.4046242774566474, "acc_stderr": 0.0374246119388725, "acc_norm": 0.4046242774566474, "acc_norm_stderr": 0.0374246119388725 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.3201058201058201, "acc_stderr": 0.024026846392873502, "acc_norm": 0.3201058201058201, "acc_norm_stderr": 0.024026846392873502 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.4513888888888889, "acc_stderr": 0.04161402398403279, "acc_norm": 0.4513888888888889, "acc_norm_stderr": 0.04161402398403279 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.29, "acc_stderr": 0.045604802157206845, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206845 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.58, "acc_stderr": 0.049604496374885836, "acc_norm": 0.58, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5028901734104047, "acc_stderr": 0.026918645383239015, "acc_norm": 0.5028901734104047, "acc_norm_stderr": 0.026918645383239015 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.4662576687116564, "acc_stderr": 0.03919415545048409, "acc_norm": 0.4662576687116564, "acc_norm_stderr": 0.03919415545048409 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.49382716049382713, "acc_stderr": 0.027818623962583302, "acc_norm": 0.49382716049382713, "acc_norm_stderr": 0.027818623962583302 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.6010362694300518, "acc_stderr": 0.035339990940656964, "acc_norm": 0.6010362694300518, "acc_norm_stderr": 0.035339990940656964 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2894736842105263, "acc_stderr": 0.04266339443159394, "acc_norm": 0.2894736842105263, "acc_norm_stderr": 0.04266339443159394 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.6, "acc_stderr": 0.021004201260420078, "acc_norm": 0.6, "acc_norm_stderr": 0.021004201260420078 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.2619047619047619, "acc_stderr": 0.0393253768039287, "acc_norm": 0.2619047619047619, "acc_norm_stderr": 0.0393253768039287 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.4542483660130719, "acc_stderr": 0.028509807802626567, "acc_norm": 0.4542483660130719, "acc_norm_stderr": 0.028509807802626567 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.52, "acc_stderr": 0.050211673156867795, "acc_norm": 0.52, "acc_norm_stderr": 0.050211673156867795 }, "harness|ko_mmlu_international_law|5": { "acc": 0.628099173553719, "acc_stderr": 0.04412015806624504, "acc_norm": 0.628099173553719, "acc_norm_stderr": 0.04412015806624504 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.4407894736842105, "acc_stderr": 0.040403110624904356, "acc_norm": 0.4407894736842105, "acc_norm_stderr": 0.040403110624904356 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.39705882352941174, "acc_stderr": 0.019794488900024103, "acc_norm": 0.39705882352941174, "acc_norm_stderr": 0.019794488900024103 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.32978723404255317, "acc_stderr": 0.028045946942042398, "acc_norm": 0.32978723404255317, "acc_norm_stderr": 0.028045946942042398 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.33035714285714285, "acc_stderr": 0.044642857142857116, "acc_norm": 0.33035714285714285, "acc_norm_stderr": 0.044642857142857116 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.25925925925925924, "acc_stderr": 0.029886910547626985, "acc_norm": 0.25925925925925924, "acc_norm_stderr": 0.029886910547626985 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2212290502793296, "acc_stderr": 0.013882164598887265, "acc_norm": 0.2212290502793296, "acc_norm_stderr": 0.013882164598887265 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.51, "acc_stderr": 0.05024183937956912, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956912 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.3786764705882353, "acc_stderr": 0.029465133639776125, "acc_norm": 0.3786764705882353, "acc_norm_stderr": 0.029465133639776125 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.37551020408163266, "acc_stderr": 0.031001209039894836, "acc_norm": 0.37551020408163266, "acc_norm_stderr": 0.031001209039894836 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.5738396624472574, "acc_stderr": 0.03219035703131775, "acc_norm": 0.5738396624472574, "acc_norm_stderr": 0.03219035703131775 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.32073011734028684, "acc_stderr": 0.011921199991782622, "acc_norm": 0.32073011734028684, "acc_norm_stderr": 0.011921199991782622 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.5, "acc_stderr": 0.03509312031717982, "acc_norm": 0.5, "acc_norm_stderr": 0.03509312031717982 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.5151515151515151, "acc_stderr": 0.03902551007374448, "acc_norm": 0.5151515151515151, "acc_norm_stderr": 0.03902551007374448 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2802937576499388, "mc1_stderr": 0.015723139524608746, "mc2": 0.4370086270534907, "mc2_stderr": 0.01529592374430609 }, "harness|ko_commongen_v2|2": { "acc": 0.35064935064935066, "acc_stderr": 0.0164055569038933, "acc_norm": 0.39315230224321135, "acc_norm_stderr": 0.01679326280128708 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "DooDooHyun/AIFT-Yi-Ko-6B-ao-instruct-all-v0.54", "model_sha": "b5c70d1bb01aa3b9b6f756c301d908c95d80900b", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }