{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.37457337883959047, "acc_stderr": 0.014144193471893456, "acc_norm": 0.43430034129692835, "acc_norm_stderr": 0.014484703048857355 }, "harness|ko_hellaswag|10": { "acc": 0.4070902210714997, "acc_stderr": 0.004902878806733046, "acc_norm": 0.5408285202150966, "acc_norm_stderr": 0.004973117975062488 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.5380116959064327, "acc_stderr": 0.03823727092882307, "acc_norm": 0.5380116959064327, "acc_norm_stderr": 0.03823727092882307 }, "harness|ko_mmlu_management|5": { "acc": 0.5533980582524272, "acc_stderr": 0.04922424153458935, "acc_norm": 0.5533980582524272, "acc_norm_stderr": 0.04922424153458935 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.5197956577266922, "acc_stderr": 0.017865944827291633, "acc_norm": 0.5197956577266922, "acc_norm_stderr": 0.017865944827291633 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.43703703703703706, "acc_stderr": 0.04284958639753399, "acc_norm": 0.43703703703703706, "acc_norm_stderr": 0.04284958639753399 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.29, "acc_stderr": 0.045604802157206824, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206824 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.3446808510638298, "acc_stderr": 0.03106898596312215, "acc_norm": 0.3446808510638298, "acc_norm_stderr": 0.03106898596312215 }, "harness|ko_mmlu_virology|5": { "acc": 0.40963855421686746, "acc_stderr": 0.038284011150790206, "acc_norm": 0.40963855421686746, "acc_norm_stderr": 0.038284011150790206 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5241157556270096, "acc_stderr": 0.028365041542564584, "acc_norm": 0.5241157556270096, "acc_norm_stderr": 0.028365041542564584 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.4170403587443946, "acc_stderr": 0.03309266936071721, "acc_norm": 0.4170403587443946, "acc_norm_stderr": 0.03309266936071721 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.5114503816793893, "acc_stderr": 0.043841400240780176, "acc_norm": 0.5114503816793893, "acc_norm_stderr": 0.043841400240780176 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.36, "acc_stderr": 0.04824181513244218, "acc_norm": 0.36, "acc_norm_stderr": 0.04824181513244218 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5454545454545454, "acc_stderr": 0.035476014940069384, "acc_norm": 0.5454545454545454, "acc_norm_stderr": 0.035476014940069384 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.47586206896551725, "acc_stderr": 0.041618085035015295, "acc_norm": 0.47586206896551725, "acc_norm_stderr": 0.041618085035015295 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.18627450980392157, "acc_stderr": 0.03873958714149352, "acc_norm": 0.18627450980392157, "acc_norm_stderr": 0.03873958714149352 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.42016806722689076, "acc_stderr": 0.03206183783236153, "acc_norm": 0.42016806722689076, "acc_norm_stderr": 0.03206183783236153 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.4153846153846154, "acc_stderr": 0.024985354923102318, "acc_norm": 0.4153846153846154, "acc_norm_stderr": 0.024985354923102318 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.47, "acc_stderr": 0.050161355804659205, "acc_norm": 0.47, "acc_norm_stderr": 0.050161355804659205 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.38, "acc_stderr": 0.04878317312145633, "acc_norm": 0.38, "acc_norm_stderr": 0.04878317312145633 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.48148148148148145, "acc_stderr": 0.04830366024635331, "acc_norm": 0.48148148148148145, "acc_norm_stderr": 0.04830366024635331 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3497536945812808, "acc_stderr": 0.03355400904969566, "acc_norm": 0.3497536945812808, "acc_norm_stderr": 0.03355400904969566 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.4612903225806452, "acc_stderr": 0.028358634859836914, "acc_norm": 0.4612903225806452, "acc_norm_stderr": 0.028358634859836914 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6495726495726496, "acc_stderr": 0.0312561082442188, "acc_norm": 0.6495726495726496, "acc_norm_stderr": 0.0312561082442188 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.4528301886792453, "acc_stderr": 0.030635627957961823, "acc_norm": 0.4528301886792453, "acc_norm_stderr": 0.030635627957961823 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.4909090909090909, "acc_stderr": 0.04788339768702861, "acc_norm": 0.4909090909090909, "acc_norm_stderr": 0.04788339768702861 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.29259259259259257, "acc_stderr": 0.027738969632176095, "acc_norm": 0.29259259259259257, "acc_norm_stderr": 0.027738969632176095 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.31788079470198677, "acc_stderr": 0.038020397601079024, "acc_norm": 0.31788079470198677, "acc_norm_stderr": 0.038020397601079024 }, "harness|ko_mmlu_sociology|5": { "acc": 0.572139303482587, "acc_stderr": 0.03498541988407795, "acc_norm": 0.572139303482587, "acc_norm_stderr": 0.03498541988407795 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.4046242774566474, "acc_stderr": 0.03742461193887248, "acc_norm": 0.4046242774566474, "acc_norm_stderr": 0.03742461193887248 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.30687830687830686, "acc_stderr": 0.023752928712112126, "acc_norm": 0.30687830687830686, "acc_norm_stderr": 0.023752928712112126 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3958333333333333, "acc_stderr": 0.04089465449325583, "acc_norm": 0.3958333333333333, "acc_norm_stderr": 0.04089465449325583 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.35, "acc_stderr": 0.0479372485441102, "acc_norm": 0.35, "acc_norm_stderr": 0.0479372485441102 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.7, "acc_stderr": 0.046056618647183814, "acc_norm": 0.7, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.48265895953757226, "acc_stderr": 0.02690290045866664, "acc_norm": 0.48265895953757226, "acc_norm_stderr": 0.02690290045866664 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.39263803680981596, "acc_stderr": 0.03836740907831029, "acc_norm": 0.39263803680981596, "acc_norm_stderr": 0.03836740907831029 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.48148148148148145, "acc_stderr": 0.027801656212323667, "acc_norm": 0.48148148148148145, "acc_norm_stderr": 0.027801656212323667 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.28, "acc_stderr": 0.045126085985421276, "acc_norm": 0.28, "acc_norm_stderr": 0.045126085985421276 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.5595854922279793, "acc_stderr": 0.03582724530036094, "acc_norm": 0.5595854922279793, "acc_norm_stderr": 0.03582724530036094 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.21929824561403508, "acc_stderr": 0.03892431106518753, "acc_norm": 0.21929824561403508, "acc_norm_stderr": 0.03892431106518753 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.5394495412844037, "acc_stderr": 0.021370494609995096, "acc_norm": 0.5394495412844037, "acc_norm_stderr": 0.021370494609995096 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.3492063492063492, "acc_stderr": 0.04263906892795133, "acc_norm": 0.3492063492063492, "acc_norm_stderr": 0.04263906892795133 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.4542483660130719, "acc_stderr": 0.028509807802626567, "acc_norm": 0.4542483660130719, "acc_norm_stderr": 0.028509807802626567 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.42, "acc_stderr": 0.049604496374885836, "acc_norm": 0.42, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6363636363636364, "acc_stderr": 0.043913262867240704, "acc_norm": 0.6363636363636364, "acc_norm_stderr": 0.043913262867240704 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.4407894736842105, "acc_stderr": 0.040403110624904356, "acc_norm": 0.4407894736842105, "acc_norm_stderr": 0.040403110624904356 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.3480392156862745, "acc_stderr": 0.019270998708223974, "acc_norm": 0.3480392156862745, "acc_norm_stderr": 0.019270998708223974 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.35815602836879434, "acc_stderr": 0.02860208586275942, "acc_norm": 0.35815602836879434, "acc_norm_stderr": 0.02860208586275942 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.24107142857142858, "acc_stderr": 0.04059867246952688, "acc_norm": 0.24107142857142858, "acc_norm_stderr": 0.04059867246952688 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.2916666666666667, "acc_stderr": 0.03099866630456053, "acc_norm": 0.2916666666666667, "acc_norm_stderr": 0.03099866630456053 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24022346368715083, "acc_stderr": 0.014288343803925295, "acc_norm": 0.24022346368715083, "acc_norm_stderr": 0.014288343803925295 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.44, "acc_stderr": 0.04988876515698589, "acc_norm": 0.44, "acc_norm_stderr": 0.04988876515698589 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.47, "acc_stderr": 0.05016135580465919, "acc_norm": 0.47, "acc_norm_stderr": 0.05016135580465919 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.3639705882352941, "acc_stderr": 0.02922719246003203, "acc_norm": 0.3639705882352941, "acc_norm_stderr": 0.02922719246003203 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.4775510204081633, "acc_stderr": 0.03197694118713673, "acc_norm": 0.4775510204081633, "acc_norm_stderr": 0.03197694118713673 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.5949367088607594, "acc_stderr": 0.031955147413706725, "acc_norm": 0.5949367088607594, "acc_norm_stderr": 0.031955147413706725 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.35919165580182527, "acc_stderr": 0.012253386187584245, "acc_norm": 0.35919165580182527, "acc_norm_stderr": 0.012253386187584245 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.5049019607843137, "acc_stderr": 0.03509143375606786, "acc_norm": 0.5049019607843137, "acc_norm_stderr": 0.03509143375606786 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.503030303030303, "acc_stderr": 0.03904272341431856, "acc_norm": 0.503030303030303, "acc_norm_stderr": 0.03904272341431856 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.27539779681762544, "mc1_stderr": 0.015638135667775523, "mc2": 0.4457475184349363, "mc2_stderr": 0.015091782961916999 }, "harness|ko_commongen_v2|2": { "acc": 0.38134592680047225, "acc_stderr": 0.016699301768828074, "acc_norm": 0.40613931523022434, "acc_norm_stderr": 0.016884749503191392 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "kiyoonyoo/ko-en-trans-platypus-13b-v3", "model_sha": "3c27d710886ff8b6a9fcf321fae0e2f76eaeafa3", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }