{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.3532423208191126, "acc_stderr": 0.013967822714840053, "acc_norm": 0.3984641638225256, "acc_norm_stderr": 0.014306946052735565 }, "harness|ko_hellaswag|10": { "acc": 0.3676558454491137, "acc_stderr": 0.004811815959388833, "acc_norm": 0.47410874327823144, "acc_norm_stderr": 0.004983087049281746 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.4502923976608187, "acc_stderr": 0.038158273659132366, "acc_norm": 0.4502923976608187, "acc_norm_stderr": 0.038158273659132366 }, "harness|ko_mmlu_management|5": { "acc": 0.5922330097087378, "acc_stderr": 0.048657775704107696, "acc_norm": 0.5922330097087378, "acc_norm_stderr": 0.048657775704107696 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.5108556832694764, "acc_stderr": 0.017875748840242407, "acc_norm": 0.5108556832694764, "acc_norm_stderr": 0.017875748840242407 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.3925925925925926, "acc_stderr": 0.04218506215368879, "acc_norm": 0.3925925925925926, "acc_norm_stderr": 0.04218506215368879 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.32, "acc_stderr": 0.04688261722621503, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621503 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.3659574468085106, "acc_stderr": 0.0314895582974553, "acc_norm": 0.3659574468085106, "acc_norm_stderr": 0.0314895582974553 }, "harness|ko_mmlu_virology|5": { "acc": 0.35542168674698793, "acc_stderr": 0.03726214354322415, "acc_norm": 0.35542168674698793, "acc_norm_stderr": 0.03726214354322415 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.4662379421221865, "acc_stderr": 0.02833327710956279, "acc_norm": 0.4662379421221865, "acc_norm_stderr": 0.02833327710956279 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.45739910313901344, "acc_stderr": 0.033435777055830646, "acc_norm": 0.45739910313901344, "acc_norm_stderr": 0.033435777055830646 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.48091603053435117, "acc_stderr": 0.04382094705550989, "acc_norm": 0.48091603053435117, "acc_norm_stderr": 0.04382094705550989 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.42, "acc_stderr": 0.049604496374885836, "acc_norm": 0.42, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5202020202020202, "acc_stderr": 0.03559443565563919, "acc_norm": 0.5202020202020202, "acc_norm_stderr": 0.03559443565563919 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.41379310344827586, "acc_stderr": 0.04104269211806232, "acc_norm": 0.41379310344827586, "acc_norm_stderr": 0.04104269211806232 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.18627450980392157, "acc_stderr": 0.03873958714149352, "acc_norm": 0.18627450980392157, "acc_norm_stderr": 0.03873958714149352 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.44537815126050423, "acc_stderr": 0.0322841062671639, "acc_norm": 0.44537815126050423, "acc_norm_stderr": 0.0322841062671639 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.4358974358974359, "acc_stderr": 0.025141801511177495, "acc_norm": 0.4358974358974359, "acc_norm_stderr": 0.025141801511177495 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.43, "acc_stderr": 0.04975698519562428, "acc_norm": 0.43, "acc_norm_stderr": 0.04975698519562428 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.4, "acc_stderr": 0.049236596391733084, "acc_norm": 0.4, "acc_norm_stderr": 0.049236596391733084 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5277777777777778, "acc_stderr": 0.04826217294139894, "acc_norm": 0.5277777777777778, "acc_norm_stderr": 0.04826217294139894 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.33497536945812806, "acc_stderr": 0.033208527423483104, "acc_norm": 0.33497536945812806, "acc_norm_stderr": 0.033208527423483104 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.4645161290322581, "acc_stderr": 0.02837228779796295, "acc_norm": 0.4645161290322581, "acc_norm_stderr": 0.02837228779796295 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6495726495726496, "acc_stderr": 0.0312561082442188, "acc_norm": 0.6495726495726496, "acc_norm_stderr": 0.0312561082442188 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.44150943396226416, "acc_stderr": 0.030561590426731833, "acc_norm": 0.44150943396226416, "acc_norm_stderr": 0.030561590426731833 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.4727272727272727, "acc_stderr": 0.04782001791380063, "acc_norm": 0.4727272727272727, "acc_norm_stderr": 0.04782001791380063 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.2962962962962963, "acc_stderr": 0.027840811495871913, "acc_norm": 0.2962962962962963, "acc_norm_stderr": 0.027840811495871913 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2781456953642384, "acc_stderr": 0.03658603262763743, "acc_norm": 0.2781456953642384, "acc_norm_stderr": 0.03658603262763743 }, "harness|ko_mmlu_sociology|5": { "acc": 0.572139303482587, "acc_stderr": 0.03498541988407795, "acc_norm": 0.572139303482587, "acc_norm_stderr": 0.03498541988407795 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.3930635838150289, "acc_stderr": 0.03724249595817731, "acc_norm": 0.3930635838150289, "acc_norm_stderr": 0.03724249595817731 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2857142857142857, "acc_stderr": 0.02326651221373057, "acc_norm": 0.2857142857142857, "acc_norm_stderr": 0.02326651221373057 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3541666666666667, "acc_stderr": 0.039994111357535424, "acc_norm": 0.3541666666666667, "acc_norm_stderr": 0.039994111357535424 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.59, "acc_stderr": 0.04943110704237102, "acc_norm": 0.59, "acc_norm_stderr": 0.04943110704237102 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5202312138728323, "acc_stderr": 0.02689704999638286, "acc_norm": 0.5202312138728323, "acc_norm_stderr": 0.02689704999638286 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.4049079754601227, "acc_stderr": 0.03856672163548913, "acc_norm": 0.4049079754601227, "acc_norm_stderr": 0.03856672163548913 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.4444444444444444, "acc_stderr": 0.027648477877413324, "acc_norm": 0.4444444444444444, "acc_norm_stderr": 0.027648477877413324 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.36, "acc_stderr": 0.04824181513244218, "acc_norm": 0.36, "acc_norm_stderr": 0.04824181513244218 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.5129533678756477, "acc_stderr": 0.03607228061047749, "acc_norm": 0.5129533678756477, "acc_norm_stderr": 0.03607228061047749 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.23684210526315788, "acc_stderr": 0.03999423879281338, "acc_norm": 0.23684210526315788, "acc_norm_stderr": 0.03999423879281338 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.47706422018348627, "acc_stderr": 0.0214147570581755, "acc_norm": 0.47706422018348627, "acc_norm_stderr": 0.0214147570581755 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.36507936507936506, "acc_stderr": 0.04306241259127153, "acc_norm": 0.36507936507936506, "acc_norm_stderr": 0.04306241259127153 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.434640522875817, "acc_stderr": 0.028384256704883037, "acc_norm": 0.434640522875817, "acc_norm_stderr": 0.028384256704883037 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.42, "acc_stderr": 0.049604496374885836, "acc_norm": 0.42, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6611570247933884, "acc_stderr": 0.043207678075366705, "acc_norm": 0.6611570247933884, "acc_norm_stderr": 0.043207678075366705 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.4144736842105263, "acc_stderr": 0.04008973785779206, "acc_norm": 0.4144736842105263, "acc_norm_stderr": 0.04008973785779206 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.3562091503267974, "acc_stderr": 0.0193733324207245, "acc_norm": 0.3562091503267974, "acc_norm_stderr": 0.0193733324207245 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3333333333333333, "acc_stderr": 0.02812163604063989, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.02812163604063989 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.29464285714285715, "acc_stderr": 0.043270409325787296, "acc_norm": 0.29464285714285715, "acc_norm_stderr": 0.043270409325787296 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.32407407407407407, "acc_stderr": 0.03191923445686186, "acc_norm": 0.32407407407407407, "acc_norm_stderr": 0.03191923445686186 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2581005586592179, "acc_stderr": 0.01463518561652782, "acc_norm": 0.2581005586592179, "acc_norm_stderr": 0.01463518561652782 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.38, "acc_stderr": 0.04878317312145632, "acc_norm": 0.38, "acc_norm_stderr": 0.04878317312145632 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.45, "acc_stderr": 0.05, "acc_norm": 0.45, "acc_norm_stderr": 0.05 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.34558823529411764, "acc_stderr": 0.028888193103988644, "acc_norm": 0.34558823529411764, "acc_norm_stderr": 0.028888193103988644 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.5387755102040817, "acc_stderr": 0.03191282052669278, "acc_norm": 0.5387755102040817, "acc_norm_stderr": 0.03191282052669278 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.5569620253164557, "acc_stderr": 0.03233532777533484, "acc_norm": 0.5569620253164557, "acc_norm_stderr": 0.03233532777533484 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3494132985658409, "acc_stderr": 0.012177306252786702, "acc_norm": 0.3494132985658409, "acc_norm_stderr": 0.012177306252786702 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.4852941176470588, "acc_stderr": 0.03507793834791324, "acc_norm": 0.4852941176470588, "acc_norm_stderr": 0.03507793834791324 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.503030303030303, "acc_stderr": 0.03904272341431856, "acc_norm": 0.503030303030303, "acc_norm_stderr": 0.03904272341431856 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2962056303549572, "mc1_stderr": 0.015983595101811392, "mc2": 0.4656302357866705, "mc2_stderr": 0.01570298909368207 }, "harness|ko_commongen_v2|2": { "acc": 0.43683589138134593, "acc_stderr": 0.017052633559856076, "acc_norm": 0.45808736717827625, "acc_norm_stderr": 0.017129852117911144 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "HwiyeolJo/testttt", "model_sha": "7409deaa6c866195b31dc8482414b4d64a40c372", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }