{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.4104095563139932, "acc_stderr": 0.014374922192642662, "acc_norm": 0.47013651877133106, "acc_norm_stderr": 0.014585305840007105 }, "harness|ko_hellaswag|10": { "acc": 0.427504481179048, "acc_stderr": 0.004937054233711568, "acc_norm": 0.5733917546305517, "acc_norm_stderr": 0.004935735300348869 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.49707602339181284, "acc_stderr": 0.03834759370936839, "acc_norm": 0.49707602339181284, "acc_norm_stderr": 0.03834759370936839 }, "harness|ko_mmlu_management|5": { "acc": 0.5242718446601942, "acc_stderr": 0.049449010929737795, "acc_norm": 0.5242718446601942, "acc_norm_stderr": 0.049449010929737795 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.5555555555555556, "acc_stderr": 0.017769250583533246, "acc_norm": 0.5555555555555556, "acc_norm_stderr": 0.017769250583533246 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.45185185185185184, "acc_stderr": 0.042992689054808624, "acc_norm": 0.45185185185185184, "acc_norm_stderr": 0.042992689054808624 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.29, "acc_stderr": 0.045604802157206824, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206824 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.42127659574468085, "acc_stderr": 0.03227834510146267, "acc_norm": 0.42127659574468085, "acc_norm_stderr": 0.03227834510146267 }, "harness|ko_mmlu_virology|5": { "acc": 0.40963855421686746, "acc_stderr": 0.038284011150790206, "acc_norm": 0.40963855421686746, "acc_norm_stderr": 0.038284011150790206 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.49517684887459806, "acc_stderr": 0.02839677044411129, "acc_norm": 0.49517684887459806, "acc_norm_stderr": 0.02839677044411129 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.5515695067264574, "acc_stderr": 0.033378837362550984, "acc_norm": 0.5515695067264574, "acc_norm_stderr": 0.033378837362550984 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.46564885496183206, "acc_stderr": 0.043749285605997376, "acc_norm": 0.46564885496183206, "acc_norm_stderr": 0.043749285605997376 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.4, "acc_stderr": 0.049236596391733084, "acc_norm": 0.4, "acc_norm_stderr": 0.049236596391733084 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5808080808080808, "acc_stderr": 0.035155207286704175, "acc_norm": 0.5808080808080808, "acc_norm_stderr": 0.035155207286704175 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.38620689655172413, "acc_stderr": 0.04057324734419034, "acc_norm": 0.38620689655172413, "acc_norm_stderr": 0.04057324734419034 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.23529411764705882, "acc_stderr": 0.04220773659171452, "acc_norm": 0.23529411764705882, "acc_norm_stderr": 0.04220773659171452 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.4789915966386555, "acc_stderr": 0.03244980849990029, "acc_norm": 0.4789915966386555, "acc_norm_stderr": 0.03244980849990029 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.45384615384615384, "acc_stderr": 0.02524277098712617, "acc_norm": 0.45384615384615384, "acc_norm_stderr": 0.02524277098712617 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.53, "acc_stderr": 0.05016135580465919, "acc_norm": 0.53, "acc_norm_stderr": 0.05016135580465919 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.4, "acc_stderr": 0.049236596391733084, "acc_norm": 0.4, "acc_norm_stderr": 0.049236596391733084 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5092592592592593, "acc_stderr": 0.04832853553437055, "acc_norm": 0.5092592592592593, "acc_norm_stderr": 0.04832853553437055 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.41379310344827586, "acc_stderr": 0.03465304488406796, "acc_norm": 0.41379310344827586, "acc_norm_stderr": 0.03465304488406796 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.46774193548387094, "acc_stderr": 0.028384747788813326, "acc_norm": 0.46774193548387094, "acc_norm_stderr": 0.028384747788813326 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6538461538461539, "acc_stderr": 0.031166957367235903, "acc_norm": 0.6538461538461539, "acc_norm_stderr": 0.031166957367235903 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.44150943396226416, "acc_stderr": 0.030561590426731833, "acc_norm": 0.44150943396226416, "acc_norm_stderr": 0.030561590426731833 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5818181818181818, "acc_stderr": 0.04724577405731572, "acc_norm": 0.5818181818181818, "acc_norm_stderr": 0.04724577405731572 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.26666666666666666, "acc_stderr": 0.026962424325073824, "acc_norm": 0.26666666666666666, "acc_norm_stderr": 0.026962424325073824 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2913907284768212, "acc_stderr": 0.03710185726119994, "acc_norm": 0.2913907284768212, "acc_norm_stderr": 0.03710185726119994 }, "harness|ko_mmlu_sociology|5": { "acc": 0.6069651741293532, "acc_stderr": 0.0345368246603156, "acc_norm": 0.6069651741293532, "acc_norm_stderr": 0.0345368246603156 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.4161849710982659, "acc_stderr": 0.037585177754049466, "acc_norm": 0.4161849710982659, "acc_norm_stderr": 0.037585177754049466 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.291005291005291, "acc_stderr": 0.023393826500484875, "acc_norm": 0.291005291005291, "acc_norm_stderr": 0.023393826500484875 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.375, "acc_stderr": 0.04048439222695598, "acc_norm": 0.375, "acc_norm_stderr": 0.04048439222695598 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.63, "acc_stderr": 0.04852365870939101, "acc_norm": 0.63, "acc_norm_stderr": 0.04852365870939101 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5, "acc_stderr": 0.026919095102908273, "acc_norm": 0.5, "acc_norm_stderr": 0.026919095102908273 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.50920245398773, "acc_stderr": 0.03927705600787443, "acc_norm": 0.50920245398773, "acc_norm_stderr": 0.03927705600787443 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.4722222222222222, "acc_stderr": 0.027777777777777797, "acc_norm": 0.4722222222222222, "acc_norm_stderr": 0.027777777777777797 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.32, "acc_stderr": 0.04688261722621504, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621504 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.5647668393782384, "acc_stderr": 0.03578038165008586, "acc_norm": 0.5647668393782384, "acc_norm_stderr": 0.03578038165008586 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2894736842105263, "acc_stderr": 0.04266339443159394, "acc_norm": 0.2894736842105263, "acc_norm_stderr": 0.04266339443159394 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.6220183486238532, "acc_stderr": 0.02078918706672812, "acc_norm": 0.6220183486238532, "acc_norm_stderr": 0.02078918706672812 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.3253968253968254, "acc_stderr": 0.041905964388711366, "acc_norm": 0.3253968253968254, "acc_norm_stderr": 0.041905964388711366 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.42483660130718953, "acc_stderr": 0.028304576673141114, "acc_norm": 0.42483660130718953, "acc_norm_stderr": 0.028304576673141114 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.41, "acc_stderr": 0.04943110704237102, "acc_norm": 0.41, "acc_norm_stderr": 0.04943110704237102 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6446280991735537, "acc_stderr": 0.0436923632657398, "acc_norm": 0.6446280991735537, "acc_norm_stderr": 0.0436923632657398 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.40789473684210525, "acc_stderr": 0.03999309712777471, "acc_norm": 0.40789473684210525, "acc_norm_stderr": 0.03999309712777471 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.3954248366013072, "acc_stderr": 0.01978046595477752, "acc_norm": 0.3954248366013072, "acc_norm_stderr": 0.01978046595477752 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3617021276595745, "acc_stderr": 0.028663820147199492, "acc_norm": 0.3617021276595745, "acc_norm_stderr": 0.028663820147199492 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.19642857142857142, "acc_stderr": 0.037709700493470166, "acc_norm": 0.19642857142857142, "acc_norm_stderr": 0.037709700493470166 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.3425925925925926, "acc_stderr": 0.032365852526021574, "acc_norm": 0.3425925925925926, "acc_norm_stderr": 0.032365852526021574 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2424581005586592, "acc_stderr": 0.01433352205921789, "acc_norm": 0.2424581005586592, "acc_norm_stderr": 0.01433352205921789 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.35, "acc_stderr": 0.0479372485441102, "acc_norm": 0.35, "acc_norm_stderr": 0.0479372485441102 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.4, "acc_stderr": 0.049236596391733084, "acc_norm": 0.4, "acc_norm_stderr": 0.049236596391733084 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.45588235294117646, "acc_stderr": 0.030254372573976698, "acc_norm": 0.45588235294117646, "acc_norm_stderr": 0.030254372573976698 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.49795918367346936, "acc_stderr": 0.0320089533497105, "acc_norm": 0.49795918367346936, "acc_norm_stderr": 0.0320089533497105 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.6455696202531646, "acc_stderr": 0.0311373042971858, "acc_norm": 0.6455696202531646, "acc_norm_stderr": 0.0311373042971858 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3644067796610169, "acc_stderr": 0.012291694983056474, "acc_norm": 0.3644067796610169, "acc_norm_stderr": 0.012291694983056474 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.5245098039215687, "acc_stderr": 0.035050931943487976, "acc_norm": 0.5245098039215687, "acc_norm_stderr": 0.035050931943487976 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.5393939393939394, "acc_stderr": 0.03892207016552012, "acc_norm": 0.5393939393939394, "acc_norm_stderr": 0.03892207016552012 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.31211750305997554, "mc1_stderr": 0.01622075676952091, "mc2": 0.4713446696460188, "mc2_stderr": 0.01534716845498683 }, "harness|ko_commongen_v2|2": { "acc": 0.4935064935064935, "acc_stderr": 0.01718890435907731, "acc_norm": 0.5667060212514758, "acc_norm_stderr": 0.017036683641893098 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "HumanF-MarkrAI/pub-llama-13B-v6", "model_sha": "af533b0f41590d5d8c5ced917b19a82ac98bb201", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }