{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.34044368600682595, "acc_stderr": 0.013847460518892976, "acc_norm": 0.4044368600682594, "acc_norm_stderr": 0.014342036483436175 }, "harness|ko_hellaswag|10": { "acc": 0.37183827922724555, "acc_stderr": 0.004823078145064962, "acc_norm": 0.4825731925911173, "acc_norm_stderr": 0.004986749760948692 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.5614035087719298, "acc_stderr": 0.038057975055904594, "acc_norm": 0.5614035087719298, "acc_norm_stderr": 0.038057975055904594 }, "harness|ko_mmlu_management|5": { "acc": 0.5922330097087378, "acc_stderr": 0.04865777570410769, "acc_norm": 0.5922330097087378, "acc_norm_stderr": 0.04865777570410769 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.4878671775223499, "acc_stderr": 0.017874698667491345, "acc_norm": 0.4878671775223499, "acc_norm_stderr": 0.017874698667491345 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4074074074074074, "acc_stderr": 0.042446332383532286, "acc_norm": 0.4074074074074074, "acc_norm_stderr": 0.042446332383532286 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.43829787234042555, "acc_stderr": 0.03243618636108102, "acc_norm": 0.43829787234042555, "acc_norm_stderr": 0.03243618636108102 }, "harness|ko_mmlu_virology|5": { "acc": 0.3795180722891566, "acc_stderr": 0.037777988227480165, "acc_norm": 0.3795180722891566, "acc_norm_stderr": 0.037777988227480165 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5337620578778135, "acc_stderr": 0.0283332771095628, "acc_norm": 0.5337620578778135, "acc_norm_stderr": 0.0283332771095628 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.43946188340807174, "acc_stderr": 0.03331092511038179, "acc_norm": 0.43946188340807174, "acc_norm_stderr": 0.03331092511038179 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.5419847328244275, "acc_stderr": 0.04369802690578757, "acc_norm": 0.5419847328244275, "acc_norm_stderr": 0.04369802690578757 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.44, "acc_stderr": 0.04988876515698589, "acc_norm": 0.44, "acc_norm_stderr": 0.04988876515698589 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5707070707070707, "acc_stderr": 0.035265527246011986, "acc_norm": 0.5707070707070707, "acc_norm_stderr": 0.035265527246011986 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.5793103448275863, "acc_stderr": 0.0411391498118926, "acc_norm": 0.5793103448275863, "acc_norm_stderr": 0.0411391498118926 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.29411764705882354, "acc_stderr": 0.04533838195929776, "acc_norm": 0.29411764705882354, "acc_norm_stderr": 0.04533838195929776 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.5546218487394958, "acc_stderr": 0.03228410626716391, "acc_norm": 0.5546218487394958, "acc_norm_stderr": 0.03228410626716391 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.5, "acc_stderr": 0.02535100632816969, "acc_norm": 0.5, "acc_norm_stderr": 0.02535100632816969 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.62, "acc_stderr": 0.048783173121456316, "acc_norm": 0.62, "acc_norm_stderr": 0.048783173121456316 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.39, "acc_stderr": 0.04902071300001974, "acc_norm": 0.39, "acc_norm_stderr": 0.04902071300001974 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.6018518518518519, "acc_stderr": 0.04732332615978814, "acc_norm": 0.6018518518518519, "acc_norm_stderr": 0.04732332615978814 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.42857142857142855, "acc_stderr": 0.034819048444388045, "acc_norm": 0.42857142857142855, "acc_norm_stderr": 0.034819048444388045 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.5129032258064516, "acc_stderr": 0.02843453315268186, "acc_norm": 0.5129032258064516, "acc_norm_stderr": 0.02843453315268186 }, "harness|ko_mmlu_marketing|5": { "acc": 0.7350427350427351, "acc_stderr": 0.028911208802749472, "acc_norm": 0.7350427350427351, "acc_norm_stderr": 0.028911208802749472 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.5018867924528302, "acc_stderr": 0.030772653642075664, "acc_norm": 0.5018867924528302, "acc_norm_stderr": 0.030772653642075664 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.4909090909090909, "acc_stderr": 0.04788339768702861, "acc_norm": 0.4909090909090909, "acc_norm_stderr": 0.04788339768702861 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.37037037037037035, "acc_stderr": 0.02944316932303154, "acc_norm": 0.37037037037037035, "acc_norm_stderr": 0.02944316932303154 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.36423841059602646, "acc_stderr": 0.03929111781242741, "acc_norm": 0.36423841059602646, "acc_norm_stderr": 0.03929111781242741 }, "harness|ko_mmlu_sociology|5": { "acc": 0.6467661691542289, "acc_stderr": 0.03379790611796777, "acc_norm": 0.6467661691542289, "acc_norm_stderr": 0.03379790611796777 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.4161849710982659, "acc_stderr": 0.037585177754049466, "acc_norm": 0.4161849710982659, "acc_norm_stderr": 0.037585177754049466 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.36507936507936506, "acc_stderr": 0.02479606060269995, "acc_norm": 0.36507936507936506, "acc_norm_stderr": 0.02479606060269995 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.4930555555555556, "acc_stderr": 0.04180806750294938, "acc_norm": 0.4930555555555556, "acc_norm_stderr": 0.04180806750294938 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.43, "acc_stderr": 0.049756985195624284, "acc_norm": 0.43, "acc_norm_stderr": 0.049756985195624284 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.62, "acc_stderr": 0.04878317312145634, "acc_norm": 0.62, "acc_norm_stderr": 0.04878317312145634 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5317919075144508, "acc_stderr": 0.026864624366756656, "acc_norm": 0.5317919075144508, "acc_norm_stderr": 0.026864624366756656 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.44171779141104295, "acc_stderr": 0.03901591825836184, "acc_norm": 0.44171779141104295, "acc_norm_stderr": 0.03901591825836184 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.49382716049382713, "acc_stderr": 0.027818623962583295, "acc_norm": 0.49382716049382713, "acc_norm_stderr": 0.027818623962583295 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.35, "acc_stderr": 0.047937248544110196, "acc_norm": 0.35, "acc_norm_stderr": 0.047937248544110196 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.5440414507772021, "acc_stderr": 0.03594413711272437, "acc_norm": 0.5440414507772021, "acc_norm_stderr": 0.03594413711272437 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.3684210526315789, "acc_stderr": 0.04537815354939391, "acc_norm": 0.3684210526315789, "acc_norm_stderr": 0.04537815354939391 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.5871559633027523, "acc_stderr": 0.021109128133413913, "acc_norm": 0.5871559633027523, "acc_norm_stderr": 0.021109128133413913 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.373015873015873, "acc_stderr": 0.04325506042017086, "acc_norm": 0.373015873015873, "acc_norm_stderr": 0.04325506042017086 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5065359477124183, "acc_stderr": 0.028627470550556047, "acc_norm": 0.5065359477124183, "acc_norm_stderr": 0.028627470550556047 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.48, "acc_stderr": 0.050211673156867795, "acc_norm": 0.48, "acc_norm_stderr": 0.050211673156867795 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6942148760330579, "acc_stderr": 0.04205953933884124, "acc_norm": 0.6942148760330579, "acc_norm_stderr": 0.04205953933884124 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.47368421052631576, "acc_stderr": 0.04063302731486671, "acc_norm": 0.47368421052631576, "acc_norm_stderr": 0.04063302731486671 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.4019607843137255, "acc_stderr": 0.01983517648437538, "acc_norm": 0.4019607843137255, "acc_norm_stderr": 0.01983517648437538 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3049645390070922, "acc_stderr": 0.027464708442022128, "acc_norm": 0.3049645390070922, "acc_norm_stderr": 0.027464708442022128 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.3482142857142857, "acc_stderr": 0.04521829902833585, "acc_norm": 0.3482142857142857, "acc_norm_stderr": 0.04521829902833585 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.46296296296296297, "acc_stderr": 0.03400603625538272, "acc_norm": 0.46296296296296297, "acc_norm_stderr": 0.03400603625538272 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24581005586592178, "acc_stderr": 0.014400296429225612, "acc_norm": 0.24581005586592178, "acc_norm_stderr": 0.014400296429225612 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.44, "acc_stderr": 0.04988876515698589, "acc_norm": 0.44, "acc_norm_stderr": 0.04988876515698589 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.57, "acc_stderr": 0.04975698519562428, "acc_norm": 0.57, "acc_norm_stderr": 0.04975698519562428 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4264705882352941, "acc_stderr": 0.03004261583271487, "acc_norm": 0.4264705882352941, "acc_norm_stderr": 0.03004261583271487 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.6081632653061224, "acc_stderr": 0.031251275910891656, "acc_norm": 0.6081632653061224, "acc_norm_stderr": 0.031251275910891656 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.6033755274261603, "acc_stderr": 0.031843998738112236, "acc_norm": 0.6033755274261603, "acc_norm_stderr": 0.031843998738112236 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.34876140808344197, "acc_stderr": 0.012172035157127115, "acc_norm": 0.34876140808344197, "acc_norm_stderr": 0.012172035157127115 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.5343137254901961, "acc_stderr": 0.03501038327635897, "acc_norm": 0.5343137254901961, "acc_norm_stderr": 0.03501038327635897 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.5757575757575758, "acc_stderr": 0.038592681420702636, "acc_norm": 0.5757575757575758, "acc_norm_stderr": 0.038592681420702636 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.3157894736842105, "mc1_stderr": 0.01627228795791694, "mc2": 0.4887242465522298, "mc2_stderr": 0.015611726455962618 }, "harness|ko_commongen_v2|2": { "acc": 0.48760330578512395, "acc_stderr": 0.017185069732676528, "acc_norm": 0.5478158205430933, "acc_norm_stderr": 0.017111567130916796 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "12thD/ko-Llama-3-8B-sft-v0.3", "model_sha": "134a44b329a37805306c77e45e932d839cae8baa", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }