{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.2901023890784983, "acc_stderr": 0.013261573677520776, "acc_norm": 0.3361774744027304, "acc_norm_stderr": 0.013804855026205756 }, "harness|ko_hellaswag|10": { "acc": 0.3567018522206732, "acc_stderr": 0.004780467270911769, "acc_norm": 0.4511053574985063, "acc_norm_stderr": 0.004965866098318165 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.30994152046783624, "acc_stderr": 0.035469769593931624, "acc_norm": 0.30994152046783624, "acc_norm_stderr": 0.035469769593931624 }, "harness|ko_mmlu_management|5": { "acc": 0.2524271844660194, "acc_stderr": 0.04301250399690876, "acc_norm": 0.2524271844660194, "acc_norm_stderr": 0.04301250399690876 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.23116219667943805, "acc_stderr": 0.015075523238101081, "acc_norm": 0.23116219667943805, "acc_norm_stderr": 0.015075523238101081 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.23703703703703705, "acc_stderr": 0.03673731683969506, "acc_norm": 0.23703703703703705, "acc_norm_stderr": 0.03673731683969506 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.28, "acc_stderr": 0.045126085985421276, "acc_norm": 0.28, "acc_norm_stderr": 0.045126085985421276 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.225531914893617, "acc_stderr": 0.02732107841738753, "acc_norm": 0.225531914893617, "acc_norm_stderr": 0.02732107841738753 }, "harness|ko_mmlu_virology|5": { "acc": 0.3313253012048193, "acc_stderr": 0.03664314777288085, "acc_norm": 0.3313253012048193, "acc_norm_stderr": 0.03664314777288085 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.2765273311897106, "acc_stderr": 0.025403832978179615, "acc_norm": 0.2765273311897106, "acc_norm_stderr": 0.025403832978179615 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.2645739910313901, "acc_stderr": 0.029605103217038336, "acc_norm": 0.2645739910313901, "acc_norm_stderr": 0.029605103217038336 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.26717557251908397, "acc_stderr": 0.03880848301082395, "acc_norm": 0.26717557251908397, "acc_norm_stderr": 0.03880848301082395 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.17676767676767677, "acc_stderr": 0.027178752639044915, "acc_norm": 0.17676767676767677, "acc_norm_stderr": 0.027178752639044915 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.2482758620689655, "acc_stderr": 0.036001056927277716, "acc_norm": 0.2482758620689655, "acc_norm_stderr": 0.036001056927277716 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.3333333333333333, "acc_stderr": 0.04690650298201942, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.04690650298201942 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.24789915966386555, "acc_stderr": 0.028047967224176896, "acc_norm": 0.24789915966386555, "acc_norm_stderr": 0.028047967224176896 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.26153846153846155, "acc_stderr": 0.022282141204204426, "acc_norm": 0.26153846153846155, "acc_norm_stderr": 0.022282141204204426 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.24, "acc_stderr": 0.04292346959909283, "acc_norm": 0.24, "acc_norm_stderr": 0.04292346959909283 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.2777777777777778, "acc_stderr": 0.043300437496507416, "acc_norm": 0.2777777777777778, "acc_norm_stderr": 0.043300437496507416 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.16748768472906403, "acc_stderr": 0.026273086047535407, "acc_norm": 0.16748768472906403, "acc_norm_stderr": 0.026273086047535407 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.2806451612903226, "acc_stderr": 0.025560604721022884, "acc_norm": 0.2806451612903226, "acc_norm_stderr": 0.025560604721022884 }, "harness|ko_mmlu_marketing|5": { "acc": 0.29914529914529914, "acc_stderr": 0.029996951858349497, "acc_norm": 0.29914529914529914, "acc_norm_stderr": 0.029996951858349497 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.21132075471698114, "acc_stderr": 0.025125766484827852, "acc_norm": 0.21132075471698114, "acc_norm_stderr": 0.025125766484827852 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.3181818181818182, "acc_stderr": 0.04461272175910507, "acc_norm": 0.3181818181818182, "acc_norm_stderr": 0.04461272175910507 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.24074074074074073, "acc_stderr": 0.026067159222275805, "acc_norm": 0.24074074074074073, "acc_norm_stderr": 0.026067159222275805 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.33774834437086093, "acc_stderr": 0.03861557546255169, "acc_norm": 0.33774834437086093, "acc_norm_stderr": 0.03861557546255169 }, "harness|ko_mmlu_sociology|5": { "acc": 0.24875621890547264, "acc_stderr": 0.030567675938916707, "acc_norm": 0.24875621890547264, "acc_norm_stderr": 0.030567675938916707 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.1907514450867052, "acc_stderr": 0.029957851329869334, "acc_norm": 0.1907514450867052, "acc_norm_stderr": 0.029957851329869334 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2698412698412698, "acc_stderr": 0.022860838309232072, "acc_norm": 0.2698412698412698, "acc_norm_stderr": 0.022860838309232072 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.20833333333333334, "acc_stderr": 0.03396116205845333, "acc_norm": 0.20833333333333334, "acc_norm_stderr": 0.03396116205845333 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.29, "acc_stderr": 0.04560480215720684, "acc_norm": 0.29, "acc_norm_stderr": 0.04560480215720684 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.23699421965317918, "acc_stderr": 0.02289408248992599, "acc_norm": 0.23699421965317918, "acc_norm_stderr": 0.02289408248992599 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.25766871165644173, "acc_stderr": 0.03436150827846917, "acc_norm": 0.25766871165644173, "acc_norm_stderr": 0.03436150827846917 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.27469135802469136, "acc_stderr": 0.024836057868294677, "acc_norm": 0.27469135802469136, "acc_norm_stderr": 0.024836057868294677 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.19689119170984457, "acc_stderr": 0.028697873971860677, "acc_norm": 0.19689119170984457, "acc_norm_stderr": 0.028697873971860677 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.22807017543859648, "acc_stderr": 0.03947152782669415, "acc_norm": 0.22807017543859648, "acc_norm_stderr": 0.03947152782669415 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.21467889908256882, "acc_stderr": 0.017604304149256483, "acc_norm": 0.21467889908256882, "acc_norm_stderr": 0.017604304149256483 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.23015873015873015, "acc_stderr": 0.037649508797906045, "acc_norm": 0.23015873015873015, "acc_norm_stderr": 0.037649508797906045 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.24509803921568626, "acc_stderr": 0.024630048979824768, "acc_norm": 0.24509803921568626, "acc_norm_stderr": 0.024630048979824768 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.29, "acc_stderr": 0.045604802157206845, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206845 }, "harness|ko_mmlu_international_law|5": { "acc": 0.3140495867768595, "acc_stderr": 0.04236964753041019, "acc_norm": 0.3140495867768595, "acc_norm_stderr": 0.04236964753041019 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.23684210526315788, "acc_stderr": 0.03459777606810535, "acc_norm": 0.23684210526315788, "acc_norm_stderr": 0.03459777606810535 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.25, "acc_stderr": 0.01751781884501444, "acc_norm": 0.25, "acc_norm_stderr": 0.01751781884501444 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.2765957446808511, "acc_stderr": 0.026684564340460987, "acc_norm": 0.2765957446808511, "acc_norm_stderr": 0.026684564340460987 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.21428571428571427, "acc_stderr": 0.03894641120044792, "acc_norm": 0.21428571428571427, "acc_norm_stderr": 0.03894641120044792 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.3055555555555556, "acc_stderr": 0.031415546294025445, "acc_norm": 0.3055555555555556, "acc_norm_stderr": 0.031415546294025445 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.27262569832402234, "acc_stderr": 0.014893391735249608, "acc_norm": 0.27262569832402234, "acc_norm_stderr": 0.014893391735249608 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.22058823529411764, "acc_stderr": 0.025187786660227255, "acc_norm": 0.22058823529411764, "acc_norm_stderr": 0.025187786660227255 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.20816326530612245, "acc_stderr": 0.025991117672813296, "acc_norm": 0.20816326530612245, "acc_norm_stderr": 0.025991117672813296 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.3206751054852321, "acc_stderr": 0.030381931949990407, "acc_norm": 0.3206751054852321, "acc_norm_stderr": 0.030381931949990407 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.2627118644067797, "acc_stderr": 0.011240545514995676, "acc_norm": 0.2627118644067797, "acc_norm_stderr": 0.011240545514995676 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.25, "acc_stderr": 0.03039153369274154, "acc_norm": 0.25, "acc_norm_stderr": 0.03039153369274154 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.2727272727272727, "acc_stderr": 0.0347769116216366, "acc_norm": 0.2727272727272727, "acc_norm_stderr": 0.0347769116216366 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.26805385556915545, "mc1_stderr": 0.015506204722834557, "mc2": 0.4175867544708941, "mc2_stderr": 0.014944385749223169 }, "harness|ko_commongen_v2|2": { "acc": 0.26564344746162927, "acc_stderr": 0.015185107107791253, "acc_norm": 0.39787485242030696, "acc_norm_stderr": 0.016827959054733388 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "AIFT/AIFT-instruct-dpo-v1.3-42dot_LLM-SFT-1.3B", "model_sha": "aa6cc70a444c3c9f933ceefec4e84add872a37b1", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }