{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.19027303754266212, "acc_stderr": 0.011470424179225702, "acc_norm": 0.2363481228668942, "acc_norm_stderr": 0.012414960524301832 }, "harness|ko_hellaswag|10": { "acc": 0.2703644692292372, "acc_stderr": 0.004432403734882273, "acc_norm": 0.2969527982473611, "acc_norm_stderr": 0.004559817589182076 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.22807017543859648, "acc_stderr": 0.03218093795602357, "acc_norm": 0.22807017543859648, "acc_norm_stderr": 0.03218093795602357 }, "harness|ko_mmlu_management|5": { "acc": 0.1650485436893204, "acc_stderr": 0.03675668832233188, "acc_norm": 0.1650485436893204, "acc_norm_stderr": 0.03675668832233188 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.27586206896551724, "acc_stderr": 0.01598281477469563, "acc_norm": 0.27586206896551724, "acc_norm_stderr": 0.01598281477469563 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.3037037037037037, "acc_stderr": 0.03972552884785137, "acc_norm": 0.3037037037037037, "acc_norm_stderr": 0.03972552884785137 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.28, "acc_stderr": 0.045126085985421255, "acc_norm": 0.28, "acc_norm_stderr": 0.045126085985421255 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.32340425531914896, "acc_stderr": 0.030579442773610337, "acc_norm": 0.32340425531914896, "acc_norm_stderr": 0.030579442773610337 }, "harness|ko_mmlu_virology|5": { "acc": 0.28313253012048195, "acc_stderr": 0.03507295431370519, "acc_norm": 0.28313253012048195, "acc_norm_stderr": 0.03507295431370519 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.2990353697749196, "acc_stderr": 0.026003301117885142, "acc_norm": 0.2990353697749196, "acc_norm_stderr": 0.026003301117885142 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.29596412556053814, "acc_stderr": 0.030636591348699817, "acc_norm": 0.29596412556053814, "acc_norm_stderr": 0.030636591348699817 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.24427480916030533, "acc_stderr": 0.03768335959728744, "acc_norm": 0.24427480916030533, "acc_norm_stderr": 0.03768335959728744 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.24, "acc_stderr": 0.04292346959909282, "acc_norm": 0.24, "acc_norm_stderr": 0.04292346959909282 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.20202020202020202, "acc_stderr": 0.028606204289229872, "acc_norm": 0.20202020202020202, "acc_norm_stderr": 0.028606204289229872 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.2413793103448276, "acc_stderr": 0.03565998174135303, "acc_norm": 0.2413793103448276, "acc_norm_stderr": 0.03565998174135303 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.16666666666666666, "acc_stderr": 0.03708284662416542, "acc_norm": 0.16666666666666666, "acc_norm_stderr": 0.03708284662416542 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.2773109243697479, "acc_stderr": 0.02907937453948001, "acc_norm": 0.2773109243697479, "acc_norm_stderr": 0.02907937453948001 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.2743589743589744, "acc_stderr": 0.02262276576749322, "acc_norm": 0.2743589743589744, "acc_norm_stderr": 0.02262276576749322 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.28, "acc_stderr": 0.045126085985421296, "acc_norm": 0.28, "acc_norm_stderr": 0.045126085985421296 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.2, "acc_stderr": 0.040201512610368445, "acc_norm": 0.2, "acc_norm_stderr": 0.040201512610368445 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.2037037037037037, "acc_stderr": 0.038935425188248475, "acc_norm": 0.2037037037037037, "acc_norm_stderr": 0.038935425188248475 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.29064039408866993, "acc_stderr": 0.031947400722655395, "acc_norm": 0.29064039408866993, "acc_norm_stderr": 0.031947400722655395 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.26129032258064516, "acc_stderr": 0.02499305339776482, "acc_norm": 0.26129032258064516, "acc_norm_stderr": 0.02499305339776482 }, "harness|ko_mmlu_marketing|5": { "acc": 0.32905982905982906, "acc_stderr": 0.030782321577688163, "acc_norm": 0.32905982905982906, "acc_norm_stderr": 0.030782321577688163 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.22264150943396227, "acc_stderr": 0.025604233470899105, "acc_norm": 0.22264150943396227, "acc_norm_stderr": 0.025604233470899105 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.21818181818181817, "acc_stderr": 0.03955932861795833, "acc_norm": 0.21818181818181817, "acc_norm_stderr": 0.03955932861795833 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.29259259259259257, "acc_stderr": 0.027738969632176088, "acc_norm": 0.29259259259259257, "acc_norm_stderr": 0.027738969632176088 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2582781456953642, "acc_stderr": 0.035737053147634576, "acc_norm": 0.2582781456953642, "acc_norm_stderr": 0.035737053147634576 }, "harness|ko_mmlu_sociology|5": { "acc": 0.24875621890547264, "acc_stderr": 0.030567675938916718, "acc_norm": 0.24875621890547264, "acc_norm_stderr": 0.030567675938916718 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.17341040462427745, "acc_stderr": 0.02886810787497064, "acc_norm": 0.17341040462427745, "acc_norm_stderr": 0.02886810787497064 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2275132275132275, "acc_stderr": 0.021591269407823774, "acc_norm": 0.2275132275132275, "acc_norm_stderr": 0.021591269407823774 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2361111111111111, "acc_stderr": 0.03551446610810826, "acc_norm": 0.2361111111111111, "acc_norm_stderr": 0.03551446610810826 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.33, "acc_stderr": 0.04725815626252606, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252606 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.22, "acc_stderr": 0.04163331998932269, "acc_norm": 0.22, "acc_norm_stderr": 0.04163331998932269 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.2138728323699422, "acc_stderr": 0.022075709251757177, "acc_norm": 0.2138728323699422, "acc_norm_stderr": 0.022075709251757177 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.27607361963190186, "acc_stderr": 0.0351238528370505, "acc_norm": 0.27607361963190186, "acc_norm_stderr": 0.0351238528370505 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.25925925925925924, "acc_stderr": 0.024383665531035457, "acc_norm": 0.25925925925925924, "acc_norm_stderr": 0.024383665531035457 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.26, "acc_stderr": 0.0440844002276808, "acc_norm": 0.26, "acc_norm_stderr": 0.0440844002276808 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.19689119170984457, "acc_stderr": 0.028697873971860695, "acc_norm": 0.19689119170984457, "acc_norm_stderr": 0.028697873971860695 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2543859649122807, "acc_stderr": 0.040969851398436716, "acc_norm": 0.2543859649122807, "acc_norm_stderr": 0.040969851398436716 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.24954128440366974, "acc_stderr": 0.018553897629501614, "acc_norm": 0.24954128440366974, "acc_norm_stderr": 0.018553897629501614 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.12698412698412698, "acc_stderr": 0.029780417522688434, "acc_norm": 0.12698412698412698, "acc_norm_stderr": 0.029780417522688434 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.22549019607843138, "acc_stderr": 0.023929155517351294, "acc_norm": 0.22549019607843138, "acc_norm_stderr": 0.023929155517351294 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.18, "acc_stderr": 0.038612291966536955, "acc_norm": 0.18, "acc_norm_stderr": 0.038612291966536955 }, "harness|ko_mmlu_international_law|5": { "acc": 0.2809917355371901, "acc_stderr": 0.04103203830514512, "acc_norm": 0.2809917355371901, "acc_norm_stderr": 0.04103203830514512 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.19078947368421054, "acc_stderr": 0.03197565821032501, "acc_norm": 0.19078947368421054, "acc_norm_stderr": 0.03197565821032501 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.2369281045751634, "acc_stderr": 0.017201662169789796, "acc_norm": 0.2369281045751634, "acc_norm_stderr": 0.017201662169789796 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.2624113475177305, "acc_stderr": 0.02624492034984301, "acc_norm": 0.2624113475177305, "acc_norm_stderr": 0.02624492034984301 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.21428571428571427, "acc_stderr": 0.03894641120044792, "acc_norm": 0.21428571428571427, "acc_norm_stderr": 0.03894641120044792 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.44907407407407407, "acc_stderr": 0.03392238405321617, "acc_norm": 0.44907407407407407, "acc_norm_stderr": 0.03392238405321617 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2424581005586592, "acc_stderr": 0.01433352205921789, "acc_norm": 0.2424581005586592, "acc_norm_stderr": 0.01433352205921789 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.34, "acc_stderr": 0.04760952285695235, "acc_norm": 0.34, "acc_norm_stderr": 0.04760952285695235 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.44485294117647056, "acc_stderr": 0.03018753206032938, "acc_norm": 0.44485294117647056, "acc_norm_stderr": 0.03018753206032938 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.2612244897959184, "acc_stderr": 0.028123429335142773, "acc_norm": 0.2612244897959184, "acc_norm_stderr": 0.028123429335142773 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.24050632911392406, "acc_stderr": 0.027820781981149685, "acc_norm": 0.24050632911392406, "acc_norm_stderr": 0.027820781981149685 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.227509778357236, "acc_stderr": 0.010707188576864226, "acc_norm": 0.227509778357236, "acc_norm_stderr": 0.010707188576864226 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.24509803921568626, "acc_stderr": 0.030190282453501936, "acc_norm": 0.24509803921568626, "acc_norm_stderr": 0.030190282453501936 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.21818181818181817, "acc_stderr": 0.03225078108306289, "acc_norm": 0.21818181818181817, "acc_norm_stderr": 0.03225078108306289 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2668298653610771, "mc1_stderr": 0.015483691939237269, "mc2": 0.42019223039185516, "mc2_stderr": 0.01650268606738961 }, "harness|ko_commongen_v2|2": { "acc": 0.21959858323494688, "acc_stderr": 0.01423274308558026, "acc_norm": 0.31641086186540734, "acc_norm_stderr": 0.015989617951065474 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "leo911kim/Exodia-7B", "model_sha": "b3e1f98b934da7498bb18ce0cb9e0fc857593656", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }