results
dict
versions
dict
config_general
dict
{ "harness|ko_arc_challenge|25": { "acc": 0.2235494880546075, "acc_stderr": 0.012174896631202605, "acc_norm": 0.2815699658703072, "acc_norm_stderr": 0.013143376735009015 }, "harness|ko_hellaswag|10": { "acc": 0.3345947022505477, "acc_stderr": 0.004708842600177431, "acc_norm": 0.4135630352519418, "acc_norm_stderr": 0.0049146550633294974 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.27485380116959063, "acc_stderr": 0.03424042924691585, "acc_norm": 0.27485380116959063, "acc_norm_stderr": 0.03424042924691585 }, "harness|ko_mmlu_management|5": { "acc": 0.27184466019417475, "acc_stderr": 0.044052680241409216, "acc_norm": 0.27184466019417475, "acc_norm_stderr": 0.044052680241409216 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.26947637292464877, "acc_stderr": 0.015866243073215065, "acc_norm": 0.26947637292464877, "acc_norm_stderr": 0.015866243073215065 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.26666666666666666, "acc_stderr": 0.038201699145179055, "acc_norm": 0.26666666666666666, "acc_norm_stderr": 0.038201699145179055 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.2127659574468085, "acc_stderr": 0.026754391348039783, "acc_norm": 0.2127659574468085, "acc_norm_stderr": 0.026754391348039783 }, "harness|ko_mmlu_virology|5": { "acc": 0.24096385542168675, "acc_stderr": 0.033293941190735296, "acc_norm": 0.24096385542168675, "acc_norm_stderr": 0.033293941190735296 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.2379421221864952, "acc_stderr": 0.024185150647818707, "acc_norm": 0.2379421221864952, "acc_norm_stderr": 0.024185150647818707 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.2825112107623318, "acc_stderr": 0.030216831011508766, "acc_norm": 0.2825112107623318, "acc_norm_stderr": 0.030216831011508766 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.21374045801526717, "acc_stderr": 0.0359546161177469, "acc_norm": 0.21374045801526717, "acc_norm_stderr": 0.0359546161177469 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.24, "acc_stderr": 0.042923469599092816, "acc_norm": 0.24, "acc_norm_stderr": 0.042923469599092816 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.2474747474747475, "acc_stderr": 0.03074630074212451, "acc_norm": 0.2474747474747475, "acc_norm_stderr": 0.03074630074212451 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.22758620689655173, "acc_stderr": 0.03493950380131184, "acc_norm": 0.22758620689655173, "acc_norm_stderr": 0.03493950380131184 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.22549019607843138, "acc_stderr": 0.041583075330832865, "acc_norm": 0.22549019607843138, "acc_norm_stderr": 0.041583075330832865 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.31512605042016806, "acc_stderr": 0.030176808288974337, "acc_norm": 0.31512605042016806, "acc_norm_stderr": 0.030176808288974337 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.2205128205128205, "acc_stderr": 0.02102067268082791, "acc_norm": 0.2205128205128205, "acc_norm_stderr": 0.02102067268082791 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.18, "acc_stderr": 0.038612291966536955, "acc_norm": 0.18, "acc_norm_stderr": 0.038612291966536955 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.25, "acc_stderr": 0.04186091791394607, "acc_norm": 0.25, "acc_norm_stderr": 0.04186091791394607 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.2660098522167488, "acc_stderr": 0.03108982600293752, "acc_norm": 0.2660098522167488, "acc_norm_stderr": 0.03108982600293752 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.3, "acc_stderr": 0.02606936229533513, "acc_norm": 0.3, "acc_norm_stderr": 0.02606936229533513 }, "harness|ko_mmlu_marketing|5": { "acc": 0.23076923076923078, "acc_stderr": 0.027601921381417607, "acc_norm": 0.23076923076923078, "acc_norm_stderr": 0.027601921381417607 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.25660377358490566, "acc_stderr": 0.026880647889051968, "acc_norm": 0.25660377358490566, "acc_norm_stderr": 0.026880647889051968 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.2545454545454545, "acc_stderr": 0.04172343038705383, "acc_norm": 0.2545454545454545, "acc_norm_stderr": 0.04172343038705383 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.2962962962962963, "acc_stderr": 0.02784081149587194, "acc_norm": 0.2962962962962963, "acc_norm_stderr": 0.02784081149587194 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.304635761589404, "acc_stderr": 0.03757949922943342, "acc_norm": 0.304635761589404, "acc_norm_stderr": 0.03757949922943342 }, "harness|ko_mmlu_sociology|5": { "acc": 0.25870646766169153, "acc_stderr": 0.03096590312357303, "acc_norm": 0.25870646766169153, "acc_norm_stderr": 0.03096590312357303 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.2254335260115607, "acc_stderr": 0.03186209851641144, "acc_norm": 0.2254335260115607, "acc_norm_stderr": 0.03186209851641144 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2566137566137566, "acc_stderr": 0.022494510767503154, "acc_norm": 0.2566137566137566, "acc_norm_stderr": 0.022494510767503154 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2638888888888889, "acc_stderr": 0.03685651095897532, "acc_norm": 0.2638888888888889, "acc_norm_stderr": 0.03685651095897532 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.23, "acc_stderr": 0.04229525846816505, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816505 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.22, "acc_stderr": 0.04163331998932269, "acc_norm": 0.22, "acc_norm_stderr": 0.04163331998932269 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.24855491329479767, "acc_stderr": 0.023267528432100174, "acc_norm": 0.24855491329479767, "acc_norm_stderr": 0.023267528432100174 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.31901840490797545, "acc_stderr": 0.03661997551073836, "acc_norm": 0.31901840490797545, "acc_norm_stderr": 0.03661997551073836 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.2623456790123457, "acc_stderr": 0.024477222856135114, "acc_norm": 0.2623456790123457, "acc_norm_stderr": 0.024477222856135114 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.33678756476683935, "acc_stderr": 0.03410780251836184, "acc_norm": 0.33678756476683935, "acc_norm_stderr": 0.03410780251836184 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.20175438596491227, "acc_stderr": 0.037752050135836386, "acc_norm": 0.20175438596491227, "acc_norm_stderr": 0.037752050135836386 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.24220183486238533, "acc_stderr": 0.01836817630659862, "acc_norm": 0.24220183486238533, "acc_norm_stderr": 0.01836817630659862 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.23015873015873015, "acc_stderr": 0.03764950879790606, "acc_norm": 0.23015873015873015, "acc_norm_stderr": 0.03764950879790606 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.23529411764705882, "acc_stderr": 0.024288619466046102, "acc_norm": 0.23529411764705882, "acc_norm_stderr": 0.024288619466046102 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.18, "acc_stderr": 0.03861229196653695, "acc_norm": 0.18, "acc_norm_stderr": 0.03861229196653695 }, "harness|ko_mmlu_international_law|5": { "acc": 0.256198347107438, "acc_stderr": 0.039849796533028704, "acc_norm": 0.256198347107438, "acc_norm_stderr": 0.039849796533028704 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.21710526315789475, "acc_stderr": 0.033550453048829226, "acc_norm": 0.21710526315789475, "acc_norm_stderr": 0.033550453048829226 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.24019607843137256, "acc_stderr": 0.01728276069516743, "acc_norm": 0.24019607843137256, "acc_norm_stderr": 0.01728276069516743 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.2553191489361702, "acc_stderr": 0.02601199293090201, "acc_norm": 0.2553191489361702, "acc_norm_stderr": 0.02601199293090201 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.21428571428571427, "acc_stderr": 0.03894641120044793, "acc_norm": 0.21428571428571427, "acc_norm_stderr": 0.03894641120044793 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.46296296296296297, "acc_stderr": 0.03400603625538272, "acc_norm": 0.46296296296296297, "acc_norm_stderr": 0.03400603625538272 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24692737430167597, "acc_stderr": 0.014422292204808852, "acc_norm": 0.24692737430167597, "acc_norm_stderr": 0.014422292204808852 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4411764705882353, "acc_stderr": 0.030161911930767102, "acc_norm": 0.4411764705882353, "acc_norm_stderr": 0.030161911930767102 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.3795918367346939, "acc_stderr": 0.03106721126287249, "acc_norm": 0.3795918367346939, "acc_norm_stderr": 0.03106721126287249 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.2109704641350211, "acc_stderr": 0.02655837250266192, "acc_norm": 0.2109704641350211, "acc_norm_stderr": 0.02655837250266192 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.23468057366362452, "acc_stderr": 0.010824026872449344, "acc_norm": 0.23468057366362452, "acc_norm_stderr": 0.010824026872449344 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.25, "acc_stderr": 0.03039153369274154, "acc_norm": 0.25, "acc_norm_stderr": 0.03039153369274154 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.22424242424242424, "acc_stderr": 0.03256866661681102, "acc_norm": 0.22424242424242424, "acc_norm_stderr": 0.03256866661681102 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.25091799265605874, "mc1_stderr": 0.015176985027707682, "mc2": 0.4116568832959107, "mc2_stderr": 0.015044504977529799 }, "harness|ko_commongen_v2|2": { "acc": 0.27744982290436837, "acc_stderr": 0.015393630236605975, "acc_norm": 0.3400236127508855, "acc_norm_stderr": 0.016286717220737674 } }
{ "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }
{ "model_name": "EleutherAI/polyglot-ko-1.3b", "model_sha": "557e162cf6e944fdbae05bab2e45d066a125eacb", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null }
{ "harness|ko_arc_challenge|25": { "acc": 0.2858361774744027, "acc_stderr": 0.013203196088537365, "acc_norm": 0.33532423208191126, "acc_norm_stderr": 0.013796182947785562 }, "harness|ko_hellaswag|10": { "acc": 0.385381398127863, "acc_stderr": 0.004856906473719383, "acc_norm": 0.5027882891854212, "acc_norm_stderr": 0.004989703824167094 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.30994152046783624, "acc_stderr": 0.03546976959393161, "acc_norm": 0.30994152046783624, "acc_norm_stderr": 0.03546976959393161 }, "harness|ko_mmlu_management|5": { "acc": 0.18446601941747573, "acc_stderr": 0.03840423627288276, "acc_norm": 0.18446601941747573, "acc_norm_stderr": 0.03840423627288276 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.25925925925925924, "acc_stderr": 0.015671006009339572, "acc_norm": 0.25925925925925924, "acc_norm_stderr": 0.015671006009339572 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.22962962962962963, "acc_stderr": 0.036333844140734636, "acc_norm": 0.22962962962962963, "acc_norm_stderr": 0.036333844140734636 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.27, "acc_stderr": 0.0446196043338474, "acc_norm": 0.27, "acc_norm_stderr": 0.0446196043338474 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.2127659574468085, "acc_stderr": 0.026754391348039787, "acc_norm": 0.2127659574468085, "acc_norm_stderr": 0.026754391348039787 }, "harness|ko_mmlu_virology|5": { "acc": 0.21686746987951808, "acc_stderr": 0.03208284450356365, "acc_norm": 0.21686746987951808, "acc_norm_stderr": 0.03208284450356365 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.31189710610932475, "acc_stderr": 0.02631185807185416, "acc_norm": 0.31189710610932475, "acc_norm_stderr": 0.02631185807185416 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.20179372197309417, "acc_stderr": 0.02693611191280227, "acc_norm": 0.20179372197309417, "acc_norm_stderr": 0.02693611191280227 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.22900763358778625, "acc_stderr": 0.036853466317118506, "acc_norm": 0.22900763358778625, "acc_norm_stderr": 0.036853466317118506 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.23, "acc_stderr": 0.04229525846816506, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816506 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.24242424242424243, "acc_stderr": 0.03053289223393203, "acc_norm": 0.24242424242424243, "acc_norm_stderr": 0.03053289223393203 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.2896551724137931, "acc_stderr": 0.03780019230438014, "acc_norm": 0.2896551724137931, "acc_norm_stderr": 0.03780019230438014 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.3137254901960784, "acc_stderr": 0.04617034827006716, "acc_norm": 0.3137254901960784, "acc_norm_stderr": 0.04617034827006716 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.23109243697478993, "acc_stderr": 0.027381406927868963, "acc_norm": 0.23109243697478993, "acc_norm_stderr": 0.027381406927868963 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.21025641025641026, "acc_stderr": 0.020660597485026928, "acc_norm": 0.21025641025641026, "acc_norm_stderr": 0.020660597485026928 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.26, "acc_stderr": 0.0440844002276808, "acc_norm": 0.26, "acc_norm_stderr": 0.0440844002276808 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.21, "acc_stderr": 0.040936018074033256, "acc_norm": 0.21, "acc_norm_stderr": 0.040936018074033256 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.25, "acc_stderr": 0.04186091791394607, "acc_norm": 0.25, "acc_norm_stderr": 0.04186091791394607 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.26108374384236455, "acc_stderr": 0.0309037969521145, "acc_norm": 0.26108374384236455, "acc_norm_stderr": 0.0309037969521145 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.25161290322580643, "acc_stderr": 0.024685979286239963, "acc_norm": 0.25161290322580643, "acc_norm_stderr": 0.024685979286239963 }, "harness|ko_mmlu_marketing|5": { "acc": 0.23076923076923078, "acc_stderr": 0.027601921381417604, "acc_norm": 0.23076923076923078, "acc_norm_stderr": 0.027601921381417604 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.23773584905660378, "acc_stderr": 0.026199808807561932, "acc_norm": 0.23773584905660378, "acc_norm_stderr": 0.026199808807561932 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.21818181818181817, "acc_stderr": 0.03955932861795833, "acc_norm": 0.21818181818181817, "acc_norm_stderr": 0.03955932861795833 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.26296296296296295, "acc_stderr": 0.02684205787383371, "acc_norm": 0.26296296296296295, "acc_norm_stderr": 0.02684205787383371 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2847682119205298, "acc_stderr": 0.03684881521389024, "acc_norm": 0.2847682119205298, "acc_norm_stderr": 0.03684881521389024 }, "harness|ko_mmlu_sociology|5": { "acc": 0.263681592039801, "acc_stderr": 0.03115715086935554, "acc_norm": 0.263681592039801, "acc_norm_stderr": 0.03115715086935554 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.24855491329479767, "acc_stderr": 0.03295304696818317, "acc_norm": 0.24855491329479767, "acc_norm_stderr": 0.03295304696818317 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2671957671957672, "acc_stderr": 0.022789673145776578, "acc_norm": 0.2671957671957672, "acc_norm_stderr": 0.022789673145776578 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2569444444444444, "acc_stderr": 0.036539469694421, "acc_norm": 0.2569444444444444, "acc_norm_stderr": 0.036539469694421 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.2, "acc_stderr": 0.04020151261036846, "acc_norm": 0.2, "acc_norm_stderr": 0.04020151261036846 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.26, "acc_stderr": 0.04408440022768079, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768079 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.23121387283236994, "acc_stderr": 0.022698657167855716, "acc_norm": 0.23121387283236994, "acc_norm_stderr": 0.022698657167855716 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.2883435582822086, "acc_stderr": 0.035590395316173425, "acc_norm": 0.2883435582822086, "acc_norm_stderr": 0.035590395316173425 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.2777777777777778, "acc_stderr": 0.024922001168886338, "acc_norm": 0.2777777777777778, "acc_norm_stderr": 0.024922001168886338 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.33, "acc_stderr": 0.047258156262526045, "acc_norm": 0.33, "acc_norm_stderr": 0.047258156262526045 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.2694300518134715, "acc_stderr": 0.03201867122877794, "acc_norm": 0.2694300518134715, "acc_norm_stderr": 0.03201867122877794 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.23684210526315788, "acc_stderr": 0.03999423879281336, "acc_norm": 0.23684210526315788, "acc_norm_stderr": 0.03999423879281336 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.26422018348623855, "acc_stderr": 0.0189041641715102, "acc_norm": 0.26422018348623855, "acc_norm_stderr": 0.0189041641715102 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.19047619047619047, "acc_stderr": 0.035122074123020534, "acc_norm": 0.19047619047619047, "acc_norm_stderr": 0.035122074123020534 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.2581699346405229, "acc_stderr": 0.02505850331695815, "acc_norm": 0.2581699346405229, "acc_norm_stderr": 0.02505850331695815 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.22, "acc_stderr": 0.041633319989322674, "acc_norm": 0.22, "acc_norm_stderr": 0.041633319989322674 }, "harness|ko_mmlu_international_law|5": { "acc": 0.36363636363636365, "acc_stderr": 0.043913262867240704, "acc_norm": 0.36363636363636365, "acc_norm_stderr": 0.043913262867240704 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.28289473684210525, "acc_stderr": 0.03665349695640767, "acc_norm": 0.28289473684210525, "acc_norm_stderr": 0.03665349695640767 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.20098039215686275, "acc_stderr": 0.016211938889655574, "acc_norm": 0.20098039215686275, "acc_norm_stderr": 0.016211938889655574 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.22695035460992907, "acc_stderr": 0.02498710636564298, "acc_norm": 0.22695035460992907, "acc_norm_stderr": 0.02498710636564298 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.25, "acc_stderr": 0.04109974682633932, "acc_norm": 0.25, "acc_norm_stderr": 0.04109974682633932 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.3611111111111111, "acc_stderr": 0.03275773486100999, "acc_norm": 0.3611111111111111, "acc_norm_stderr": 0.03275773486100999 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24692737430167597, "acc_stderr": 0.014422292204808852, "acc_norm": 0.24692737430167597, "acc_norm_stderr": 0.014422292204808852 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.27, "acc_stderr": 0.04461960433384741, "acc_norm": 0.27, "acc_norm_stderr": 0.04461960433384741 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.44485294117647056, "acc_stderr": 0.030187532060329383, "acc_norm": 0.44485294117647056, "acc_norm_stderr": 0.030187532060329383 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.22857142857142856, "acc_stderr": 0.026882144922307748, "acc_norm": 0.22857142857142856, "acc_norm_stderr": 0.026882144922307748 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.32489451476793246, "acc_stderr": 0.030486039389105303, "acc_norm": 0.32489451476793246, "acc_norm_stderr": 0.030486039389105303 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.25684485006518903, "acc_stderr": 0.011158455853098857, "acc_norm": 0.25684485006518903, "acc_norm_stderr": 0.011158455853098857 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.27941176470588236, "acc_stderr": 0.031493281045079556, "acc_norm": 0.27941176470588236, "acc_norm_stderr": 0.031493281045079556 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.23030303030303031, "acc_stderr": 0.032876667586034886, "acc_norm": 0.23030303030303031, "acc_norm_stderr": 0.032876667586034886 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2350061199510404, "mc1_stderr": 0.014843061507731613, "mc2": 0.390667104295536, "mc2_stderr": 0.014736649975849761 }, "harness|ko_commongen_v2|2": { "acc": 0.30932703659976385, "acc_stderr": 0.01589132050552089, "acc_norm": 0.3990554899645809, "acc_norm_stderr": 0.0168363772928493 } }
{ "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }
{ "model_name": "EleutherAI/polyglot-ko-12.8b", "model_sha": "09dfc839067bf44e7f52976eca8adbc17f04e1b0", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null }
{ "harness|ko_arc_challenge|25": { "acc": 0.2525597269624573, "acc_stderr": 0.01269672898020771, "acc_norm": 0.3046075085324232, "acc_norm_stderr": 0.013449522109932494 }, "harness|ko_hellaswag|10": { "acc": 0.3511252738498307, "acc_stderr": 0.004763465139038552, "acc_norm": 0.4420434176458873, "acc_norm_stderr": 0.004956147046108961 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.2222222222222222, "acc_stderr": 0.03188578017686398, "acc_norm": 0.2222222222222222, "acc_norm_stderr": 0.03188578017686398 }, "harness|ko_mmlu_management|5": { "acc": 0.30097087378640774, "acc_stderr": 0.045416094465039476, "acc_norm": 0.30097087378640774, "acc_norm_stderr": 0.045416094465039476 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.23627075351213284, "acc_stderr": 0.015190473717037497, "acc_norm": 0.23627075351213284, "acc_norm_stderr": 0.015190473717037497 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.23703703703703705, "acc_stderr": 0.03673731683969506, "acc_norm": 0.23703703703703705, "acc_norm_stderr": 0.03673731683969506 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.27, "acc_stderr": 0.04461960433384741, "acc_norm": 0.27, "acc_norm_stderr": 0.04461960433384741 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.2, "acc_stderr": 0.026148818018424502, "acc_norm": 0.2, "acc_norm_stderr": 0.026148818018424502 }, "harness|ko_mmlu_virology|5": { "acc": 0.25301204819277107, "acc_stderr": 0.03384429155233135, "acc_norm": 0.25301204819277107, "acc_norm_stderr": 0.03384429155233135 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.24437299035369775, "acc_stderr": 0.024406162094668886, "acc_norm": 0.24437299035369775, "acc_norm_stderr": 0.024406162094668886 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.21524663677130046, "acc_stderr": 0.027584066602208263, "acc_norm": 0.21524663677130046, "acc_norm_stderr": 0.027584066602208263 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.19083969465648856, "acc_stderr": 0.034465133507525954, "acc_norm": 0.19083969465648856, "acc_norm_stderr": 0.034465133507525954 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.21, "acc_stderr": 0.040936018074033256, "acc_norm": 0.21, "acc_norm_stderr": 0.040936018074033256 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.31313131313131315, "acc_stderr": 0.033042050878136525, "acc_norm": 0.31313131313131315, "acc_norm_stderr": 0.033042050878136525 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.2413793103448276, "acc_stderr": 0.03565998174135302, "acc_norm": 0.2413793103448276, "acc_norm_stderr": 0.03565998174135302 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.19607843137254902, "acc_stderr": 0.03950581861179962, "acc_norm": 0.19607843137254902, "acc_norm_stderr": 0.03950581861179962 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.3487394957983193, "acc_stderr": 0.030956636328566548, "acc_norm": 0.3487394957983193, "acc_norm_stderr": 0.030956636328566548 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.3564102564102564, "acc_stderr": 0.024283140529467295, "acc_norm": 0.3564102564102564, "acc_norm_stderr": 0.024283140529467295 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.16, "acc_stderr": 0.03684529491774709, "acc_norm": 0.16, "acc_norm_stderr": 0.03684529491774709 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.32, "acc_stderr": 0.04688261722621504, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621504 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.21296296296296297, "acc_stderr": 0.03957835471980981, "acc_norm": 0.21296296296296297, "acc_norm_stderr": 0.03957835471980981 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.29064039408866993, "acc_stderr": 0.03194740072265541, "acc_norm": 0.29064039408866993, "acc_norm_stderr": 0.03194740072265541 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.3032258064516129, "acc_stderr": 0.026148685930671742, "acc_norm": 0.3032258064516129, "acc_norm_stderr": 0.026148685930671742 }, "harness|ko_mmlu_marketing|5": { "acc": 0.19658119658119658, "acc_stderr": 0.02603538609895129, "acc_norm": 0.19658119658119658, "acc_norm_stderr": 0.02603538609895129 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.3283018867924528, "acc_stderr": 0.02890159361241178, "acc_norm": 0.3283018867924528, "acc_norm_stderr": 0.02890159361241178 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.23636363636363636, "acc_stderr": 0.04069306319721376, "acc_norm": 0.23636363636363636, "acc_norm_stderr": 0.04069306319721376 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.25925925925925924, "acc_stderr": 0.02671924078371216, "acc_norm": 0.25925925925925924, "acc_norm_stderr": 0.02671924078371216 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.32450331125827814, "acc_stderr": 0.03822746937658753, "acc_norm": 0.32450331125827814, "acc_norm_stderr": 0.03822746937658753 }, "harness|ko_mmlu_sociology|5": { "acc": 0.26865671641791045, "acc_stderr": 0.03134328358208954, "acc_norm": 0.26865671641791045, "acc_norm_stderr": 0.03134328358208954 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.32947976878612717, "acc_stderr": 0.03583901754736412, "acc_norm": 0.32947976878612717, "acc_norm_stderr": 0.03583901754736412 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.24603174603174602, "acc_stderr": 0.022182037202948368, "acc_norm": 0.24603174603174602, "acc_norm_stderr": 0.022182037202948368 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2569444444444444, "acc_stderr": 0.03653946969442099, "acc_norm": 0.2569444444444444, "acc_norm_stderr": 0.03653946969442099 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.4, "acc_stderr": 0.049236596391733084, "acc_norm": 0.4, "acc_norm_stderr": 0.049236596391733084 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.26, "acc_stderr": 0.044084400227680794, "acc_norm": 0.26, "acc_norm_stderr": 0.044084400227680794 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.2398843930635838, "acc_stderr": 0.022989592543123567, "acc_norm": 0.2398843930635838, "acc_norm_stderr": 0.022989592543123567 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.3312883435582822, "acc_stderr": 0.03697983910025588, "acc_norm": 0.3312883435582822, "acc_norm_stderr": 0.03697983910025588 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.26851851851851855, "acc_stderr": 0.024659685185967277, "acc_norm": 0.26851851851851855, "acc_norm_stderr": 0.024659685185967277 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.35, "acc_stderr": 0.0479372485441102, "acc_norm": 0.35, "acc_norm_stderr": 0.0479372485441102 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.3626943005181347, "acc_stderr": 0.034697137917043715, "acc_norm": 0.3626943005181347, "acc_norm_stderr": 0.034697137917043715 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.22807017543859648, "acc_stderr": 0.03947152782669415, "acc_norm": 0.22807017543859648, "acc_norm_stderr": 0.03947152782669415 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.25688073394495414, "acc_stderr": 0.018732492928342448, "acc_norm": 0.25688073394495414, "acc_norm_stderr": 0.018732492928342448 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.3492063492063492, "acc_stderr": 0.04263906892795132, "acc_norm": 0.3492063492063492, "acc_norm_stderr": 0.04263906892795132 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.2549019607843137, "acc_stderr": 0.024954184324879905, "acc_norm": 0.2549019607843137, "acc_norm_stderr": 0.024954184324879905 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.26, "acc_stderr": 0.04408440022768079, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768079 }, "harness|ko_mmlu_international_law|5": { "acc": 0.1652892561983471, "acc_stderr": 0.03390780612972776, "acc_norm": 0.1652892561983471, "acc_norm_stderr": 0.03390780612972776 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.29605263157894735, "acc_stderr": 0.03715062154998904, "acc_norm": 0.29605263157894735, "acc_norm_stderr": 0.03715062154998904 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.21895424836601307, "acc_stderr": 0.016729937565537537, "acc_norm": 0.21895424836601307, "acc_norm_stderr": 0.016729937565537537 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.24822695035460993, "acc_stderr": 0.025770015644290396, "acc_norm": 0.24822695035460993, "acc_norm_stderr": 0.025770015644290396 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.15178571428571427, "acc_stderr": 0.03405702838185692, "acc_norm": 0.15178571428571427, "acc_norm_stderr": 0.03405702838185692 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4722222222222222, "acc_stderr": 0.0340470532865388, "acc_norm": 0.4722222222222222, "acc_norm_stderr": 0.0340470532865388 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.26145251396648045, "acc_stderr": 0.014696599650364546, "acc_norm": 0.26145251396648045, "acc_norm_stderr": 0.014696599650364546 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.26, "acc_stderr": 0.04408440022768078, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768078 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.22, "acc_stderr": 0.04163331998932269, "acc_norm": 0.22, "acc_norm_stderr": 0.04163331998932269 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.44485294117647056, "acc_stderr": 0.030187532060329383, "acc_norm": 0.44485294117647056, "acc_norm_stderr": 0.030187532060329383 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.39591836734693875, "acc_stderr": 0.03130802899065685, "acc_norm": 0.39591836734693875, "acc_norm_stderr": 0.03130802899065685 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.23628691983122363, "acc_stderr": 0.02765215314415926, "acc_norm": 0.23628691983122363, "acc_norm_stderr": 0.02765215314415926 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.24445893089960888, "acc_stderr": 0.010976425013113912, "acc_norm": 0.24445893089960888, "acc_norm_stderr": 0.010976425013113912 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.2647058823529412, "acc_stderr": 0.03096451792692341, "acc_norm": 0.2647058823529412, "acc_norm_stderr": 0.03096451792692341 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.296969696969697, "acc_stderr": 0.03567969772268046, "acc_norm": 0.296969696969697, "acc_norm_stderr": 0.03567969772268046 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.24969400244798043, "mc1_stderr": 0.015152286907148125, "mc2": 0.40454723614569765, "mc2_stderr": 0.014981033793701278 }, "harness|ko_commongen_v2|2": { "acc": 0.282172373081464, "acc_stderr": 0.015473271583988433, "acc_norm": 0.3707201889020071, "acc_norm_stderr": 0.016605801289212605 } }
{ "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }
{ "model_name": "EleutherAI/polyglot-ko-3.8b", "model_sha": "3c696a71c16b4a4622b7cabf6c5da4ba5a73b548", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null }
{ "harness|ko_arc_challenge|25": { "acc": 0.2687713310580205, "acc_stderr": 0.012955065963710675, "acc_norm": 0.32764505119453924, "acc_norm_stderr": 0.013715847940719339 }, "harness|ko_hellaswag|10": { "acc": 0.3690499900418243, "acc_stderr": 0.004815613144385398, "acc_norm": 0.4814777932682733, "acc_norm_stderr": 0.004986356526063965 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.17543859649122806, "acc_stderr": 0.029170885500727665, "acc_norm": 0.17543859649122806, "acc_norm_stderr": 0.029170885500727665 }, "harness|ko_mmlu_management|5": { "acc": 0.3592233009708738, "acc_stderr": 0.04750458399041693, "acc_norm": 0.3592233009708738, "acc_norm_stderr": 0.04750458399041693 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.20178799489144317, "acc_stderr": 0.014351702181636861, "acc_norm": 0.20178799489144317, "acc_norm_stderr": 0.014351702181636861 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.22962962962962963, "acc_stderr": 0.036333844140734636, "acc_norm": 0.22962962962962963, "acc_norm_stderr": 0.036333844140734636 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.22, "acc_stderr": 0.0416333199893227, "acc_norm": 0.22, "acc_norm_stderr": 0.0416333199893227 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.17446808510638298, "acc_stderr": 0.02480944233550398, "acc_norm": 0.17446808510638298, "acc_norm_stderr": 0.02480944233550398 }, "harness|ko_mmlu_virology|5": { "acc": 0.1927710843373494, "acc_stderr": 0.030709824050565264, "acc_norm": 0.1927710843373494, "acc_norm_stderr": 0.030709824050565264 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.24115755627009647, "acc_stderr": 0.024296594034763426, "acc_norm": 0.24115755627009647, "acc_norm_stderr": 0.024296594034763426 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.11659192825112108, "acc_stderr": 0.021539639816244467, "acc_norm": 0.11659192825112108, "acc_norm_stderr": 0.021539639816244467 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.2748091603053435, "acc_stderr": 0.03915345408847835, "acc_norm": 0.2748091603053435, "acc_norm_stderr": 0.03915345408847835 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.24, "acc_stderr": 0.04292346959909284, "acc_norm": 0.24, "acc_norm_stderr": 0.04292346959909284 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.35353535353535354, "acc_stderr": 0.03406086723547153, "acc_norm": 0.35353535353535354, "acc_norm_stderr": 0.03406086723547153 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.2413793103448276, "acc_stderr": 0.03565998174135302, "acc_norm": 0.2413793103448276, "acc_norm_stderr": 0.03565998174135302 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.37254901960784315, "acc_stderr": 0.048108401480826346, "acc_norm": 0.37254901960784315, "acc_norm_stderr": 0.048108401480826346 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.3487394957983193, "acc_stderr": 0.030956636328566548, "acc_norm": 0.3487394957983193, "acc_norm_stderr": 0.030956636328566548 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.3641025641025641, "acc_stderr": 0.024396672985094778, "acc_norm": 0.3641025641025641, "acc_norm_stderr": 0.024396672985094778 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.18, "acc_stderr": 0.03861229196653694, "acc_norm": 0.18, "acc_norm_stderr": 0.03861229196653694 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.16, "acc_stderr": 0.03684529491774709, "acc_norm": 0.16, "acc_norm_stderr": 0.03684529491774709 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.21296296296296297, "acc_stderr": 0.03957835471980981, "acc_norm": 0.21296296296296297, "acc_norm_stderr": 0.03957835471980981 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.28078817733990147, "acc_stderr": 0.0316185633535861, "acc_norm": 0.28078817733990147, "acc_norm_stderr": 0.0316185633535861 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.3161290322580645, "acc_stderr": 0.026450874489042764, "acc_norm": 0.3161290322580645, "acc_norm_stderr": 0.026450874489042764 }, "harness|ko_mmlu_marketing|5": { "acc": 0.19658119658119658, "acc_stderr": 0.02603538609895129, "acc_norm": 0.19658119658119658, "acc_norm_stderr": 0.02603538609895129 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.2981132075471698, "acc_stderr": 0.028152837942493864, "acc_norm": 0.2981132075471698, "acc_norm_stderr": 0.028152837942493864 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.22727272727272727, "acc_stderr": 0.040139645540727735, "acc_norm": 0.22727272727272727, "acc_norm_stderr": 0.040139645540727735 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.26296296296296295, "acc_stderr": 0.02684205787383371, "acc_norm": 0.26296296296296295, "acc_norm_stderr": 0.02684205787383371 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.33112582781456956, "acc_stderr": 0.038425817186598696, "acc_norm": 0.33112582781456956, "acc_norm_stderr": 0.038425817186598696 }, "harness|ko_mmlu_sociology|5": { "acc": 0.26865671641791045, "acc_stderr": 0.03134328358208954, "acc_norm": 0.26865671641791045, "acc_norm_stderr": 0.03134328358208954 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.3352601156069364, "acc_stderr": 0.03599586301247078, "acc_norm": 0.3352601156069364, "acc_norm_stderr": 0.03599586301247078 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2698412698412698, "acc_stderr": 0.022860838309232072, "acc_norm": 0.2698412698412698, "acc_norm_stderr": 0.022860838309232072 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2638888888888889, "acc_stderr": 0.03685651095897532, "acc_norm": 0.2638888888888889, "acc_norm_stderr": 0.03685651095897532 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.26, "acc_stderr": 0.044084400227680794, "acc_norm": 0.26, "acc_norm_stderr": 0.044084400227680794 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.2138728323699422, "acc_stderr": 0.022075709251757173, "acc_norm": 0.2138728323699422, "acc_norm_stderr": 0.022075709251757173 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.2331288343558282, "acc_stderr": 0.0332201579577674, "acc_norm": 0.2331288343558282, "acc_norm_stderr": 0.0332201579577674 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.22530864197530864, "acc_stderr": 0.02324620264781975, "acc_norm": 0.22530864197530864, "acc_norm_stderr": 0.02324620264781975 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.36787564766839376, "acc_stderr": 0.034801756684660366, "acc_norm": 0.36787564766839376, "acc_norm_stderr": 0.034801756684660366 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.23684210526315788, "acc_stderr": 0.039994238792813365, "acc_norm": 0.23684210526315788, "acc_norm_stderr": 0.039994238792813365 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.3504587155963303, "acc_stderr": 0.02045607759982446, "acc_norm": 0.3504587155963303, "acc_norm_stderr": 0.02045607759982446 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.36507936507936506, "acc_stderr": 0.04306241259127153, "acc_norm": 0.36507936507936506, "acc_norm_stderr": 0.04306241259127153 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.29411764705882354, "acc_stderr": 0.026090162504279053, "acc_norm": 0.29411764705882354, "acc_norm_stderr": 0.026090162504279053 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.21, "acc_stderr": 0.040936018074033256, "acc_norm": 0.21, "acc_norm_stderr": 0.040936018074033256 }, "harness|ko_mmlu_international_law|5": { "acc": 0.14049586776859505, "acc_stderr": 0.031722334260021585, "acc_norm": 0.14049586776859505, "acc_norm_stderr": 0.031722334260021585 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.3355263157894737, "acc_stderr": 0.038424985593952694, "acc_norm": 0.3355263157894737, "acc_norm_stderr": 0.038424985593952694 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.2173202614379085, "acc_stderr": 0.016684820929148598, "acc_norm": 0.2173202614379085, "acc_norm_stderr": 0.016684820929148598 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.24113475177304963, "acc_stderr": 0.02551873104953776, "acc_norm": 0.24113475177304963, "acc_norm_stderr": 0.02551873104953776 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.16964285714285715, "acc_stderr": 0.03562367850095391, "acc_norm": 0.16964285714285715, "acc_norm_stderr": 0.03562367850095391 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4722222222222222, "acc_stderr": 0.0340470532865388, "acc_norm": 0.4722222222222222, "acc_norm_stderr": 0.0340470532865388 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.27262569832402234, "acc_stderr": 0.014893391735249608, "acc_norm": 0.27262569832402234, "acc_norm_stderr": 0.014893391735249608 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.18, "acc_stderr": 0.038612291966536934, "acc_norm": 0.18, "acc_norm_stderr": 0.038612291966536934 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4485294117647059, "acc_stderr": 0.030211479609121593, "acc_norm": 0.4485294117647059, "acc_norm_stderr": 0.030211479609121593 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.37142857142857144, "acc_stderr": 0.030932858792789834, "acc_norm": 0.37142857142857144, "acc_norm_stderr": 0.030932858792789834 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.20253164556962025, "acc_stderr": 0.026160568246601464, "acc_norm": 0.20253164556962025, "acc_norm_stderr": 0.026160568246601464 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.2438070404172099, "acc_stderr": 0.010966507972178475, "acc_norm": 0.2438070404172099, "acc_norm_stderr": 0.010966507972178475 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.25980392156862747, "acc_stderr": 0.030778554678693257, "acc_norm": 0.25980392156862747, "acc_norm_stderr": 0.030778554678693257 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.2727272727272727, "acc_stderr": 0.03477691162163659, "acc_norm": 0.2727272727272727, "acc_norm_stderr": 0.03477691162163659 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2533659730722154, "mc1_stderr": 0.01522589934082683, "mc2": 0.3923103125697379, "mc2_stderr": 0.014648106435610566 }, "harness|ko_commongen_v2|2": { "acc": 0.2857142857142857, "acc_stderr": 0.01553162078698674, "acc_norm": 0.3565525383707202, "acc_norm_stderr": 0.016467706981527448 } }
{ "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }
{ "model_name": "EleutherAI/polyglot-ko-5.8b", "model_sha": "581a4c3eebfac23536b3c9676bcfb05c6a97baa2", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null }
README.md exists but content is empty. Use the Edit dataset card button to edit it.
Downloads last month
0
Edit dataset card