|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.2986348122866894, |
|
"acc_stderr": 0.013374078615068759, |
|
"acc_norm": 0.35494880546075086, |
|
"acc_norm_stderr": 0.013983036904094095 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.4047998406691894, |
|
"acc_stderr": 0.004898501014225842, |
|
"acc_norm": 0.5283808006373233, |
|
"acc_norm_stderr": 0.004981736689518753 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.2807017543859649, |
|
"acc_stderr": 0.03446296217088426, |
|
"acc_norm": 0.2807017543859649, |
|
"acc_norm_stderr": 0.03446296217088426 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.1941747572815534, |
|
"acc_stderr": 0.03916667762822584, |
|
"acc_norm": 0.1941747572815534, |
|
"acc_norm_stderr": 0.03916667762822584 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.27458492975734355, |
|
"acc_stderr": 0.015959829933084032, |
|
"acc_norm": 0.27458492975734355, |
|
"acc_norm_stderr": 0.015959829933084032 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.32592592592592595, |
|
"acc_stderr": 0.040491220417025055, |
|
"acc_norm": 0.32592592592592595, |
|
"acc_norm_stderr": 0.040491220417025055 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.225531914893617, |
|
"acc_stderr": 0.02732107841738753, |
|
"acc_norm": 0.225531914893617, |
|
"acc_norm_stderr": 0.02732107841738753 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.18674698795180722, |
|
"acc_stderr": 0.03033874914450058, |
|
"acc_norm": 0.18674698795180722, |
|
"acc_norm_stderr": 0.03033874914450058 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.3086816720257235, |
|
"acc_stderr": 0.026236965881153266, |
|
"acc_norm": 0.3086816720257235, |
|
"acc_norm_stderr": 0.026236965881153266 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.20179372197309417, |
|
"acc_stderr": 0.02693611191280227, |
|
"acc_norm": 0.20179372197309417, |
|
"acc_norm_stderr": 0.02693611191280227 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.22900763358778625, |
|
"acc_stderr": 0.036853466317118506, |
|
"acc_norm": 0.22900763358778625, |
|
"acc_norm_stderr": 0.036853466317118506 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.24, |
|
"acc_stderr": 0.04292346959909283, |
|
"acc_norm": 0.24, |
|
"acc_norm_stderr": 0.04292346959909283 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.2676767676767677, |
|
"acc_stderr": 0.031544498882702866, |
|
"acc_norm": 0.2676767676767677, |
|
"acc_norm_stderr": 0.031544498882702866 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.31724137931034485, |
|
"acc_stderr": 0.038783523721386236, |
|
"acc_norm": 0.31724137931034485, |
|
"acc_norm_stderr": 0.038783523721386236 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.28431372549019607, |
|
"acc_stderr": 0.04488482852329017, |
|
"acc_norm": 0.28431372549019607, |
|
"acc_norm_stderr": 0.04488482852329017 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.226890756302521, |
|
"acc_stderr": 0.027205371538279472, |
|
"acc_norm": 0.226890756302521, |
|
"acc_norm_stderr": 0.027205371538279472 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.24871794871794872, |
|
"acc_stderr": 0.0219169577092138, |
|
"acc_norm": 0.24871794871794872, |
|
"acc_norm_stderr": 0.0219169577092138 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.2222222222222222, |
|
"acc_stderr": 0.040191074725573483, |
|
"acc_norm": 0.2222222222222222, |
|
"acc_norm_stderr": 0.040191074725573483 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.2512315270935961, |
|
"acc_stderr": 0.030516530732694436, |
|
"acc_norm": 0.2512315270935961, |
|
"acc_norm_stderr": 0.030516530732694436 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.24838709677419354, |
|
"acc_stderr": 0.02458002892148101, |
|
"acc_norm": 0.24838709677419354, |
|
"acc_norm_stderr": 0.02458002892148101 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.2777777777777778, |
|
"acc_stderr": 0.02934311479809447, |
|
"acc_norm": 0.2777777777777778, |
|
"acc_norm_stderr": 0.02934311479809447 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.22264150943396227, |
|
"acc_stderr": 0.02560423347089911, |
|
"acc_norm": 0.22264150943396227, |
|
"acc_norm_stderr": 0.02560423347089911 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.22727272727272727, |
|
"acc_stderr": 0.04013964554072774, |
|
"acc_norm": 0.22727272727272727, |
|
"acc_norm_stderr": 0.04013964554072774 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.24444444444444444, |
|
"acc_stderr": 0.02620276653465215, |
|
"acc_norm": 0.24444444444444444, |
|
"acc_norm_stderr": 0.02620276653465215 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.24503311258278146, |
|
"acc_stderr": 0.03511807571804725, |
|
"acc_norm": 0.24503311258278146, |
|
"acc_norm_stderr": 0.03511807571804725 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.24875621890547264, |
|
"acc_stderr": 0.030567675938916714, |
|
"acc_norm": 0.24875621890547264, |
|
"acc_norm_stderr": 0.030567675938916714 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.28901734104046245, |
|
"acc_stderr": 0.03456425745087, |
|
"acc_norm": 0.28901734104046245, |
|
"acc_norm_stderr": 0.03456425745087 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.2804232804232804, |
|
"acc_stderr": 0.023135287974325618, |
|
"acc_norm": 0.2804232804232804, |
|
"acc_norm_stderr": 0.023135287974325618 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.2986111111111111, |
|
"acc_stderr": 0.038270523579507554, |
|
"acc_norm": 0.2986111111111111, |
|
"acc_norm_stderr": 0.038270523579507554 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.16, |
|
"acc_stderr": 0.03684529491774709, |
|
"acc_norm": 0.16, |
|
"acc_norm_stderr": 0.03684529491774709 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.045126085985421276, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.045126085985421276 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.24566473988439305, |
|
"acc_stderr": 0.02317629820399201, |
|
"acc_norm": 0.24566473988439305, |
|
"acc_norm_stderr": 0.02317629820399201 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.3006134969325153, |
|
"acc_stderr": 0.03602511318806771, |
|
"acc_norm": 0.3006134969325153, |
|
"acc_norm_stderr": 0.03602511318806771 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.29012345679012347, |
|
"acc_stderr": 0.02525117393649502, |
|
"acc_norm": 0.29012345679012347, |
|
"acc_norm_stderr": 0.02525117393649502 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621504, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621504 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.23834196891191708, |
|
"acc_stderr": 0.030748905363909895, |
|
"acc_norm": 0.23834196891191708, |
|
"acc_norm_stderr": 0.030748905363909895 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2719298245614035, |
|
"acc_stderr": 0.04185774424022057, |
|
"acc_norm": 0.2719298245614035, |
|
"acc_norm_stderr": 0.04185774424022057 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.23669724770642203, |
|
"acc_stderr": 0.01822407811729906, |
|
"acc_norm": 0.23669724770642203, |
|
"acc_norm_stderr": 0.01822407811729906 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.15079365079365079, |
|
"acc_stderr": 0.03200686497287392, |
|
"acc_norm": 0.15079365079365079, |
|
"acc_norm_stderr": 0.03200686497287392 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.2647058823529412, |
|
"acc_stderr": 0.025261691219729487, |
|
"acc_norm": 0.2647058823529412, |
|
"acc_norm_stderr": 0.025261691219729487 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.23, |
|
"acc_stderr": 0.04229525846816506, |
|
"acc_norm": 0.23, |
|
"acc_norm_stderr": 0.04229525846816506 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.36363636363636365, |
|
"acc_stderr": 0.043913262867240704, |
|
"acc_norm": 0.36363636363636365, |
|
"acc_norm_stderr": 0.043913262867240704 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.2894736842105263, |
|
"acc_stderr": 0.03690677986137283, |
|
"acc_norm": 0.2894736842105263, |
|
"acc_norm_stderr": 0.03690677986137283 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.24673202614379086, |
|
"acc_stderr": 0.0174408203674025, |
|
"acc_norm": 0.24673202614379086, |
|
"acc_norm_stderr": 0.0174408203674025 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.23404255319148937, |
|
"acc_stderr": 0.025257861359432403, |
|
"acc_norm": 0.23404255319148937, |
|
"acc_norm_stderr": 0.025257861359432403 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04109974682633932, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04109974682633932 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.4166666666666667, |
|
"acc_stderr": 0.03362277436608044, |
|
"acc_norm": 0.4166666666666667, |
|
"acc_norm_stderr": 0.03362277436608044 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2547486033519553, |
|
"acc_stderr": 0.014572650383409162, |
|
"acc_norm": 0.2547486033519553, |
|
"acc_norm_stderr": 0.014572650383409162 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.04512608598542127, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.04512608598542127 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.0440844002276808, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.0440844002276808 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.33455882352941174, |
|
"acc_stderr": 0.028661996202335317, |
|
"acc_norm": 0.33455882352941174, |
|
"acc_norm_stderr": 0.028661996202335317 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.3836734693877551, |
|
"acc_stderr": 0.031130880396235922, |
|
"acc_norm": 0.3836734693877551, |
|
"acc_norm_stderr": 0.031130880396235922 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.1940928270042194, |
|
"acc_stderr": 0.025744902532290927, |
|
"acc_norm": 0.1940928270042194, |
|
"acc_norm_stderr": 0.025744902532290927 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.2542372881355932, |
|
"acc_stderr": 0.011121129007840673, |
|
"acc_norm": 0.2542372881355932, |
|
"acc_norm_stderr": 0.011121129007840673 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.25980392156862747, |
|
"acc_stderr": 0.030778554678693264, |
|
"acc_norm": 0.25980392156862747, |
|
"acc_norm_stderr": 0.030778554678693264 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.24848484848484848, |
|
"acc_stderr": 0.03374402644139405, |
|
"acc_norm": 0.24848484848484848, |
|
"acc_norm_stderr": 0.03374402644139405 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.23745410036719705, |
|
"mc1_stderr": 0.014896277441041866, |
|
"mc2": 0.3991183406834575, |
|
"mc2_stderr": 0.01489709315395438 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.3305785123966942, |
|
"acc_stderr": 0.0161734232988457, |
|
"acc_norm": 0.3837072018890201, |
|
"acc_norm_stderr": 0.016718924637231826 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "DILAB-HYU/koquality-polyglot-12.8b", |
|
"model_sha": "7cca798e18e44cebbde1c6d1f59162882c2bf254", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |