|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.2986348122866894, |
|
"acc_stderr": 0.013374078615068754, |
|
"acc_norm": 0.34982935153583616, |
|
"acc_norm_stderr": 0.013936809212158296 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.39294961163114916, |
|
"acc_stderr": 0.004874076250521577, |
|
"acc_norm": 0.4985062736506672, |
|
"acc_norm_stderr": 0.004989759144812286 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.24561403508771928, |
|
"acc_stderr": 0.0330140594698725, |
|
"acc_norm": 0.24561403508771928, |
|
"acc_norm_stderr": 0.0330140594698725 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.17475728155339806, |
|
"acc_stderr": 0.037601780060266196, |
|
"acc_norm": 0.17475728155339806, |
|
"acc_norm_stderr": 0.037601780060266196 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.26053639846743293, |
|
"acc_stderr": 0.015696008563807096, |
|
"acc_norm": 0.26053639846743293, |
|
"acc_norm_stderr": 0.015696008563807096 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.26666666666666666, |
|
"acc_stderr": 0.038201699145179055, |
|
"acc_norm": 0.26666666666666666, |
|
"acc_norm_stderr": 0.038201699145179055 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.22, |
|
"acc_stderr": 0.04163331998932269, |
|
"acc_norm": 0.22, |
|
"acc_norm_stderr": 0.04163331998932269 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.25957446808510637, |
|
"acc_stderr": 0.028659179374292326, |
|
"acc_norm": 0.25957446808510637, |
|
"acc_norm_stderr": 0.028659179374292326 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.24096385542168675, |
|
"acc_stderr": 0.03329394119073529, |
|
"acc_norm": 0.24096385542168675, |
|
"acc_norm_stderr": 0.03329394119073529 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.21864951768488747, |
|
"acc_stderr": 0.02347558141786111, |
|
"acc_norm": 0.21864951768488747, |
|
"acc_norm_stderr": 0.02347558141786111 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.2556053811659193, |
|
"acc_stderr": 0.029275891003969927, |
|
"acc_norm": 0.2556053811659193, |
|
"acc_norm_stderr": 0.029275891003969927 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.1984732824427481, |
|
"acc_stderr": 0.0349814938546247, |
|
"acc_norm": 0.1984732824427481, |
|
"acc_norm_stderr": 0.0349814938546247 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.04408440022768079, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.04408440022768079 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.2474747474747475, |
|
"acc_stderr": 0.0307463007421245, |
|
"acc_norm": 0.2474747474747475, |
|
"acc_norm_stderr": 0.0307463007421245 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.2413793103448276, |
|
"acc_stderr": 0.03565998174135303, |
|
"acc_norm": 0.2413793103448276, |
|
"acc_norm_stderr": 0.03565998174135303 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.23529411764705882, |
|
"acc_stderr": 0.04220773659171453, |
|
"acc_norm": 0.23529411764705882, |
|
"acc_norm_stderr": 0.04220773659171453 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.25630252100840334, |
|
"acc_stderr": 0.02835962087053395, |
|
"acc_norm": 0.25630252100840334, |
|
"acc_norm_stderr": 0.02835962087053395 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.20512820512820512, |
|
"acc_stderr": 0.020473233173551986, |
|
"acc_norm": 0.20512820512820512, |
|
"acc_norm_stderr": 0.020473233173551986 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.044619604333847394, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.044619604333847394 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.23, |
|
"acc_stderr": 0.04229525846816505, |
|
"acc_norm": 0.23, |
|
"acc_norm_stderr": 0.04229525846816505 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.25925925925925924, |
|
"acc_stderr": 0.04236511258094633, |
|
"acc_norm": 0.25925925925925924, |
|
"acc_norm_stderr": 0.04236511258094633 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.2315270935960591, |
|
"acc_stderr": 0.029678333141444455, |
|
"acc_norm": 0.2315270935960591, |
|
"acc_norm_stderr": 0.029678333141444455 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.23548387096774193, |
|
"acc_stderr": 0.02413763242933771, |
|
"acc_norm": 0.23548387096774193, |
|
"acc_norm_stderr": 0.02413763242933771 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.23931623931623933, |
|
"acc_stderr": 0.02795182680892433, |
|
"acc_norm": 0.23931623931623933, |
|
"acc_norm_stderr": 0.02795182680892433 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.3283018867924528, |
|
"acc_stderr": 0.02890159361241178, |
|
"acc_norm": 0.3283018867924528, |
|
"acc_norm_stderr": 0.02890159361241178 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.2909090909090909, |
|
"acc_stderr": 0.04350271442923243, |
|
"acc_norm": 0.2909090909090909, |
|
"acc_norm_stderr": 0.04350271442923243 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.22592592592592592, |
|
"acc_stderr": 0.02549753263960954, |
|
"acc_norm": 0.22592592592592592, |
|
"acc_norm_stderr": 0.02549753263960954 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.271523178807947, |
|
"acc_stderr": 0.03631329803969654, |
|
"acc_norm": 0.271523178807947, |
|
"acc_norm_stderr": 0.03631329803969654 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.24378109452736318, |
|
"acc_stderr": 0.030360490154014652, |
|
"acc_norm": 0.24378109452736318, |
|
"acc_norm_stderr": 0.030360490154014652 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.2832369942196532, |
|
"acc_stderr": 0.03435568056047875, |
|
"acc_norm": 0.2832369942196532, |
|
"acc_norm_stderr": 0.03435568056047875 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.23809523809523808, |
|
"acc_stderr": 0.02193587808118476, |
|
"acc_norm": 0.23809523809523808, |
|
"acc_norm_stderr": 0.02193587808118476 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.03621034121889507, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.03621034121889507 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.044084400227680794, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.044084400227680794 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621505, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621505 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.24277456647398843, |
|
"acc_stderr": 0.023083658586984204, |
|
"acc_norm": 0.24277456647398843, |
|
"acc_norm_stderr": 0.023083658586984204 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.2085889570552147, |
|
"acc_stderr": 0.03192193448934725, |
|
"acc_norm": 0.2085889570552147, |
|
"acc_norm_stderr": 0.03192193448934725 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.2345679012345679, |
|
"acc_stderr": 0.02357688174400572, |
|
"acc_norm": 0.2345679012345679, |
|
"acc_norm_stderr": 0.02357688174400572 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.04461960433384739, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.04461960433384739 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.22279792746113988, |
|
"acc_stderr": 0.030031147977641545, |
|
"acc_norm": 0.22279792746113988, |
|
"acc_norm_stderr": 0.030031147977641545 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2719298245614035, |
|
"acc_stderr": 0.04185774424022057, |
|
"acc_norm": 0.2719298245614035, |
|
"acc_norm_stderr": 0.04185774424022057 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.22568807339449543, |
|
"acc_stderr": 0.017923087667803053, |
|
"acc_norm": 0.22568807339449543, |
|
"acc_norm_stderr": 0.017923087667803053 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.25396825396825395, |
|
"acc_stderr": 0.03893259610604672, |
|
"acc_norm": 0.25396825396825395, |
|
"acc_norm_stderr": 0.03893259610604672 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.20588235294117646, |
|
"acc_stderr": 0.023152722439402303, |
|
"acc_norm": 0.20588235294117646, |
|
"acc_norm_stderr": 0.023152722439402303 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.34, |
|
"acc_stderr": 0.04760952285695235, |
|
"acc_norm": 0.34, |
|
"acc_norm_stderr": 0.04760952285695235 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.23140495867768596, |
|
"acc_stderr": 0.03849856098794088, |
|
"acc_norm": 0.23140495867768596, |
|
"acc_norm_stderr": 0.03849856098794088 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.28289473684210525, |
|
"acc_stderr": 0.03665349695640767, |
|
"acc_norm": 0.28289473684210525, |
|
"acc_norm_stderr": 0.03665349695640767 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.25980392156862747, |
|
"acc_stderr": 0.01774089950917779, |
|
"acc_norm": 0.25980392156862747, |
|
"acc_norm_stderr": 0.01774089950917779 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.24822695035460993, |
|
"acc_stderr": 0.025770015644290392, |
|
"acc_norm": 0.24822695035460993, |
|
"acc_norm_stderr": 0.025770015644290392 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.2857142857142857, |
|
"acc_stderr": 0.042878587513404544, |
|
"acc_norm": 0.2857142857142857, |
|
"acc_norm_stderr": 0.042878587513404544 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.32407407407407407, |
|
"acc_stderr": 0.03191923445686185, |
|
"acc_norm": 0.32407407407407407, |
|
"acc_norm_stderr": 0.03191923445686185 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.26927374301675977, |
|
"acc_stderr": 0.014835616582882603, |
|
"acc_norm": 0.26927374301675977, |
|
"acc_norm_stderr": 0.014835616582882603 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621504, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621504 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.19, |
|
"acc_stderr": 0.03942772444036624, |
|
"acc_norm": 0.19, |
|
"acc_norm_stderr": 0.03942772444036624 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.2977941176470588, |
|
"acc_stderr": 0.02777829870154544, |
|
"acc_norm": 0.2977941176470588, |
|
"acc_norm_stderr": 0.02777829870154544 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.23673469387755103, |
|
"acc_stderr": 0.027212835884073142, |
|
"acc_norm": 0.23673469387755103, |
|
"acc_norm_stderr": 0.027212835884073142 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.27848101265822783, |
|
"acc_stderr": 0.029178682304842538, |
|
"acc_norm": 0.27848101265822783, |
|
"acc_norm_stderr": 0.029178682304842538 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.24315514993481094, |
|
"acc_stderr": 0.010956556654417355, |
|
"acc_norm": 0.24315514993481094, |
|
"acc_norm_stderr": 0.010956556654417355 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.2549019607843137, |
|
"acc_stderr": 0.030587591351604246, |
|
"acc_norm": 0.2549019607843137, |
|
"acc_norm_stderr": 0.030587591351604246 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.21212121212121213, |
|
"acc_stderr": 0.03192271569548299, |
|
"acc_norm": 0.21212121212121213, |
|
"acc_norm_stderr": 0.03192271569548299 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.26438188494492043, |
|
"mc1_stderr": 0.015438211119522512, |
|
"mc2": 0.4146123621762204, |
|
"mc2_stderr": 0.015628722266635826 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.27390791027154665, |
|
"acc_stderr": 0.01533249947479102, |
|
"acc_norm": 0.29988193624557263, |
|
"acc_norm_stderr": 0.01575344761542946 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "MrBananaHuman/polyglot-ko-5.8b", |
|
"model_sha": "05f8800a617b483b1c502d8c965ff434e63d387c", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |