|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.371160409556314, |
|
"acc_stderr": 0.014117971901142817, |
|
"acc_norm": 0.4283276450511945, |
|
"acc_norm_stderr": 0.014460496367599022 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.37950607448715395, |
|
"acc_stderr": 0.004842723234022034, |
|
"acc_norm": 0.481876120294762, |
|
"acc_norm_stderr": 0.00498650229693118 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.4619883040935672, |
|
"acc_stderr": 0.03823727092882307, |
|
"acc_norm": 0.4619883040935672, |
|
"acc_norm_stderr": 0.03823727092882307 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6019417475728155, |
|
"acc_stderr": 0.048467482539772386, |
|
"acc_norm": 0.6019417475728155, |
|
"acc_norm_stderr": 0.048467482539772386 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.4840357598978289, |
|
"acc_stderr": 0.017870847506081717, |
|
"acc_norm": 0.4840357598978289, |
|
"acc_norm_stderr": 0.017870847506081717 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.37777777777777777, |
|
"acc_stderr": 0.04188307537595853, |
|
"acc_norm": 0.37777777777777777, |
|
"acc_norm_stderr": 0.04188307537595853 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.4085106382978723, |
|
"acc_stderr": 0.03213418026701576, |
|
"acc_norm": 0.4085106382978723, |
|
"acc_norm_stderr": 0.03213418026701576 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.4457831325301205, |
|
"acc_stderr": 0.038695433234721015, |
|
"acc_norm": 0.4457831325301205, |
|
"acc_norm_stderr": 0.038695433234721015 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.41479099678456594, |
|
"acc_stderr": 0.027982680459759556, |
|
"acc_norm": 0.41479099678456594, |
|
"acc_norm_stderr": 0.027982680459759556 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.4663677130044843, |
|
"acc_stderr": 0.033481800170603065, |
|
"acc_norm": 0.4663677130044843, |
|
"acc_norm_stderr": 0.033481800170603065 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.366412213740458, |
|
"acc_stderr": 0.04225875451969638, |
|
"acc_norm": 0.366412213740458, |
|
"acc_norm_stderr": 0.04225875451969638 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.04824181513244218, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.04824181513244218 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5606060606060606, |
|
"acc_stderr": 0.035360859475294805, |
|
"acc_norm": 0.5606060606060606, |
|
"acc_norm_stderr": 0.035360859475294805 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.3793103448275862, |
|
"acc_stderr": 0.04043461861916747, |
|
"acc_norm": 0.3793103448275862, |
|
"acc_norm_stderr": 0.04043461861916747 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.24509803921568626, |
|
"acc_stderr": 0.04280105837364396, |
|
"acc_norm": 0.24509803921568626, |
|
"acc_norm_stderr": 0.04280105837364396 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.46218487394957986, |
|
"acc_stderr": 0.032385469487589795, |
|
"acc_norm": 0.46218487394957986, |
|
"acc_norm_stderr": 0.032385469487589795 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4076923076923077, |
|
"acc_stderr": 0.024915243985987837, |
|
"acc_norm": 0.4076923076923077, |
|
"acc_norm_stderr": 0.024915243985987837 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.63, |
|
"acc_stderr": 0.048523658709390974, |
|
"acc_norm": 0.63, |
|
"acc_norm_stderr": 0.048523658709390974 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621505, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621505 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5092592592592593, |
|
"acc_stderr": 0.04832853553437055, |
|
"acc_norm": 0.5092592592592593, |
|
"acc_norm_stderr": 0.04832853553437055 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.42857142857142855, |
|
"acc_stderr": 0.03481904844438803, |
|
"acc_norm": 0.42857142857142855, |
|
"acc_norm_stderr": 0.03481904844438803 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.43548387096774194, |
|
"acc_stderr": 0.02820622559150274, |
|
"acc_norm": 0.43548387096774194, |
|
"acc_norm_stderr": 0.02820622559150274 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.688034188034188, |
|
"acc_stderr": 0.030351527323344944, |
|
"acc_norm": 0.688034188034188, |
|
"acc_norm_stderr": 0.030351527323344944 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.45660377358490567, |
|
"acc_stderr": 0.030656748696739428, |
|
"acc_norm": 0.45660377358490567, |
|
"acc_norm_stderr": 0.030656748696739428 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5272727272727272, |
|
"acc_stderr": 0.04782001791380061, |
|
"acc_norm": 0.5272727272727272, |
|
"acc_norm_stderr": 0.04782001791380061 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.34814814814814815, |
|
"acc_stderr": 0.029045600290616255, |
|
"acc_norm": 0.34814814814814815, |
|
"acc_norm_stderr": 0.029045600290616255 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3708609271523179, |
|
"acc_stderr": 0.03943966699183629, |
|
"acc_norm": 0.3708609271523179, |
|
"acc_norm_stderr": 0.03943966699183629 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.5870646766169154, |
|
"acc_stderr": 0.03481520803367348, |
|
"acc_norm": 0.5870646766169154, |
|
"acc_norm_stderr": 0.03481520803367348 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.36416184971098264, |
|
"acc_stderr": 0.03669072477416908, |
|
"acc_norm": 0.36416184971098264, |
|
"acc_norm_stderr": 0.03669072477416908 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.35978835978835977, |
|
"acc_stderr": 0.024718075944129277, |
|
"acc_norm": 0.35978835978835977, |
|
"acc_norm_stderr": 0.024718075944129277 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3402777777777778, |
|
"acc_stderr": 0.039621355734862175, |
|
"acc_norm": 0.3402777777777778, |
|
"acc_norm_stderr": 0.039621355734862175 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.04878317312145632, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.04878317312145632 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.57, |
|
"acc_stderr": 0.04975698519562427, |
|
"acc_norm": 0.57, |
|
"acc_norm_stderr": 0.04975698519562427 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.48265895953757226, |
|
"acc_stderr": 0.026902900458666633, |
|
"acc_norm": 0.48265895953757226, |
|
"acc_norm_stderr": 0.026902900458666633 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.4233128834355828, |
|
"acc_stderr": 0.03881891213334383, |
|
"acc_norm": 0.4233128834355828, |
|
"acc_norm_stderr": 0.03881891213334383 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.4351851851851852, |
|
"acc_stderr": 0.0275860062216077, |
|
"acc_norm": 0.4351851851851852, |
|
"acc_norm_stderr": 0.0275860062216077 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.04824181513244218, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.04824181513244218 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.46113989637305697, |
|
"acc_stderr": 0.035975244117345775, |
|
"acc_norm": 0.46113989637305697, |
|
"acc_norm_stderr": 0.035975244117345775 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2719298245614035, |
|
"acc_stderr": 0.04185774424022056, |
|
"acc_norm": 0.2719298245614035, |
|
"acc_norm_stderr": 0.04185774424022056 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.45321100917431195, |
|
"acc_stderr": 0.021343255165546034, |
|
"acc_norm": 0.45321100917431195, |
|
"acc_norm_stderr": 0.021343255165546034 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.2698412698412698, |
|
"acc_stderr": 0.03970158273235172, |
|
"acc_norm": 0.2698412698412698, |
|
"acc_norm_stderr": 0.03970158273235172 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.41830065359477125, |
|
"acc_stderr": 0.02824513402438729, |
|
"acc_norm": 0.41830065359477125, |
|
"acc_norm_stderr": 0.02824513402438729 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.49, |
|
"acc_stderr": 0.05024183937956911, |
|
"acc_norm": 0.49, |
|
"acc_norm_stderr": 0.05024183937956911 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6694214876033058, |
|
"acc_stderr": 0.04294340845212094, |
|
"acc_norm": 0.6694214876033058, |
|
"acc_norm_stderr": 0.04294340845212094 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.40789473684210525, |
|
"acc_stderr": 0.03999309712777472, |
|
"acc_norm": 0.40789473684210525, |
|
"acc_norm_stderr": 0.03999309712777472 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.4084967320261438, |
|
"acc_stderr": 0.019886221037501872, |
|
"acc_norm": 0.4084967320261438, |
|
"acc_norm_stderr": 0.019886221037501872 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3475177304964539, |
|
"acc_stderr": 0.028406627809590954, |
|
"acc_norm": 0.3475177304964539, |
|
"acc_norm_stderr": 0.028406627809590954 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.38392857142857145, |
|
"acc_stderr": 0.046161430750285455, |
|
"acc_norm": 0.38392857142857145, |
|
"acc_norm_stderr": 0.046161430750285455 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.3148148148148148, |
|
"acc_stderr": 0.03167468706828978, |
|
"acc_norm": 0.3148148148148148, |
|
"acc_norm_stderr": 0.03167468706828978 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.23575418994413408, |
|
"acc_stderr": 0.014196375686290804, |
|
"acc_norm": 0.23575418994413408, |
|
"acc_norm_stderr": 0.014196375686290804 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.54, |
|
"acc_stderr": 0.05009082659620332, |
|
"acc_norm": 0.54, |
|
"acc_norm_stderr": 0.05009082659620332 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.33088235294117646, |
|
"acc_stderr": 0.028582709753898435, |
|
"acc_norm": 0.33088235294117646, |
|
"acc_norm_stderr": 0.028582709753898435 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.3510204081632653, |
|
"acc_stderr": 0.03055531675557364, |
|
"acc_norm": 0.3510204081632653, |
|
"acc_norm_stderr": 0.03055531675557364 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.5569620253164557, |
|
"acc_stderr": 0.03233532777533484, |
|
"acc_norm": 0.5569620253164557, |
|
"acc_norm_stderr": 0.03233532777533484 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.32073011734028684, |
|
"acc_stderr": 0.011921199991782629, |
|
"acc_norm": 0.32073011734028684, |
|
"acc_norm_stderr": 0.011921199991782629 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.45098039215686275, |
|
"acc_stderr": 0.03492406104163614, |
|
"acc_norm": 0.45098039215686275, |
|
"acc_norm_stderr": 0.03492406104163614 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.4484848484848485, |
|
"acc_stderr": 0.038835659779569286, |
|
"acc_norm": 0.4484848484848485, |
|
"acc_norm_stderr": 0.038835659779569286 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.27906976744186046, |
|
"mc1_stderr": 0.01570210709062788, |
|
"mc2": 0.46295306302174644, |
|
"mc2_stderr": 0.015320970978421385 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.4085005903187721, |
|
"acc_stderr": 0.01690006287942712, |
|
"acc_norm": 0.4793388429752066, |
|
"acc_norm_stderr": 0.017175671279836442 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "MNCLLM/Mistral-7B-OP-over500-grad1.0", |
|
"model_sha": "f7789c5af9b3b166070a886207090228deccf9d6", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |