|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.44197952218430037, |
|
"acc_stderr": 0.014512682523128342, |
|
"acc_norm": 0.47525597269624575, |
|
"acc_norm_stderr": 0.014593487694937738 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.4400517825134435, |
|
"acc_stderr": 0.004953787146510924, |
|
"acc_norm": 0.5938060147381, |
|
"acc_norm_stderr": 0.0049011789179008464 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.4678362573099415, |
|
"acc_stderr": 0.03826882417660368, |
|
"acc_norm": 0.4678362573099415, |
|
"acc_norm_stderr": 0.03826882417660368 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6504854368932039, |
|
"acc_stderr": 0.047211885060971716, |
|
"acc_norm": 0.6504854368932039, |
|
"acc_norm_stderr": 0.047211885060971716 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5964240102171137, |
|
"acc_stderr": 0.01754433223792644, |
|
"acc_norm": 0.5964240102171137, |
|
"acc_norm_stderr": 0.01754433223792644 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4444444444444444, |
|
"acc_stderr": 0.04292596718256981, |
|
"acc_norm": 0.4444444444444444, |
|
"acc_norm_stderr": 0.04292596718256981 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.5276595744680851, |
|
"acc_stderr": 0.03263597118409769, |
|
"acc_norm": 0.5276595744680851, |
|
"acc_norm_stderr": 0.03263597118409769 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.42168674698795183, |
|
"acc_stderr": 0.03844453181770917, |
|
"acc_norm": 0.42168674698795183, |
|
"acc_norm_stderr": 0.03844453181770917 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.5691318327974276, |
|
"acc_stderr": 0.02812534098397271, |
|
"acc_norm": 0.5691318327974276, |
|
"acc_norm_stderr": 0.02812534098397271 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5246636771300448, |
|
"acc_stderr": 0.03351695167652628, |
|
"acc_norm": 0.5246636771300448, |
|
"acc_norm_stderr": 0.03351695167652628 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.5038167938931297, |
|
"acc_stderr": 0.043851623256015534, |
|
"acc_norm": 0.5038167938931297, |
|
"acc_norm_stderr": 0.043851623256015534 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.04960449637488583, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.04960449637488583 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.6717171717171717, |
|
"acc_stderr": 0.03345678422756776, |
|
"acc_norm": 0.6717171717171717, |
|
"acc_norm_stderr": 0.03345678422756776 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.5241379310344828, |
|
"acc_stderr": 0.041618085035015295, |
|
"acc_norm": 0.5241379310344828, |
|
"acc_norm_stderr": 0.041618085035015295 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.35294117647058826, |
|
"acc_stderr": 0.047551296160629475, |
|
"acc_norm": 0.35294117647058826, |
|
"acc_norm_stderr": 0.047551296160629475 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.592436974789916, |
|
"acc_stderr": 0.03191863374478466, |
|
"acc_norm": 0.592436974789916, |
|
"acc_norm_stderr": 0.03191863374478466 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.5358974358974359, |
|
"acc_stderr": 0.02528558599001786, |
|
"acc_norm": 0.5358974358974359, |
|
"acc_norm_stderr": 0.02528558599001786 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.62, |
|
"acc_stderr": 0.048783173121456316, |
|
"acc_norm": 0.62, |
|
"acc_norm_stderr": 0.048783173121456316 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6203703703703703, |
|
"acc_stderr": 0.04691521224077742, |
|
"acc_norm": 0.6203703703703703, |
|
"acc_norm_stderr": 0.04691521224077742 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.39901477832512317, |
|
"acc_stderr": 0.03445487686264716, |
|
"acc_norm": 0.39901477832512317, |
|
"acc_norm_stderr": 0.03445487686264716 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.5516129032258065, |
|
"acc_stderr": 0.02829205683011273, |
|
"acc_norm": 0.5516129032258065, |
|
"acc_norm_stderr": 0.02829205683011273 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7905982905982906, |
|
"acc_stderr": 0.026655699653922737, |
|
"acc_norm": 0.7905982905982906, |
|
"acc_norm_stderr": 0.026655699653922737 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.5094339622641509, |
|
"acc_stderr": 0.0307673947078081, |
|
"acc_norm": 0.5094339622641509, |
|
"acc_norm_stderr": 0.0307673947078081 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5454545454545454, |
|
"acc_stderr": 0.04769300568972744, |
|
"acc_norm": 0.5454545454545454, |
|
"acc_norm_stderr": 0.04769300568972744 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3814814814814815, |
|
"acc_stderr": 0.02961671892749759, |
|
"acc_norm": 0.3814814814814815, |
|
"acc_norm_stderr": 0.02961671892749759 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.37748344370860926, |
|
"acc_stderr": 0.0395802723112157, |
|
"acc_norm": 0.37748344370860926, |
|
"acc_norm_stderr": 0.0395802723112157 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6318407960199005, |
|
"acc_stderr": 0.03410410565495302, |
|
"acc_norm": 0.6318407960199005, |
|
"acc_norm_stderr": 0.03410410565495302 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.4624277456647399, |
|
"acc_stderr": 0.0380168510452446, |
|
"acc_norm": 0.4624277456647399, |
|
"acc_norm_stderr": 0.0380168510452446 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.41798941798941797, |
|
"acc_stderr": 0.02540255550326091, |
|
"acc_norm": 0.41798941798941797, |
|
"acc_norm_stderr": 0.02540255550326091 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4583333333333333, |
|
"acc_stderr": 0.04166666666666665, |
|
"acc_norm": 0.4583333333333333, |
|
"acc_norm_stderr": 0.04166666666666665 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621504, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621504 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.67, |
|
"acc_stderr": 0.04725815626252606, |
|
"acc_norm": 0.67, |
|
"acc_norm_stderr": 0.04725815626252606 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5491329479768786, |
|
"acc_stderr": 0.02678881193156276, |
|
"acc_norm": 0.5491329479768786, |
|
"acc_norm_stderr": 0.02678881193156276 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.48466257668711654, |
|
"acc_stderr": 0.039265223787088424, |
|
"acc_norm": 0.48466257668711654, |
|
"acc_norm_stderr": 0.039265223787088424 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.5462962962962963, |
|
"acc_stderr": 0.027701228468542595, |
|
"acc_norm": 0.5462962962962963, |
|
"acc_norm_stderr": 0.027701228468542595 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.34, |
|
"acc_stderr": 0.04760952285695235, |
|
"acc_norm": 0.34, |
|
"acc_norm_stderr": 0.04760952285695235 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.6994818652849741, |
|
"acc_stderr": 0.0330881859441575, |
|
"acc_norm": 0.6994818652849741, |
|
"acc_norm_stderr": 0.0330881859441575 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.4473684210526316, |
|
"acc_stderr": 0.04677473004491199, |
|
"acc_norm": 0.4473684210526316, |
|
"acc_norm_stderr": 0.04677473004491199 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.6623853211009174, |
|
"acc_stderr": 0.020275265986638907, |
|
"acc_norm": 0.6623853211009174, |
|
"acc_norm_stderr": 0.020275265986638907 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.4126984126984127, |
|
"acc_stderr": 0.04403438954768177, |
|
"acc_norm": 0.4126984126984127, |
|
"acc_norm_stderr": 0.04403438954768177 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.4934640522875817, |
|
"acc_stderr": 0.028627470550556047, |
|
"acc_norm": 0.4934640522875817, |
|
"acc_norm_stderr": 0.028627470550556047 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.63, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.63, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6611570247933884, |
|
"acc_stderr": 0.04320767807536671, |
|
"acc_norm": 0.6611570247933884, |
|
"acc_norm_stderr": 0.04320767807536671 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5394736842105263, |
|
"acc_stderr": 0.04056242252249034, |
|
"acc_norm": 0.5394736842105263, |
|
"acc_norm_stderr": 0.04056242252249034 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.48856209150326796, |
|
"acc_stderr": 0.02022254151561087, |
|
"acc_norm": 0.48856209150326796, |
|
"acc_norm_stderr": 0.02022254151561087 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3900709219858156, |
|
"acc_stderr": 0.02909767559946393, |
|
"acc_norm": 0.3900709219858156, |
|
"acc_norm_stderr": 0.02909767559946393 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.39285714285714285, |
|
"acc_stderr": 0.04635550135609976, |
|
"acc_norm": 0.39285714285714285, |
|
"acc_norm_stderr": 0.04635550135609976 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.46296296296296297, |
|
"acc_stderr": 0.03400603625538272, |
|
"acc_norm": 0.46296296296296297, |
|
"acc_norm_stderr": 0.03400603625538272 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.329608938547486, |
|
"acc_stderr": 0.01572153107518388, |
|
"acc_norm": 0.329608938547486, |
|
"acc_norm_stderr": 0.01572153107518388 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.66, |
|
"acc_stderr": 0.04760952285695237, |
|
"acc_norm": 0.66, |
|
"acc_norm_stderr": 0.04760952285695237 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.45955882352941174, |
|
"acc_stderr": 0.030273325077345755, |
|
"acc_norm": 0.45955882352941174, |
|
"acc_norm_stderr": 0.030273325077345755 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.5142857142857142, |
|
"acc_stderr": 0.03199615232806286, |
|
"acc_norm": 0.5142857142857142, |
|
"acc_norm_stderr": 0.03199615232806286 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.7257383966244726, |
|
"acc_stderr": 0.029041333510598025, |
|
"acc_norm": 0.7257383966244726, |
|
"acc_norm_stderr": 0.029041333510598025 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3898305084745763, |
|
"acc_stderr": 0.012456386619082598, |
|
"acc_norm": 0.3898305084745763, |
|
"acc_norm_stderr": 0.012456386619082598 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.6421568627450981, |
|
"acc_stderr": 0.03364487286088299, |
|
"acc_norm": 0.6421568627450981, |
|
"acc_norm_stderr": 0.03364487286088299 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.6424242424242425, |
|
"acc_stderr": 0.03742597043806587, |
|
"acc_norm": 0.6424242424242425, |
|
"acc_norm_stderr": 0.03742597043806587 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.30599755201958384, |
|
"mc1_stderr": 0.016132229728155055, |
|
"mc2": 0.4675763272729592, |
|
"mc2_stderr": 0.01542512830327087 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.4781582054309327, |
|
"acc_stderr": 0.017173944474294375, |
|
"acc_norm": 0.5312868949232585, |
|
"acc_norm_stderr": 0.017156666859785483 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "DopeorNope/SOLAR_D-v2-10.7B", |
|
"model_sha": "e6ee5ef0a6f5c8c4497f8a439b347d2a540a2b34", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |