|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.2645051194539249, |
|
"acc_stderr": 0.012889272949313368, |
|
"acc_norm": 0.32849829351535836, |
|
"acc_norm_stderr": 0.013724978465537378 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.32682732523401714, |
|
"acc_stderr": 0.004680949283855315, |
|
"acc_norm": 0.3868751244771958, |
|
"acc_norm_stderr": 0.004860393011974685 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.38596491228070173, |
|
"acc_stderr": 0.03733756969066163, |
|
"acc_norm": 0.38596491228070173, |
|
"acc_norm_stderr": 0.03733756969066163 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.4563106796116505, |
|
"acc_stderr": 0.04931801994220414, |
|
"acc_norm": 0.4563106796116505, |
|
"acc_norm_stderr": 0.04931801994220414 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.38569604086845466, |
|
"acc_stderr": 0.01740647661921291, |
|
"acc_norm": 0.38569604086845466, |
|
"acc_norm_stderr": 0.01740647661921291 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.3037037037037037, |
|
"acc_stderr": 0.039725528847851375, |
|
"acc_norm": 0.3037037037037037, |
|
"acc_norm_stderr": 0.039725528847851375 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.04560480215720683, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.04560480215720683 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.32340425531914896, |
|
"acc_stderr": 0.030579442773610337, |
|
"acc_norm": 0.32340425531914896, |
|
"acc_norm_stderr": 0.030579442773610337 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.3192771084337349, |
|
"acc_stderr": 0.036293353299478595, |
|
"acc_norm": 0.3192771084337349, |
|
"acc_norm_stderr": 0.036293353299478595 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.41479099678456594, |
|
"acc_stderr": 0.027982680459759563, |
|
"acc_norm": 0.41479099678456594, |
|
"acc_norm_stderr": 0.027982680459759563 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.36771300448430494, |
|
"acc_stderr": 0.03236198350928276, |
|
"acc_norm": 0.36771300448430494, |
|
"acc_norm_stderr": 0.03236198350928276 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.3893129770992366, |
|
"acc_stderr": 0.04276486542814591, |
|
"acc_norm": 0.3893129770992366, |
|
"acc_norm_stderr": 0.04276486542814591 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.04943110704237102, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.04943110704237102 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.4444444444444444, |
|
"acc_stderr": 0.03540294377095367, |
|
"acc_norm": 0.4444444444444444, |
|
"acc_norm_stderr": 0.03540294377095367 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.45517241379310347, |
|
"acc_stderr": 0.04149886942192117, |
|
"acc_norm": 0.45517241379310347, |
|
"acc_norm_stderr": 0.04149886942192117 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.27450980392156865, |
|
"acc_stderr": 0.044405219061793254, |
|
"acc_norm": 0.27450980392156865, |
|
"acc_norm_stderr": 0.044405219061793254 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.39915966386554624, |
|
"acc_stderr": 0.03181110032413926, |
|
"acc_norm": 0.39915966386554624, |
|
"acc_norm_stderr": 0.03181110032413926 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.37435897435897436, |
|
"acc_stderr": 0.024537591572830524, |
|
"acc_norm": 0.37435897435897436, |
|
"acc_norm_stderr": 0.024537591572830524 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.51, |
|
"acc_stderr": 0.05024183937956912, |
|
"acc_norm": 0.51, |
|
"acc_norm_stderr": 0.05024183937956912 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.4074074074074074, |
|
"acc_stderr": 0.04750077341199985, |
|
"acc_norm": 0.4074074074074074, |
|
"acc_norm_stderr": 0.04750077341199985 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.3448275862068966, |
|
"acc_stderr": 0.033442837442804574, |
|
"acc_norm": 0.3448275862068966, |
|
"acc_norm_stderr": 0.033442837442804574 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.3774193548387097, |
|
"acc_stderr": 0.027575960723278253, |
|
"acc_norm": 0.3774193548387097, |
|
"acc_norm_stderr": 0.027575960723278253 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6111111111111112, |
|
"acc_stderr": 0.031937057262002924, |
|
"acc_norm": 0.6111111111111112, |
|
"acc_norm_stderr": 0.031937057262002924 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.43018867924528303, |
|
"acc_stderr": 0.03047144586718323, |
|
"acc_norm": 0.43018867924528303, |
|
"acc_norm_stderr": 0.03047144586718323 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.37272727272727274, |
|
"acc_stderr": 0.04631381319425463, |
|
"acc_norm": 0.37272727272727274, |
|
"acc_norm_stderr": 0.04631381319425463 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3037037037037037, |
|
"acc_stderr": 0.02803792996911499, |
|
"acc_norm": 0.3037037037037037, |
|
"acc_norm_stderr": 0.02803792996911499 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2980132450331126, |
|
"acc_stderr": 0.037345356767871984, |
|
"acc_norm": 0.2980132450331126, |
|
"acc_norm_stderr": 0.037345356767871984 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.47761194029850745, |
|
"acc_stderr": 0.035319879302087305, |
|
"acc_norm": 0.47761194029850745, |
|
"acc_norm_stderr": 0.035319879302087305 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.3699421965317919, |
|
"acc_stderr": 0.03681229633394319, |
|
"acc_norm": 0.3699421965317919, |
|
"acc_norm_stderr": 0.03681229633394319 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.31216931216931215, |
|
"acc_stderr": 0.023865206836972602, |
|
"acc_norm": 0.31216931216931215, |
|
"acc_norm_stderr": 0.023865206836972602 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3125, |
|
"acc_stderr": 0.038760854559127644, |
|
"acc_norm": 0.3125, |
|
"acc_norm_stderr": 0.038760854559127644 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.047937248544110196, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.047937248544110196 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.51, |
|
"acc_stderr": 0.05024183937956911, |
|
"acc_norm": 0.51, |
|
"acc_norm_stderr": 0.05024183937956911 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.4046242774566474, |
|
"acc_stderr": 0.02642481659400985, |
|
"acc_norm": 0.4046242774566474, |
|
"acc_norm_stderr": 0.02642481659400985 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.37423312883435583, |
|
"acc_stderr": 0.038020681028996146, |
|
"acc_norm": 0.37423312883435583, |
|
"acc_norm_stderr": 0.038020681028996146 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.3271604938271605, |
|
"acc_stderr": 0.026105673861409814, |
|
"acc_norm": 0.3271604938271605, |
|
"acc_norm_stderr": 0.026105673861409814 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.41968911917098445, |
|
"acc_stderr": 0.03561587327685884, |
|
"acc_norm": 0.41968911917098445, |
|
"acc_norm_stderr": 0.03561587327685884 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.22807017543859648, |
|
"acc_stderr": 0.03947152782669415, |
|
"acc_norm": 0.22807017543859648, |
|
"acc_norm_stderr": 0.03947152782669415 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.3834862385321101, |
|
"acc_stderr": 0.020847156641915984, |
|
"acc_norm": 0.3834862385321101, |
|
"acc_norm_stderr": 0.020847156641915984 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.42857142857142855, |
|
"acc_stderr": 0.0442626668137991, |
|
"acc_norm": 0.42857142857142855, |
|
"acc_norm_stderr": 0.0442626668137991 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.4477124183006536, |
|
"acc_stderr": 0.028472938478033526, |
|
"acc_norm": 0.4477124183006536, |
|
"acc_norm_stderr": 0.028472938478033526 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.049236596391733084, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.049236596391733084 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.5371900826446281, |
|
"acc_stderr": 0.04551711196104218, |
|
"acc_norm": 0.5371900826446281, |
|
"acc_norm_stderr": 0.04551711196104218 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.3223684210526316, |
|
"acc_stderr": 0.03803510248351585, |
|
"acc_norm": 0.3223684210526316, |
|
"acc_norm_stderr": 0.03803510248351585 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.3055555555555556, |
|
"acc_stderr": 0.01863559403442397, |
|
"acc_norm": 0.3055555555555556, |
|
"acc_norm_stderr": 0.01863559403442397 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3262411347517731, |
|
"acc_stderr": 0.027968453043563168, |
|
"acc_norm": 0.3262411347517731, |
|
"acc_norm_stderr": 0.027968453043563168 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.30357142857142855, |
|
"acc_stderr": 0.043642261558410445, |
|
"acc_norm": 0.30357142857142855, |
|
"acc_norm_stderr": 0.043642261558410445 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.41203703703703703, |
|
"acc_stderr": 0.03356787758160835, |
|
"acc_norm": 0.41203703703703703, |
|
"acc_norm_stderr": 0.03356787758160835 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.31843575418994413, |
|
"acc_stderr": 0.015581008080360274, |
|
"acc_norm": 0.31843575418994413, |
|
"acc_norm_stderr": 0.015581008080360274 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.44, |
|
"acc_stderr": 0.04988876515698589, |
|
"acc_norm": 0.44, |
|
"acc_norm_stderr": 0.04988876515698589 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.38235294117647056, |
|
"acc_stderr": 0.029520095697687758, |
|
"acc_norm": 0.38235294117647056, |
|
"acc_norm_stderr": 0.029520095697687758 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.44081632653061226, |
|
"acc_stderr": 0.03178419114175363, |
|
"acc_norm": 0.44081632653061226, |
|
"acc_norm_stderr": 0.03178419114175363 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.47257383966244726, |
|
"acc_stderr": 0.032498227183013026, |
|
"acc_norm": 0.47257383966244726, |
|
"acc_norm_stderr": 0.032498227183013026 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.2966101694915254, |
|
"acc_stderr": 0.011665946586082868, |
|
"acc_norm": 0.2966101694915254, |
|
"acc_norm_stderr": 0.011665946586082868 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.3235294117647059, |
|
"acc_stderr": 0.03283472056108567, |
|
"acc_norm": 0.3235294117647059, |
|
"acc_norm_stderr": 0.03283472056108567 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.3515151515151515, |
|
"acc_stderr": 0.0372820699868265, |
|
"acc_norm": 0.3515151515151515, |
|
"acc_norm_stderr": 0.0372820699868265 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.3243574051407589, |
|
"mc1_stderr": 0.01638797677964793, |
|
"mc2": 0.49917419306073907, |
|
"mc2_stderr": 0.016202138687957245 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.2857142857142857, |
|
"acc_stderr": 0.015531620786986743, |
|
"acc_norm": 0.3565525383707202, |
|
"acc_norm_stderr": 0.016467706981527445 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "mistralai/Mistral-7B-Instruct-v0.1", |
|
"model_sha": "7ad5799710574ba1c1d953eba3077af582f3a773", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |