|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3455631399317406, |
|
"acc_stderr": 0.013896938461145687, |
|
"acc_norm": 0.38139931740614336, |
|
"acc_norm_stderr": 0.014194389086685265 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3619796853216491, |
|
"acc_stderr": 0.004795908282584555, |
|
"acc_norm": 0.44761999601672975, |
|
"acc_norm_stderr": 0.0049623252978409915 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.47368421052631576, |
|
"acc_stderr": 0.03829509868994727, |
|
"acc_norm": 0.47368421052631576, |
|
"acc_norm_stderr": 0.03829509868994727 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5436893203883495, |
|
"acc_stderr": 0.049318019942204146, |
|
"acc_norm": 0.5436893203883495, |
|
"acc_norm_stderr": 0.049318019942204146 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.4725415070242657, |
|
"acc_stderr": 0.01785298126663395, |
|
"acc_norm": 0.4725415070242657, |
|
"acc_norm_stderr": 0.01785298126663395 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.37777777777777777, |
|
"acc_stderr": 0.04188307537595853, |
|
"acc_norm": 0.37777777777777777, |
|
"acc_norm_stderr": 0.04188307537595853 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.04560480215720684, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.04560480215720684 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.39148936170212767, |
|
"acc_stderr": 0.03190701242326812, |
|
"acc_norm": 0.39148936170212767, |
|
"acc_norm_stderr": 0.03190701242326812 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.4397590361445783, |
|
"acc_stderr": 0.03864139923699122, |
|
"acc_norm": 0.4397590361445783, |
|
"acc_norm_stderr": 0.03864139923699122 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.4855305466237942, |
|
"acc_stderr": 0.028386198084177673, |
|
"acc_norm": 0.4855305466237942, |
|
"acc_norm_stderr": 0.028386198084177673 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.4170403587443946, |
|
"acc_stderr": 0.03309266936071722, |
|
"acc_norm": 0.4170403587443946, |
|
"acc_norm_stderr": 0.03309266936071722 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.4580152671755725, |
|
"acc_stderr": 0.04369802690578756, |
|
"acc_norm": 0.4580152671755725, |
|
"acc_norm_stderr": 0.04369802690578756 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5151515151515151, |
|
"acc_stderr": 0.03560716516531061, |
|
"acc_norm": 0.5151515151515151, |
|
"acc_norm_stderr": 0.03560716516531061 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.4068965517241379, |
|
"acc_stderr": 0.040937939812662374, |
|
"acc_norm": 0.4068965517241379, |
|
"acc_norm_stderr": 0.040937939812662374 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.23529411764705882, |
|
"acc_stderr": 0.04220773659171452, |
|
"acc_norm": 0.23529411764705882, |
|
"acc_norm_stderr": 0.04220773659171452 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.4579831932773109, |
|
"acc_stderr": 0.03236361111951941, |
|
"acc_norm": 0.4579831932773109, |
|
"acc_norm_stderr": 0.03236361111951941 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.37435897435897436, |
|
"acc_stderr": 0.024537591572830524, |
|
"acc_norm": 0.37435897435897436, |
|
"acc_norm_stderr": 0.024537591572830524 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.52, |
|
"acc_stderr": 0.05021167315686779, |
|
"acc_norm": 0.52, |
|
"acc_norm_stderr": 0.05021167315686779 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621504, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621504 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.4444444444444444, |
|
"acc_stderr": 0.04803752235190193, |
|
"acc_norm": 0.4444444444444444, |
|
"acc_norm_stderr": 0.04803752235190193 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.43842364532019706, |
|
"acc_stderr": 0.03491207857486518, |
|
"acc_norm": 0.43842364532019706, |
|
"acc_norm_stderr": 0.03491207857486518 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.45161290322580644, |
|
"acc_stderr": 0.028310500348568385, |
|
"acc_norm": 0.45161290322580644, |
|
"acc_norm_stderr": 0.028310500348568385 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6752136752136753, |
|
"acc_stderr": 0.03067902276549883, |
|
"acc_norm": 0.6752136752136753, |
|
"acc_norm_stderr": 0.03067902276549883 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.43018867924528303, |
|
"acc_stderr": 0.030471445867183235, |
|
"acc_norm": 0.43018867924528303, |
|
"acc_norm_stderr": 0.030471445867183235 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.04789131426105757, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.04789131426105757 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.32592592592592595, |
|
"acc_stderr": 0.02857834836547308, |
|
"acc_norm": 0.32592592592592595, |
|
"acc_norm_stderr": 0.02857834836547308 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.23178807947019867, |
|
"acc_stderr": 0.03445406271987053, |
|
"acc_norm": 0.23178807947019867, |
|
"acc_norm_stderr": 0.03445406271987053 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.5522388059701493, |
|
"acc_stderr": 0.03516184772952167, |
|
"acc_norm": 0.5522388059701493, |
|
"acc_norm_stderr": 0.03516184772952167 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.3468208092485549, |
|
"acc_stderr": 0.03629146670159665, |
|
"acc_norm": 0.3468208092485549, |
|
"acc_norm_stderr": 0.03629146670159665 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.37566137566137564, |
|
"acc_stderr": 0.02494236893115979, |
|
"acc_norm": 0.37566137566137564, |
|
"acc_norm_stderr": 0.02494236893115979 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3194444444444444, |
|
"acc_stderr": 0.038990736873573344, |
|
"acc_norm": 0.3194444444444444, |
|
"acc_norm_stderr": 0.038990736873573344 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.04824181513244218, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.04824181513244218 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.58, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.58, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.026919095102908273, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.026919095102908273 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.4233128834355828, |
|
"acc_stderr": 0.03881891213334383, |
|
"acc_norm": 0.4233128834355828, |
|
"acc_norm_stderr": 0.03881891213334383 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.4228395061728395, |
|
"acc_stderr": 0.02748747298087159, |
|
"acc_norm": 0.4228395061728395, |
|
"acc_norm_stderr": 0.02748747298087159 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.045604802157206845, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.045604802157206845 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.46113989637305697, |
|
"acc_stderr": 0.03597524411734577, |
|
"acc_norm": 0.46113989637305697, |
|
"acc_norm_stderr": 0.03597524411734577 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2807017543859649, |
|
"acc_stderr": 0.042270544512321984, |
|
"acc_norm": 0.2807017543859649, |
|
"acc_norm_stderr": 0.042270544512321984 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.45871559633027525, |
|
"acc_stderr": 0.021364122533881695, |
|
"acc_norm": 0.45871559633027525, |
|
"acc_norm_stderr": 0.021364122533881695 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.2857142857142857, |
|
"acc_stderr": 0.0404061017820884, |
|
"acc_norm": 0.2857142857142857, |
|
"acc_norm_stderr": 0.0404061017820884 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.46405228758169936, |
|
"acc_stderr": 0.028555827516528787, |
|
"acc_norm": 0.46405228758169936, |
|
"acc_norm_stderr": 0.028555827516528787 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.5785123966942148, |
|
"acc_stderr": 0.04507732278775087, |
|
"acc_norm": 0.5785123966942148, |
|
"acc_norm_stderr": 0.04507732278775087 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.40131578947368424, |
|
"acc_stderr": 0.03988903703336285, |
|
"acc_norm": 0.40131578947368424, |
|
"acc_norm_stderr": 0.03988903703336285 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.3627450980392157, |
|
"acc_stderr": 0.01945076843250551, |
|
"acc_norm": 0.3627450980392157, |
|
"acc_norm_stderr": 0.01945076843250551 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3404255319148936, |
|
"acc_stderr": 0.02826765748265014, |
|
"acc_norm": 0.3404255319148936, |
|
"acc_norm_stderr": 0.02826765748265014 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.35714285714285715, |
|
"acc_stderr": 0.04547960999764376, |
|
"acc_norm": 0.35714285714285715, |
|
"acc_norm_stderr": 0.04547960999764376 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.4027777777777778, |
|
"acc_stderr": 0.033448873829978666, |
|
"acc_norm": 0.4027777777777778, |
|
"acc_norm_stderr": 0.033448873829978666 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2335195530726257, |
|
"acc_stderr": 0.014149575348976276, |
|
"acc_norm": 0.2335195530726257, |
|
"acc_norm_stderr": 0.014149575348976276 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.048523658709391, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.048523658709391 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.52, |
|
"acc_stderr": 0.050211673156867795, |
|
"acc_norm": 0.52, |
|
"acc_norm_stderr": 0.050211673156867795 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.4007352941176471, |
|
"acc_stderr": 0.029768263528933105, |
|
"acc_norm": 0.4007352941176471, |
|
"acc_norm_stderr": 0.029768263528933105 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.49387755102040815, |
|
"acc_stderr": 0.03200682020163908, |
|
"acc_norm": 0.49387755102040815, |
|
"acc_norm_stderr": 0.03200682020163908 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.5443037974683544, |
|
"acc_stderr": 0.032419206846933335, |
|
"acc_norm": 0.5443037974683544, |
|
"acc_norm_stderr": 0.032419206846933335 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3155149934810952, |
|
"acc_stderr": 0.011869184843058642, |
|
"acc_norm": 0.3155149934810952, |
|
"acc_norm_stderr": 0.011869184843058642 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.44607843137254904, |
|
"acc_stderr": 0.034888454513049734, |
|
"acc_norm": 0.44607843137254904, |
|
"acc_norm_stderr": 0.034888454513049734 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.41818181818181815, |
|
"acc_stderr": 0.038517163193983954, |
|
"acc_norm": 0.41818181818181815, |
|
"acc_norm_stderr": 0.038517163193983954 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.29498164014687883, |
|
"mc1_stderr": 0.015964400965589667, |
|
"mc2": 0.46872875951621523, |
|
"mc2_stderr": 0.01631020915826667 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.3624557260920897, |
|
"acc_stderr": 0.016527131240453716, |
|
"acc_norm": 0.3825265643447462, |
|
"acc_norm_stderr": 0.016709165387228817 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "MNC-Jihun/Mistral-7B-OP-u1k-ver0.6", |
|
"model_sha": "23c7a5ec9de97c7c729fb2d9dc76bba8f6cb3406", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |