|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.48378839590443684, |
|
"acc_stderr": 0.014603708567414947, |
|
"acc_norm": 0.5460750853242321, |
|
"acc_norm_stderr": 0.014549221105171867 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.45956980681139215, |
|
"acc_stderr": 0.00497344206074162, |
|
"acc_norm": 0.6365265883290181, |
|
"acc_norm_stderr": 0.004800164434233249 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.5614035087719298, |
|
"acc_stderr": 0.038057975055904594, |
|
"acc_norm": 0.5614035087719298, |
|
"acc_norm_stderr": 0.038057975055904594 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6310679611650486, |
|
"acc_stderr": 0.0477761518115674, |
|
"acc_norm": 0.6310679611650486, |
|
"acc_norm_stderr": 0.0477761518115674 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.6513409961685823, |
|
"acc_stderr": 0.01704124314349098, |
|
"acc_norm": 0.6513409961685823, |
|
"acc_norm_stderr": 0.01704124314349098 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4666666666666667, |
|
"acc_stderr": 0.043097329010363554, |
|
"acc_norm": 0.4666666666666667, |
|
"acc_norm_stderr": 0.043097329010363554 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.19, |
|
"acc_stderr": 0.03942772444036623, |
|
"acc_norm": 0.19, |
|
"acc_norm_stderr": 0.03942772444036623 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.46382978723404256, |
|
"acc_stderr": 0.032600385118357715, |
|
"acc_norm": 0.46382978723404256, |
|
"acc_norm_stderr": 0.032600385118357715 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.46987951807228917, |
|
"acc_stderr": 0.03885425420866766, |
|
"acc_norm": 0.46987951807228917, |
|
"acc_norm_stderr": 0.03885425420866766 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.6237942122186495, |
|
"acc_stderr": 0.027513925683549427, |
|
"acc_norm": 0.6237942122186495, |
|
"acc_norm_stderr": 0.027513925683549427 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5874439461883408, |
|
"acc_stderr": 0.03304062175449297, |
|
"acc_norm": 0.5874439461883408, |
|
"acc_norm_stderr": 0.03304062175449297 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.6183206106870229, |
|
"acc_stderr": 0.0426073515764456, |
|
"acc_norm": 0.6183206106870229, |
|
"acc_norm_stderr": 0.0426073515764456 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.49, |
|
"acc_stderr": 0.05024183937956914, |
|
"acc_norm": 0.49, |
|
"acc_norm_stderr": 0.05024183937956914 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.6868686868686869, |
|
"acc_stderr": 0.033042050878136525, |
|
"acc_norm": 0.6868686868686869, |
|
"acc_norm_stderr": 0.033042050878136525 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.4482758620689655, |
|
"acc_stderr": 0.04144311810878151, |
|
"acc_norm": 0.4482758620689655, |
|
"acc_norm_stderr": 0.04144311810878151 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.29411764705882354, |
|
"acc_stderr": 0.04533838195929777, |
|
"acc_norm": 0.29411764705882354, |
|
"acc_norm_stderr": 0.04533838195929777 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.6008403361344538, |
|
"acc_stderr": 0.031811100324139245, |
|
"acc_norm": 0.6008403361344538, |
|
"acc_norm_stderr": 0.031811100324139245 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.5794871794871795, |
|
"acc_stderr": 0.025028610276710862, |
|
"acc_norm": 0.5794871794871795, |
|
"acc_norm_stderr": 0.025028610276710862 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.61, |
|
"acc_stderr": 0.04902071300001974, |
|
"acc_norm": 0.61, |
|
"acc_norm_stderr": 0.04902071300001974 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.047937248544110196, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.047937248544110196 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6296296296296297, |
|
"acc_stderr": 0.04668408033024931, |
|
"acc_norm": 0.6296296296296297, |
|
"acc_norm_stderr": 0.04668408033024931 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.3842364532019704, |
|
"acc_stderr": 0.03422398565657551, |
|
"acc_norm": 0.3842364532019704, |
|
"acc_norm_stderr": 0.03422398565657551 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.5774193548387097, |
|
"acc_stderr": 0.02810096472427264, |
|
"acc_norm": 0.5774193548387097, |
|
"acc_norm_stderr": 0.02810096472427264 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7692307692307693, |
|
"acc_stderr": 0.027601921381417618, |
|
"acc_norm": 0.7692307692307693, |
|
"acc_norm_stderr": 0.027601921381417618 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.5584905660377358, |
|
"acc_stderr": 0.03056159042673184, |
|
"acc_norm": 0.5584905660377358, |
|
"acc_norm_stderr": 0.03056159042673184 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.6, |
|
"acc_stderr": 0.0469237132203465, |
|
"acc_norm": 0.6, |
|
"acc_norm_stderr": 0.0469237132203465 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3148148148148148, |
|
"acc_stderr": 0.028317533496066496, |
|
"acc_norm": 0.3148148148148148, |
|
"acc_norm_stderr": 0.028317533496066496 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.32450331125827814, |
|
"acc_stderr": 0.03822746937658754, |
|
"acc_norm": 0.32450331125827814, |
|
"acc_norm_stderr": 0.03822746937658754 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.736318407960199, |
|
"acc_stderr": 0.03115715086935558, |
|
"acc_norm": 0.736318407960199, |
|
"acc_norm_stderr": 0.03115715086935558 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.5317919075144508, |
|
"acc_stderr": 0.03804749744364764, |
|
"acc_norm": 0.5317919075144508, |
|
"acc_norm_stderr": 0.03804749744364764 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.41798941798941797, |
|
"acc_stderr": 0.02540255550326091, |
|
"acc_norm": 0.41798941798941797, |
|
"acc_norm_stderr": 0.02540255550326091 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4722222222222222, |
|
"acc_stderr": 0.04174752578923185, |
|
"acc_norm": 0.4722222222222222, |
|
"acc_norm_stderr": 0.04174752578923185 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.67, |
|
"acc_stderr": 0.047258156262526066, |
|
"acc_norm": 0.67, |
|
"acc_norm_stderr": 0.047258156262526066 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5780346820809249, |
|
"acc_stderr": 0.02658923114217426, |
|
"acc_norm": 0.5780346820809249, |
|
"acc_norm_stderr": 0.02658923114217426 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.5153374233128835, |
|
"acc_stderr": 0.039265223787088445, |
|
"acc_norm": 0.5153374233128835, |
|
"acc_norm_stderr": 0.039265223787088445 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.6049382716049383, |
|
"acc_stderr": 0.02720111766692565, |
|
"acc_norm": 0.6049382716049383, |
|
"acc_norm_stderr": 0.02720111766692565 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.04560480215720684, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.04560480215720684 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.7202072538860104, |
|
"acc_stderr": 0.032396370467357036, |
|
"acc_norm": 0.7202072538860104, |
|
"acc_norm_stderr": 0.032396370467357036 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.4298245614035088, |
|
"acc_stderr": 0.046570472605949646, |
|
"acc_norm": 0.4298245614035088, |
|
"acc_norm_stderr": 0.046570472605949646 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.673394495412844, |
|
"acc_stderr": 0.020106990889937303, |
|
"acc_norm": 0.673394495412844, |
|
"acc_norm_stderr": 0.020106990889937303 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.40476190476190477, |
|
"acc_stderr": 0.043902592653775614, |
|
"acc_norm": 0.40476190476190477, |
|
"acc_norm_stderr": 0.043902592653775614 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5588235294117647, |
|
"acc_stderr": 0.028431095444176643, |
|
"acc_norm": 0.5588235294117647, |
|
"acc_norm_stderr": 0.028431095444176643 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.61, |
|
"acc_stderr": 0.04902071300001975, |
|
"acc_norm": 0.61, |
|
"acc_norm_stderr": 0.04902071300001975 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.7107438016528925, |
|
"acc_stderr": 0.041391127276354626, |
|
"acc_norm": 0.7107438016528925, |
|
"acc_norm_stderr": 0.041391127276354626 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5197368421052632, |
|
"acc_stderr": 0.040657710025626036, |
|
"acc_norm": 0.5197368421052632, |
|
"acc_norm_stderr": 0.040657710025626036 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.5081699346405228, |
|
"acc_stderr": 0.02022513434305727, |
|
"acc_norm": 0.5081699346405228, |
|
"acc_norm_stderr": 0.02022513434305727 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.4078014184397163, |
|
"acc_stderr": 0.029316011776343555, |
|
"acc_norm": 0.4078014184397163, |
|
"acc_norm_stderr": 0.029316011776343555 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.4017857142857143, |
|
"acc_stderr": 0.04653333146973646, |
|
"acc_norm": 0.4017857142857143, |
|
"acc_norm_stderr": 0.04653333146973646 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.5324074074074074, |
|
"acc_stderr": 0.034028015813589656, |
|
"acc_norm": 0.5324074074074074, |
|
"acc_norm_stderr": 0.034028015813589656 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.3329608938547486, |
|
"acc_stderr": 0.015761716178397556, |
|
"acc_norm": 0.3329608938547486, |
|
"acc_norm_stderr": 0.015761716178397556 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.53, |
|
"acc_stderr": 0.05016135580465919, |
|
"acc_norm": 0.53, |
|
"acc_norm_stderr": 0.05016135580465919 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.67, |
|
"acc_stderr": 0.04725815626252607, |
|
"acc_norm": 0.67, |
|
"acc_norm_stderr": 0.04725815626252607 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.5220588235294118, |
|
"acc_stderr": 0.03034326422421352, |
|
"acc_norm": 0.5220588235294118, |
|
"acc_norm_stderr": 0.03034326422421352 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.6040816326530613, |
|
"acc_stderr": 0.03130802899065686, |
|
"acc_norm": 0.6040816326530613, |
|
"acc_norm_stderr": 0.03130802899065686 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.7679324894514767, |
|
"acc_stderr": 0.027479744550808524, |
|
"acc_norm": 0.7679324894514767, |
|
"acc_norm_stderr": 0.027479744550808524 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.4322033898305085, |
|
"acc_stderr": 0.01265229777711497, |
|
"acc_norm": 0.4322033898305085, |
|
"acc_norm_stderr": 0.01265229777711497 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.6666666666666666, |
|
"acc_stderr": 0.03308611113236436, |
|
"acc_norm": 0.6666666666666666, |
|
"acc_norm_stderr": 0.03308611113236436 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.7151515151515152, |
|
"acc_stderr": 0.03524390844511781, |
|
"acc_norm": 0.7151515151515152, |
|
"acc_norm_stderr": 0.03524390844511781 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.46511627906976744, |
|
"mc1_stderr": 0.017460849975873962, |
|
"mc2": 0.6250875981333179, |
|
"mc2_stderr": 0.015880152200294977 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5985832349468713, |
|
"acc_stderr": 0.01685290785872906, |
|
"acc_norm": 0.6127508854781583, |
|
"acc_norm_stderr": 0.01674757799164278 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "JaeyeonKang/CCK-v1.3.0-DPO", |
|
"model_sha": "86818a7076320a0d25d0374b0b6ea096bf4d3404", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |