|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.4197952218430034, |
|
"acc_stderr": 0.01442218122630303, |
|
"acc_norm": 0.48378839590443684, |
|
"acc_norm_stderr": 0.014603708567414947 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.4303923521210914, |
|
"acc_stderr": 0.0049411916073179105, |
|
"acc_norm": 0.5866361282613025, |
|
"acc_norm_stderr": 0.004914305798575694 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.6608187134502924, |
|
"acc_stderr": 0.03631053496488905, |
|
"acc_norm": 0.6608187134502924, |
|
"acc_norm_stderr": 0.03631053496488905 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6310679611650486, |
|
"acc_stderr": 0.0477761518115674, |
|
"acc_norm": 0.6310679611650486, |
|
"acc_norm_stderr": 0.0477761518115674 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.6564495530012772, |
|
"acc_stderr": 0.01698214563265247, |
|
"acc_norm": 0.6564495530012772, |
|
"acc_norm_stderr": 0.01698214563265247 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4148148148148148, |
|
"acc_stderr": 0.04256193767901406, |
|
"acc_norm": 0.4148148148148148, |
|
"acc_norm_stderr": 0.04256193767901406 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.49361702127659574, |
|
"acc_stderr": 0.032683358999363386, |
|
"acc_norm": 0.49361702127659574, |
|
"acc_norm_stderr": 0.032683358999363386 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.4819277108433735, |
|
"acc_stderr": 0.03889951252827216, |
|
"acc_norm": 0.4819277108433735, |
|
"acc_norm_stderr": 0.03889951252827216 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.6045016077170418, |
|
"acc_stderr": 0.027770918531427838, |
|
"acc_norm": 0.6045016077170418, |
|
"acc_norm_stderr": 0.027770918531427838 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.547085201793722, |
|
"acc_stderr": 0.03340867501923324, |
|
"acc_norm": 0.547085201793722, |
|
"acc_norm_stderr": 0.03340867501923324 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.6335877862595419, |
|
"acc_stderr": 0.04225875451969638, |
|
"acc_norm": 0.6335877862595419, |
|
"acc_norm_stderr": 0.04225875451969638 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.47, |
|
"acc_stderr": 0.05016135580465919, |
|
"acc_norm": 0.47, |
|
"acc_norm_stderr": 0.05016135580465919 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.7424242424242424, |
|
"acc_stderr": 0.03115626951964684, |
|
"acc_norm": 0.7424242424242424, |
|
"acc_norm_stderr": 0.03115626951964684 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.5310344827586206, |
|
"acc_stderr": 0.04158632762097828, |
|
"acc_norm": 0.5310344827586206, |
|
"acc_norm_stderr": 0.04158632762097828 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.3137254901960784, |
|
"acc_stderr": 0.04617034827006716, |
|
"acc_norm": 0.3137254901960784, |
|
"acc_norm_stderr": 0.04617034827006716 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.5840336134453782, |
|
"acc_stderr": 0.0320165010073961, |
|
"acc_norm": 0.5840336134453782, |
|
"acc_norm_stderr": 0.0320165010073961 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.5025641025641026, |
|
"acc_stderr": 0.025350672979412184, |
|
"acc_norm": 0.5025641025641026, |
|
"acc_norm_stderr": 0.025350672979412184 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.6, |
|
"acc_stderr": 0.049236596391733084, |
|
"acc_norm": 0.6, |
|
"acc_norm_stderr": 0.049236596391733084 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.047258156262526045, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6296296296296297, |
|
"acc_stderr": 0.04668408033024931, |
|
"acc_norm": 0.6296296296296297, |
|
"acc_norm_stderr": 0.04668408033024931 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.37438423645320196, |
|
"acc_stderr": 0.03405155380561952, |
|
"acc_norm": 0.37438423645320196, |
|
"acc_norm_stderr": 0.03405155380561952 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.6225806451612903, |
|
"acc_stderr": 0.027575960723278243, |
|
"acc_norm": 0.6225806451612903, |
|
"acc_norm_stderr": 0.027575960723278243 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.811965811965812, |
|
"acc_stderr": 0.02559819368665225, |
|
"acc_norm": 0.811965811965812, |
|
"acc_norm_stderr": 0.02559819368665225 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.5358490566037736, |
|
"acc_stderr": 0.030693675018458003, |
|
"acc_norm": 0.5358490566037736, |
|
"acc_norm_stderr": 0.030693675018458003 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5909090909090909, |
|
"acc_stderr": 0.04709306978661895, |
|
"acc_norm": 0.5909090909090909, |
|
"acc_norm_stderr": 0.04709306978661895 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.028742040903948485, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.028742040903948485 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.36423841059602646, |
|
"acc_stderr": 0.03929111781242741, |
|
"acc_norm": 0.36423841059602646, |
|
"acc_norm_stderr": 0.03929111781242741 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6915422885572139, |
|
"acc_stderr": 0.03265819588512697, |
|
"acc_norm": 0.6915422885572139, |
|
"acc_norm_stderr": 0.03265819588512697 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.5086705202312138, |
|
"acc_stderr": 0.038118909889404126, |
|
"acc_norm": 0.5086705202312138, |
|
"acc_norm_stderr": 0.038118909889404126 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.36507936507936506, |
|
"acc_stderr": 0.024796060602699947, |
|
"acc_norm": 0.36507936507936506, |
|
"acc_norm_stderr": 0.024796060602699947 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4722222222222222, |
|
"acc_stderr": 0.04174752578923185, |
|
"acc_norm": 0.4722222222222222, |
|
"acc_norm_stderr": 0.04174752578923185 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.04824181513244218, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.04824181513244218 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.72, |
|
"acc_stderr": 0.04512608598542127, |
|
"acc_norm": 0.72, |
|
"acc_norm_stderr": 0.04512608598542127 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.6040462427745664, |
|
"acc_stderr": 0.02632981334194624, |
|
"acc_norm": 0.6040462427745664, |
|
"acc_norm_stderr": 0.02632981334194624 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.5398773006134969, |
|
"acc_stderr": 0.03915857291436972, |
|
"acc_norm": 0.5398773006134969, |
|
"acc_norm_stderr": 0.03915857291436972 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.6203703703703703, |
|
"acc_stderr": 0.027002521034516478, |
|
"acc_norm": 0.6203703703703703, |
|
"acc_norm_stderr": 0.027002521034516478 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.0479372485441102, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.0479372485441102 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.7253886010362695, |
|
"acc_stderr": 0.03221024508041154, |
|
"acc_norm": 0.7253886010362695, |
|
"acc_norm_stderr": 0.03221024508041154 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.43859649122807015, |
|
"acc_stderr": 0.04668000738510455, |
|
"acc_norm": 0.43859649122807015, |
|
"acc_norm_stderr": 0.04668000738510455 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.6385321100917432, |
|
"acc_stderr": 0.02059808200993737, |
|
"acc_norm": 0.6385321100917432, |
|
"acc_norm_stderr": 0.02059808200993737 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.3492063492063492, |
|
"acc_stderr": 0.04263906892795133, |
|
"acc_norm": 0.3492063492063492, |
|
"acc_norm_stderr": 0.04263906892795133 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5359477124183006, |
|
"acc_stderr": 0.028555827516528777, |
|
"acc_norm": 0.5359477124183006, |
|
"acc_norm_stderr": 0.028555827516528777 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.55, |
|
"acc_stderr": 0.049999999999999996, |
|
"acc_norm": 0.55, |
|
"acc_norm_stderr": 0.049999999999999996 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.7272727272727273, |
|
"acc_stderr": 0.04065578140908705, |
|
"acc_norm": 0.7272727272727273, |
|
"acc_norm_stderr": 0.04065578140908705 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5723684210526315, |
|
"acc_stderr": 0.04026097083296564, |
|
"acc_norm": 0.5723684210526315, |
|
"acc_norm_stderr": 0.04026097083296564 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.477124183006536, |
|
"acc_stderr": 0.020206653187884782, |
|
"acc_norm": 0.477124183006536, |
|
"acc_norm_stderr": 0.020206653187884782 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.40070921985815605, |
|
"acc_stderr": 0.029233465745573083, |
|
"acc_norm": 0.40070921985815605, |
|
"acc_norm_stderr": 0.029233465745573083 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.38392857142857145, |
|
"acc_stderr": 0.046161430750285455, |
|
"acc_norm": 0.38392857142857145, |
|
"acc_norm_stderr": 0.046161430750285455 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.5231481481481481, |
|
"acc_stderr": 0.034063153607115065, |
|
"acc_norm": 0.5231481481481481, |
|
"acc_norm_stderr": 0.034063153607115065 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.19553072625698323, |
|
"acc_stderr": 0.013264579220945106, |
|
"acc_norm": 0.19553072625698323, |
|
"acc_norm_stderr": 0.013264579220945106 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.46, |
|
"acc_stderr": 0.05009082659620332, |
|
"acc_norm": 0.46, |
|
"acc_norm_stderr": 0.05009082659620332 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.68, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.68, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.5147058823529411, |
|
"acc_stderr": 0.03035969707904612, |
|
"acc_norm": 0.5147058823529411, |
|
"acc_norm_stderr": 0.03035969707904612 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.6326530612244898, |
|
"acc_stderr": 0.03086214492108757, |
|
"acc_norm": 0.6326530612244898, |
|
"acc_norm_stderr": 0.03086214492108757 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.7088607594936709, |
|
"acc_stderr": 0.029571601065753374, |
|
"acc_norm": 0.7088607594936709, |
|
"acc_norm_stderr": 0.029571601065753374 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.4041720990873533, |
|
"acc_stderr": 0.012533504046491367, |
|
"acc_norm": 0.4041720990873533, |
|
"acc_norm_stderr": 0.012533504046491367 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.6421568627450981, |
|
"acc_stderr": 0.03364487286088298, |
|
"acc_norm": 0.6421568627450981, |
|
"acc_norm_stderr": 0.03364487286088298 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.6545454545454545, |
|
"acc_stderr": 0.03713158067481914, |
|
"acc_norm": 0.6545454545454545, |
|
"acc_norm_stderr": 0.03713158067481914 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2533659730722154, |
|
"mc1_stderr": 0.015225899340826837, |
|
"mc2": 0.4116463111717996, |
|
"mc2_stderr": 0.014902796745251457 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.43565525383707204, |
|
"acc_stderr": 0.017047415229476327, |
|
"acc_norm": 0.4935064935064935, |
|
"acc_norm_stderr": 0.01718890435907731 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "AIdenU/SOLAR-10.7b-ko-Y24_v1.0", |
|
"model_sha": "12af074c34713e89135226b12f6f59d2036234c4", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |