|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.24744027303754265, |
|
"acc_stderr": 0.01261035266329267, |
|
"acc_norm": 0.3873720136518771, |
|
"acc_norm_stderr": 0.014235872487909869 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.2749452300338578, |
|
"acc_stderr": 0.004455741817861901, |
|
"acc_norm": 0.3882692690699064, |
|
"acc_norm_stderr": 0.00486360363836745 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.2982456140350877, |
|
"acc_stderr": 0.03508771929824563, |
|
"acc_norm": 0.2982456140350877, |
|
"acc_norm_stderr": 0.03508771929824563 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.22330097087378642, |
|
"acc_stderr": 0.04123553189891431, |
|
"acc_norm": 0.22330097087378642, |
|
"acc_norm_stderr": 0.04123553189891431 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.3397190293742018, |
|
"acc_stderr": 0.01693639411430165, |
|
"acc_norm": 0.3397190293742018, |
|
"acc_norm_stderr": 0.01693639411430165 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.34074074074074073, |
|
"acc_stderr": 0.040943762699967946, |
|
"acc_norm": 0.34074074074074073, |
|
"acc_norm_stderr": 0.040943762699967946 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.251063829787234, |
|
"acc_stderr": 0.02834696377716245, |
|
"acc_norm": 0.251063829787234, |
|
"acc_norm_stderr": 0.02834696377716245 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.19879518072289157, |
|
"acc_stderr": 0.031069390260789406, |
|
"acc_norm": 0.19879518072289157, |
|
"acc_norm_stderr": 0.031069390260789406 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.3440514469453376, |
|
"acc_stderr": 0.026981478043648026, |
|
"acc_norm": 0.3440514469453376, |
|
"acc_norm_stderr": 0.026981478043648026 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.27802690582959644, |
|
"acc_stderr": 0.03006958487449405, |
|
"acc_norm": 0.27802690582959644, |
|
"acc_norm_stderr": 0.03006958487449405 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.24427480916030533, |
|
"acc_stderr": 0.037683359597287434, |
|
"acc_norm": 0.24427480916030533, |
|
"acc_norm_stderr": 0.037683359597287434 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.23, |
|
"acc_stderr": 0.04229525846816505, |
|
"acc_norm": 0.23, |
|
"acc_norm_stderr": 0.04229525846816505 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.30303030303030304, |
|
"acc_stderr": 0.032742879140268674, |
|
"acc_norm": 0.30303030303030304, |
|
"acc_norm_stderr": 0.032742879140268674 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.296551724137931, |
|
"acc_stderr": 0.03806142687309994, |
|
"acc_norm": 0.296551724137931, |
|
"acc_norm_stderr": 0.03806142687309994 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.17647058823529413, |
|
"acc_stderr": 0.0379328118530781, |
|
"acc_norm": 0.17647058823529413, |
|
"acc_norm_stderr": 0.0379328118530781 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.25630252100840334, |
|
"acc_stderr": 0.028359620870533953, |
|
"acc_norm": 0.25630252100840334, |
|
"acc_norm_stderr": 0.028359620870533953 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.22564102564102564, |
|
"acc_stderr": 0.021193632525148533, |
|
"acc_norm": 0.22564102564102564, |
|
"acc_norm_stderr": 0.021193632525148533 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.04878317312145634, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.04878317312145634 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.3055555555555556, |
|
"acc_stderr": 0.04453197507374984, |
|
"acc_norm": 0.3055555555555556, |
|
"acc_norm_stderr": 0.04453197507374984 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.3399014778325123, |
|
"acc_stderr": 0.033327690684107895, |
|
"acc_norm": 0.3399014778325123, |
|
"acc_norm_stderr": 0.033327690684107895 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.3032258064516129, |
|
"acc_stderr": 0.02614868593067175, |
|
"acc_norm": 0.3032258064516129, |
|
"acc_norm_stderr": 0.02614868593067175 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.36752136752136755, |
|
"acc_stderr": 0.031585391577456365, |
|
"acc_norm": 0.36752136752136755, |
|
"acc_norm_stderr": 0.031585391577456365 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.2943396226415094, |
|
"acc_stderr": 0.028049186315695245, |
|
"acc_norm": 0.2943396226415094, |
|
"acc_norm_stderr": 0.028049186315695245 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.35454545454545455, |
|
"acc_stderr": 0.04582004841505417, |
|
"acc_norm": 0.35454545454545455, |
|
"acc_norm_stderr": 0.04582004841505417 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.23703703703703705, |
|
"acc_stderr": 0.02592887613276611, |
|
"acc_norm": 0.23703703703703705, |
|
"acc_norm_stderr": 0.02592887613276611 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2980132450331126, |
|
"acc_stderr": 0.037345356767871984, |
|
"acc_norm": 0.2980132450331126, |
|
"acc_norm_stderr": 0.037345356767871984 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.2885572139303483, |
|
"acc_stderr": 0.03203841040213323, |
|
"acc_norm": 0.2885572139303483, |
|
"acc_norm_stderr": 0.03203841040213323 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.24855491329479767, |
|
"acc_stderr": 0.03295304696818318, |
|
"acc_norm": 0.24855491329479767, |
|
"acc_norm_stderr": 0.03295304696818318 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.2275132275132275, |
|
"acc_stderr": 0.02159126940782378, |
|
"acc_norm": 0.2275132275132275, |
|
"acc_norm_stderr": 0.02159126940782378 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.24305555555555555, |
|
"acc_stderr": 0.03586879280080342, |
|
"acc_norm": 0.24305555555555555, |
|
"acc_norm_stderr": 0.03586879280080342 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.04020151261036845, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.04020151261036845 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252604, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252604 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.36127167630057805, |
|
"acc_stderr": 0.025862201852277906, |
|
"acc_norm": 0.36127167630057805, |
|
"acc_norm_stderr": 0.025862201852277906 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.3128834355828221, |
|
"acc_stderr": 0.036429145782924034, |
|
"acc_norm": 0.3128834355828221, |
|
"acc_norm_stderr": 0.036429145782924034 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.3487654320987654, |
|
"acc_stderr": 0.02651759772446501, |
|
"acc_norm": 0.3487654320987654, |
|
"acc_norm_stderr": 0.02651759772446501 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.044084400227680794, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.044084400227680794 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.24870466321243523, |
|
"acc_stderr": 0.03119584087770029, |
|
"acc_norm": 0.24870466321243523, |
|
"acc_norm_stderr": 0.03119584087770029 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2719298245614035, |
|
"acc_stderr": 0.04185774424022056, |
|
"acc_norm": 0.2719298245614035, |
|
"acc_norm_stderr": 0.04185774424022056 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.29174311926605506, |
|
"acc_stderr": 0.01948930096887653, |
|
"acc_norm": 0.29174311926605506, |
|
"acc_norm_stderr": 0.01948930096887653 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.23015873015873015, |
|
"acc_stderr": 0.03764950879790604, |
|
"acc_norm": 0.23015873015873015, |
|
"acc_norm_stderr": 0.03764950879790604 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.32679738562091504, |
|
"acc_stderr": 0.026857294663281416, |
|
"acc_norm": 0.32679738562091504, |
|
"acc_norm_stderr": 0.026857294663281416 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.38016528925619836, |
|
"acc_stderr": 0.04431324501968431, |
|
"acc_norm": 0.38016528925619836, |
|
"acc_norm_stderr": 0.04431324501968431 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.2894736842105263, |
|
"acc_stderr": 0.03690677986137283, |
|
"acc_norm": 0.2894736842105263, |
|
"acc_norm_stderr": 0.03690677986137283 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.29248366013071897, |
|
"acc_stderr": 0.01840341571010979, |
|
"acc_norm": 0.29248366013071897, |
|
"acc_norm_stderr": 0.01840341571010979 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.29432624113475175, |
|
"acc_stderr": 0.027187127011503796, |
|
"acc_norm": 0.29432624113475175, |
|
"acc_norm_stderr": 0.027187127011503796 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.21428571428571427, |
|
"acc_stderr": 0.03894641120044792, |
|
"acc_norm": 0.21428571428571427, |
|
"acc_norm_stderr": 0.03894641120044792 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.21296296296296297, |
|
"acc_stderr": 0.027920963147993666, |
|
"acc_norm": 0.21296296296296297, |
|
"acc_norm_stderr": 0.027920963147993666 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.1564245810055866, |
|
"acc_stderr": 0.012149144539664161, |
|
"acc_norm": 0.1564245810055866, |
|
"acc_norm_stderr": 0.012149144539664161 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.22, |
|
"acc_stderr": 0.04163331998932269, |
|
"acc_norm": 0.22, |
|
"acc_norm_stderr": 0.04163331998932269 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.21691176470588236, |
|
"acc_stderr": 0.02503584522771126, |
|
"acc_norm": 0.21691176470588236, |
|
"acc_norm_stderr": 0.02503584522771126 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.2571428571428571, |
|
"acc_stderr": 0.027979823538744543, |
|
"acc_norm": 0.2571428571428571, |
|
"acc_norm_stderr": 0.027979823538744543 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.29957805907172996, |
|
"acc_stderr": 0.0298180247497531, |
|
"acc_norm": 0.29957805907172996, |
|
"acc_norm_stderr": 0.0298180247497531 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.2770534550195567, |
|
"acc_stderr": 0.011430462443719678, |
|
"acc_norm": 0.2770534550195567, |
|
"acc_norm_stderr": 0.011430462443719678 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.29901960784313725, |
|
"acc_stderr": 0.03213325717373617, |
|
"acc_norm": 0.29901960784313725, |
|
"acc_norm_stderr": 0.03213325717373617 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.3151515151515151, |
|
"acc_stderr": 0.0362773057502241, |
|
"acc_norm": 0.3151515151515151, |
|
"acc_norm_stderr": 0.0362773057502241 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2692778457772338, |
|
"mc1_stderr": 0.01552856663708729, |
|
"mc2": 0.5091125601958423, |
|
"mc2_stderr": 0.017262344185815827 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.11570247933884298, |
|
"acc_stderr": 0.01099728182694215, |
|
"acc_norm": 0.3730814639905549, |
|
"acc_norm_stderr": 0.01662731827513743 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "Danielbrdz/Barcenas-2x10.7b-Korean", |
|
"model_sha": "bcc5f0a3e63caac0f7a8459da21379daf5fb1edd", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |