{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.24744027303754265, "acc_stderr": 0.01261035266329267, "acc_norm": 0.3873720136518771, "acc_norm_stderr": 0.014235872487909869 }, "harness|ko_hellaswag|10": { "acc": 0.2749452300338578, "acc_stderr": 0.004455741817861901, "acc_norm": 0.3882692690699064, "acc_norm_stderr": 0.00486360363836745 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.2982456140350877, "acc_stderr": 0.03508771929824563, "acc_norm": 0.2982456140350877, "acc_norm_stderr": 0.03508771929824563 }, "harness|ko_mmlu_management|5": { "acc": 0.22330097087378642, "acc_stderr": 0.04123553189891431, "acc_norm": 0.22330097087378642, "acc_norm_stderr": 0.04123553189891431 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.3397190293742018, "acc_stderr": 0.01693639411430165, "acc_norm": 0.3397190293742018, "acc_norm_stderr": 0.01693639411430165 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.34074074074074073, "acc_stderr": 0.040943762699967946, "acc_norm": 0.34074074074074073, "acc_norm_stderr": 0.040943762699967946 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.251063829787234, "acc_stderr": 0.02834696377716245, "acc_norm": 0.251063829787234, "acc_norm_stderr": 0.02834696377716245 }, "harness|ko_mmlu_virology|5": { "acc": 0.19879518072289157, "acc_stderr": 0.031069390260789406, "acc_norm": 0.19879518072289157, "acc_norm_stderr": 0.031069390260789406 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.3440514469453376, "acc_stderr": 0.026981478043648026, "acc_norm": 0.3440514469453376, "acc_norm_stderr": 0.026981478043648026 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.27802690582959644, "acc_stderr": 0.03006958487449405, "acc_norm": 0.27802690582959644, "acc_norm_stderr": 0.03006958487449405 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.24427480916030533, "acc_stderr": 0.037683359597287434, "acc_norm": 0.24427480916030533, "acc_norm_stderr": 0.037683359597287434 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.23, "acc_stderr": 0.04229525846816505, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816505 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.30303030303030304, "acc_stderr": 0.032742879140268674, "acc_norm": 0.30303030303030304, "acc_norm_stderr": 0.032742879140268674 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.296551724137931, "acc_stderr": 0.03806142687309994, "acc_norm": 0.296551724137931, "acc_norm_stderr": 0.03806142687309994 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.17647058823529413, "acc_stderr": 0.0379328118530781, "acc_norm": 0.17647058823529413, "acc_norm_stderr": 0.0379328118530781 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.25630252100840334, "acc_stderr": 0.028359620870533953, "acc_norm": 0.25630252100840334, "acc_norm_stderr": 0.028359620870533953 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.22564102564102564, "acc_stderr": 0.021193632525148533, "acc_norm": 0.22564102564102564, "acc_norm_stderr": 0.021193632525148533 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.38, "acc_stderr": 0.04878317312145634, "acc_norm": 0.38, "acc_norm_stderr": 0.04878317312145634 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.3055555555555556, "acc_stderr": 0.04453197507374984, "acc_norm": 0.3055555555555556, "acc_norm_stderr": 0.04453197507374984 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3399014778325123, "acc_stderr": 0.033327690684107895, "acc_norm": 0.3399014778325123, "acc_norm_stderr": 0.033327690684107895 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.3032258064516129, "acc_stderr": 0.02614868593067175, "acc_norm": 0.3032258064516129, "acc_norm_stderr": 0.02614868593067175 }, "harness|ko_mmlu_marketing|5": { "acc": 0.36752136752136755, "acc_stderr": 0.031585391577456365, "acc_norm": 0.36752136752136755, "acc_norm_stderr": 0.031585391577456365 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.2943396226415094, "acc_stderr": 0.028049186315695245, "acc_norm": 0.2943396226415094, "acc_norm_stderr": 0.028049186315695245 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.35454545454545455, "acc_stderr": 0.04582004841505417, "acc_norm": 0.35454545454545455, "acc_norm_stderr": 0.04582004841505417 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.23703703703703705, "acc_stderr": 0.02592887613276611, "acc_norm": 0.23703703703703705, "acc_norm_stderr": 0.02592887613276611 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2980132450331126, "acc_stderr": 0.037345356767871984, "acc_norm": 0.2980132450331126, "acc_norm_stderr": 0.037345356767871984 }, "harness|ko_mmlu_sociology|5": { "acc": 0.2885572139303483, "acc_stderr": 0.03203841040213323, "acc_norm": 0.2885572139303483, "acc_norm_stderr": 0.03203841040213323 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.24855491329479767, "acc_stderr": 0.03295304696818318, "acc_norm": 0.24855491329479767, "acc_norm_stderr": 0.03295304696818318 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2275132275132275, "acc_stderr": 0.02159126940782378, "acc_norm": 0.2275132275132275, "acc_norm_stderr": 0.02159126940782378 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.24305555555555555, "acc_stderr": 0.03586879280080342, "acc_norm": 0.24305555555555555, "acc_norm_stderr": 0.03586879280080342 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.2, "acc_stderr": 0.04020151261036845, "acc_norm": 0.2, "acc_norm_stderr": 0.04020151261036845 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.36127167630057805, "acc_stderr": 0.025862201852277906, "acc_norm": 0.36127167630057805, "acc_norm_stderr": 0.025862201852277906 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.3128834355828221, "acc_stderr": 0.036429145782924034, "acc_norm": 0.3128834355828221, "acc_norm_stderr": 0.036429145782924034 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.3487654320987654, "acc_stderr": 0.02651759772446501, "acc_norm": 0.3487654320987654, "acc_norm_stderr": 0.02651759772446501 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.26, "acc_stderr": 0.044084400227680794, "acc_norm": 0.26, "acc_norm_stderr": 0.044084400227680794 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.24870466321243523, "acc_stderr": 0.03119584087770029, "acc_norm": 0.24870466321243523, "acc_norm_stderr": 0.03119584087770029 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2719298245614035, "acc_stderr": 0.04185774424022056, "acc_norm": 0.2719298245614035, "acc_norm_stderr": 0.04185774424022056 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.29174311926605506, "acc_stderr": 0.01948930096887653, "acc_norm": 0.29174311926605506, "acc_norm_stderr": 0.01948930096887653 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.23015873015873015, "acc_stderr": 0.03764950879790604, "acc_norm": 0.23015873015873015, "acc_norm_stderr": 0.03764950879790604 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.32679738562091504, "acc_stderr": 0.026857294663281416, "acc_norm": 0.32679738562091504, "acc_norm_stderr": 0.026857294663281416 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "harness|ko_mmlu_international_law|5": { "acc": 0.38016528925619836, "acc_stderr": 0.04431324501968431, "acc_norm": 0.38016528925619836, "acc_norm_stderr": 0.04431324501968431 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.2894736842105263, "acc_stderr": 0.03690677986137283, "acc_norm": 0.2894736842105263, "acc_norm_stderr": 0.03690677986137283 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.29248366013071897, "acc_stderr": 0.01840341571010979, "acc_norm": 0.29248366013071897, "acc_norm_stderr": 0.01840341571010979 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.29432624113475175, "acc_stderr": 0.027187127011503796, "acc_norm": 0.29432624113475175, "acc_norm_stderr": 0.027187127011503796 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.21428571428571427, "acc_stderr": 0.03894641120044792, "acc_norm": 0.21428571428571427, "acc_norm_stderr": 0.03894641120044792 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.21296296296296297, "acc_stderr": 0.027920963147993666, "acc_norm": 0.21296296296296297, "acc_norm_stderr": 0.027920963147993666 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.1564245810055866, "acc_stderr": 0.012149144539664161, "acc_norm": 0.1564245810055866, "acc_norm_stderr": 0.012149144539664161 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.22, "acc_stderr": 0.04163331998932269, "acc_norm": 0.22, "acc_norm_stderr": 0.04163331998932269 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.21691176470588236, "acc_stderr": 0.02503584522771126, "acc_norm": 0.21691176470588236, "acc_norm_stderr": 0.02503584522771126 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.2571428571428571, "acc_stderr": 0.027979823538744543, "acc_norm": 0.2571428571428571, "acc_norm_stderr": 0.027979823538744543 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.29957805907172996, "acc_stderr": 0.0298180247497531, "acc_norm": 0.29957805907172996, "acc_norm_stderr": 0.0298180247497531 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.2770534550195567, "acc_stderr": 0.011430462443719678, "acc_norm": 0.2770534550195567, "acc_norm_stderr": 0.011430462443719678 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.29901960784313725, "acc_stderr": 0.03213325717373617, "acc_norm": 0.29901960784313725, "acc_norm_stderr": 0.03213325717373617 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.3151515151515151, "acc_stderr": 0.0362773057502241, "acc_norm": 0.3151515151515151, "acc_norm_stderr": 0.0362773057502241 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2692778457772338, "mc1_stderr": 0.01552856663708729, "mc2": 0.5091125601958423, "mc2_stderr": 0.017262344185815827 }, "harness|ko_commongen_v2|2": { "acc": 0.11570247933884298, "acc_stderr": 0.01099728182694215, "acc_norm": 0.3730814639905549, "acc_norm_stderr": 0.01662731827513743 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Danielbrdz/Barcenas-2x10.7b-Korean", "model_sha": "bcc5f0a3e63caac0f7a8459da21379daf5fb1edd", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }