{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.3199658703071672, "acc_stderr": 0.013631345807016195, "acc_norm": 0.3779863481228669, "acc_norm_stderr": 0.014169664520303101 }, "harness|ko_hellaswag|10": { "acc": 0.3586934873531169, "acc_stderr": 0.004786368011500456, "acc_norm": 0.4553873730332603, "acc_norm_stderr": 0.0049698795328430865 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.39766081871345027, "acc_stderr": 0.0375363895576169, "acc_norm": 0.39766081871345027, "acc_norm_stderr": 0.0375363895576169 }, "harness|ko_mmlu_management|5": { "acc": 0.4368932038834951, "acc_stderr": 0.04911147107365778, "acc_norm": 0.4368932038834951, "acc_norm_stderr": 0.04911147107365778 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.44189016602809705, "acc_stderr": 0.01775880053421442, "acc_norm": 0.44189016602809705, "acc_norm_stderr": 0.01775880053421442 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.35555555555555557, "acc_stderr": 0.04135176749720386, "acc_norm": 0.35555555555555557, "acc_norm_stderr": 0.04135176749720386 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.36, "acc_stderr": 0.048241815132442176, "acc_norm": 0.36, "acc_norm_stderr": 0.048241815132442176 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.35319148936170214, "acc_stderr": 0.03124532520276193, "acc_norm": 0.35319148936170214, "acc_norm_stderr": 0.03124532520276193 }, "harness|ko_mmlu_virology|5": { "acc": 0.3253012048192771, "acc_stderr": 0.03647168523683227, "acc_norm": 0.3253012048192771, "acc_norm_stderr": 0.03647168523683227 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.3890675241157556, "acc_stderr": 0.027690337536485372, "acc_norm": 0.3890675241157556, "acc_norm_stderr": 0.027690337536485372 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.33183856502242154, "acc_stderr": 0.031602951437766785, "acc_norm": 0.33183856502242154, "acc_norm_stderr": 0.031602951437766785 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.5190839694656488, "acc_stderr": 0.04382094705550988, "acc_norm": 0.5190839694656488, "acc_norm_stderr": 0.04382094705550988 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.4595959595959596, "acc_stderr": 0.035507024651313425, "acc_norm": 0.4595959595959596, "acc_norm_stderr": 0.035507024651313425 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.3724137931034483, "acc_stderr": 0.0402873153294756, "acc_norm": 0.3724137931034483, "acc_norm_stderr": 0.0402873153294756 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.21568627450980393, "acc_stderr": 0.04092563958237655, "acc_norm": 0.21568627450980393, "acc_norm_stderr": 0.04092563958237655 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.4369747899159664, "acc_stderr": 0.03221943636566196, "acc_norm": 0.4369747899159664, "acc_norm_stderr": 0.03221943636566196 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.39487179487179486, "acc_stderr": 0.02478431694215637, "acc_norm": 0.39487179487179486, "acc_norm_stderr": 0.02478431694215637 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.43, "acc_stderr": 0.049756985195624284, "acc_norm": 0.43, "acc_norm_stderr": 0.049756985195624284 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.4722222222222222, "acc_stderr": 0.048262172941398944, "acc_norm": 0.4722222222222222, "acc_norm_stderr": 0.048262172941398944 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.30049261083743845, "acc_stderr": 0.03225799476233485, "acc_norm": 0.30049261083743845, "acc_norm_stderr": 0.03225799476233485 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.4064516129032258, "acc_stderr": 0.027941727346256315, "acc_norm": 0.4064516129032258, "acc_norm_stderr": 0.027941727346256315 }, "harness|ko_mmlu_marketing|5": { "acc": 0.5470085470085471, "acc_stderr": 0.03261099873098619, "acc_norm": 0.5470085470085471, "acc_norm_stderr": 0.03261099873098619 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.37735849056603776, "acc_stderr": 0.029832808114796005, "acc_norm": 0.37735849056603776, "acc_norm_stderr": 0.029832808114796005 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.4, "acc_stderr": 0.0469237132203465, "acc_norm": 0.4, "acc_norm_stderr": 0.0469237132203465 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.24814814814814815, "acc_stderr": 0.0263357394040558, "acc_norm": 0.24814814814814815, "acc_norm_stderr": 0.0263357394040558 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.31788079470198677, "acc_stderr": 0.038020397601079024, "acc_norm": 0.31788079470198677, "acc_norm_stderr": 0.038020397601079024 }, "harness|ko_mmlu_sociology|5": { "acc": 0.5174129353233831, "acc_stderr": 0.03533389234739245, "acc_norm": 0.5174129353233831, "acc_norm_stderr": 0.03533389234739245 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.3930635838150289, "acc_stderr": 0.03724249595817731, "acc_norm": 0.3930635838150289, "acc_norm_stderr": 0.03724249595817731 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.29894179894179895, "acc_stderr": 0.02357760479165581, "acc_norm": 0.29894179894179895, "acc_norm_stderr": 0.02357760479165581 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3611111111111111, "acc_stderr": 0.040166600304512336, "acc_norm": 0.3611111111111111, "acc_norm_stderr": 0.040166600304512336 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.42, "acc_stderr": 0.049604496374885836, "acc_norm": 0.42, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.54, "acc_stderr": 0.05009082659620333, "acc_norm": 0.54, "acc_norm_stderr": 0.05009082659620333 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.45375722543352603, "acc_stderr": 0.026803720583206188, "acc_norm": 0.45375722543352603, "acc_norm_stderr": 0.026803720583206188 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.31901840490797545, "acc_stderr": 0.03661997551073836, "acc_norm": 0.31901840490797545, "acc_norm_stderr": 0.03661997551073836 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.3888888888888889, "acc_stderr": 0.027125115513166865, "acc_norm": 0.3888888888888889, "acc_norm_stderr": 0.027125115513166865 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.29, "acc_stderr": 0.045604802157206845, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206845 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.48704663212435234, "acc_stderr": 0.0360722806104775, "acc_norm": 0.48704663212435234, "acc_norm_stderr": 0.0360722806104775 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.23684210526315788, "acc_stderr": 0.039994238792813365, "acc_norm": 0.23684210526315788, "acc_norm_stderr": 0.039994238792813365 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.42752293577981654, "acc_stderr": 0.02121091020430043, "acc_norm": 0.42752293577981654, "acc_norm_stderr": 0.02121091020430043 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.3253968253968254, "acc_stderr": 0.04190596438871136, "acc_norm": 0.3253968253968254, "acc_norm_stderr": 0.04190596438871136 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.4215686274509804, "acc_stderr": 0.028275490156791434, "acc_norm": 0.4215686274509804, "acc_norm_stderr": 0.028275490156791434 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_international_law|5": { "acc": 0.5371900826446281, "acc_stderr": 0.04551711196104218, "acc_norm": 0.5371900826446281, "acc_norm_stderr": 0.04551711196104218 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.3815789473684211, "acc_stderr": 0.03953173377749194, "acc_norm": 0.3815789473684211, "acc_norm_stderr": 0.03953173377749194 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.28921568627450983, "acc_stderr": 0.018342529845275908, "acc_norm": 0.28921568627450983, "acc_norm_stderr": 0.018342529845275908 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.2801418439716312, "acc_stderr": 0.02678917235114023, "acc_norm": 0.2801418439716312, "acc_norm_stderr": 0.02678917235114023 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.19642857142857142, "acc_stderr": 0.03770970049347019, "acc_norm": 0.19642857142857142, "acc_norm_stderr": 0.03770970049347019 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4027777777777778, "acc_stderr": 0.03344887382997866, "acc_norm": 0.4027777777777778, "acc_norm_stderr": 0.03344887382997866 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2424581005586592, "acc_stderr": 0.01433352205921789, "acc_norm": 0.2424581005586592, "acc_norm_stderr": 0.01433352205921789 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.35, "acc_stderr": 0.04793724854411019, "acc_norm": 0.35, "acc_norm_stderr": 0.04793724854411019 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.40441176470588236, "acc_stderr": 0.029812630701569746, "acc_norm": 0.40441176470588236, "acc_norm_stderr": 0.029812630701569746 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.46530612244897956, "acc_stderr": 0.03193207024425314, "acc_norm": 0.46530612244897956, "acc_norm_stderr": 0.03193207024425314 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.27848101265822783, "acc_stderr": 0.029178682304842538, "acc_norm": 0.27848101265822783, "acc_norm_stderr": 0.029178682304842538 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.2790091264667536, "acc_stderr": 0.011455208832803538, "acc_norm": 0.2790091264667536, "acc_norm_stderr": 0.011455208832803538 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.29901960784313725, "acc_stderr": 0.03213325717373616, "acc_norm": 0.29901960784313725, "acc_norm_stderr": 0.03213325717373616 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.3515151515151515, "acc_stderr": 0.0372820699868265, "acc_norm": 0.3515151515151515, "acc_norm_stderr": 0.0372820699868265 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.29253365973072215, "mc1_stderr": 0.015925597445286165, "mc2": 0.46212361951327446, "mc2_stderr": 0.015249362527618285 }, "harness|ko_commongen_v2|2": { "acc": 0.3955135773317591, "acc_stderr": 0.016810815902206046, "acc_norm": 0.4757969303423849, "acc_norm_stderr": 0.017170202466520748 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Jenti-Kaeri/ko-llama2-13b-platypus", "model_sha": "f20decdd9d1525560ce299352c7ee6421c5ec81d", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }