{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.30802047781569963, "acc_stderr": 0.01349142951729204, "acc_norm": 0.3515358361774744, "acc_norm_stderr": 0.013952413699600938 }, "harness|ko_hellaswag|10": { "acc": 0.3610834495120494, "acc_stderr": 0.004793330525656211, "acc_norm": 0.45120493925512845, "acc_norm_stderr": 0.004965963647210315 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.2573099415204678, "acc_stderr": 0.03352799844161865, "acc_norm": 0.2573099415204678, "acc_norm_stderr": 0.03352799844161865 }, "harness|ko_mmlu_management|5": { "acc": 0.4077669902912621, "acc_stderr": 0.048657775704107696, "acc_norm": 0.4077669902912621, "acc_norm_stderr": 0.048657775704107696 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.2503192848020434, "acc_stderr": 0.01549108895149458, "acc_norm": 0.2503192848020434, "acc_norm_stderr": 0.01549108895149458 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.23703703703703705, "acc_stderr": 0.03673731683969506, "acc_norm": 0.23703703703703705, "acc_norm_stderr": 0.03673731683969506 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.21, "acc_stderr": 0.040936018074033256, "acc_norm": 0.21, "acc_norm_stderr": 0.040936018074033256 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.20425531914893616, "acc_stderr": 0.026355158413349424, "acc_norm": 0.20425531914893616, "acc_norm_stderr": 0.026355158413349424 }, "harness|ko_mmlu_virology|5": { "acc": 0.19879518072289157, "acc_stderr": 0.031069390260789396, "acc_norm": 0.19879518072289157, "acc_norm_stderr": 0.031069390260789396 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.24115755627009647, "acc_stderr": 0.024296594034763426, "acc_norm": 0.24115755627009647, "acc_norm_stderr": 0.024296594034763426 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.1210762331838565, "acc_stderr": 0.021894174113185737, "acc_norm": 0.1210762331838565, "acc_norm_stderr": 0.021894174113185737 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.3053435114503817, "acc_stderr": 0.04039314978724561, "acc_norm": 0.3053435114503817, "acc_norm_stderr": 0.04039314978724561 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.24, "acc_stderr": 0.04292346959909284, "acc_norm": 0.24, "acc_norm_stderr": 0.04292346959909284 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.3686868686868687, "acc_stderr": 0.034373055019806184, "acc_norm": 0.3686868686868687, "acc_norm_stderr": 0.034373055019806184 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.2482758620689655, "acc_stderr": 0.036001056927277716, "acc_norm": 0.2482758620689655, "acc_norm_stderr": 0.036001056927277716 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.4019607843137255, "acc_stderr": 0.04878608714466996, "acc_norm": 0.4019607843137255, "acc_norm_stderr": 0.04878608714466996 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.3487394957983193, "acc_stderr": 0.030956636328566548, "acc_norm": 0.3487394957983193, "acc_norm_stderr": 0.030956636328566548 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.36666666666666664, "acc_stderr": 0.024433016466052452, "acc_norm": 0.36666666666666664, "acc_norm_stderr": 0.024433016466052452 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.22, "acc_stderr": 0.0416333199893227, "acc_norm": 0.22, "acc_norm_stderr": 0.0416333199893227 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.19, "acc_stderr": 0.03942772444036623, "acc_norm": 0.19, "acc_norm_stderr": 0.03942772444036623 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.23148148148148148, "acc_stderr": 0.04077494709252626, "acc_norm": 0.23148148148148148, "acc_norm_stderr": 0.04077494709252626 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.28078817733990147, "acc_stderr": 0.0316185633535861, "acc_norm": 0.28078817733990147, "acc_norm_stderr": 0.0316185633535861 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.3161290322580645, "acc_stderr": 0.026450874489042764, "acc_norm": 0.3161290322580645, "acc_norm_stderr": 0.026450874489042764 }, "harness|ko_mmlu_marketing|5": { "acc": 0.2094017094017094, "acc_stderr": 0.026655699653922737, "acc_norm": 0.2094017094017094, "acc_norm_stderr": 0.026655699653922737 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.3018867924528302, "acc_stderr": 0.02825420034443866, "acc_norm": 0.3018867924528302, "acc_norm_stderr": 0.02825420034443866 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.24545454545454545, "acc_stderr": 0.041220665028782834, "acc_norm": 0.24545454545454545, "acc_norm_stderr": 0.041220665028782834 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.26296296296296295, "acc_stderr": 0.02684205787383371, "acc_norm": 0.26296296296296295, "acc_norm_stderr": 0.02684205787383371 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.33112582781456956, "acc_stderr": 0.038425817186598696, "acc_norm": 0.33112582781456956, "acc_norm_stderr": 0.038425817186598696 }, "harness|ko_mmlu_sociology|5": { "acc": 0.2736318407960199, "acc_stderr": 0.03152439186555401, "acc_norm": 0.2736318407960199, "acc_norm_stderr": 0.03152439186555401 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.34104046242774566, "acc_stderr": 0.036146654241808254, "acc_norm": 0.34104046242774566, "acc_norm_stderr": 0.036146654241808254 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2671957671957672, "acc_stderr": 0.02278967314577656, "acc_norm": 0.2671957671957672, "acc_norm_stderr": 0.02278967314577656 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2638888888888889, "acc_stderr": 0.03685651095897532, "acc_norm": 0.2638888888888889, "acc_norm_stderr": 0.03685651095897532 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.26, "acc_stderr": 0.044084400227680794, "acc_norm": 0.26, "acc_norm_stderr": 0.044084400227680794 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.22832369942196531, "acc_stderr": 0.022598703804321624, "acc_norm": 0.22832369942196531, "acc_norm_stderr": 0.022598703804321624 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.2331288343558282, "acc_stderr": 0.0332201579577674, "acc_norm": 0.2331288343558282, "acc_norm_stderr": 0.0332201579577674 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.2345679012345679, "acc_stderr": 0.02357688174400572, "acc_norm": 0.2345679012345679, "acc_norm_stderr": 0.02357688174400572 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.37305699481865284, "acc_stderr": 0.03490205592048574, "acc_norm": 0.37305699481865284, "acc_norm_stderr": 0.03490205592048574 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.23684210526315788, "acc_stderr": 0.039994238792813365, "acc_norm": 0.23684210526315788, "acc_norm_stderr": 0.039994238792813365 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.3559633027522936, "acc_stderr": 0.020528559278244214, "acc_norm": 0.3559633027522936, "acc_norm_stderr": 0.020528559278244214 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.36507936507936506, "acc_stderr": 0.04306241259127153, "acc_norm": 0.36507936507936506, "acc_norm_stderr": 0.04306241259127153 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.29411764705882354, "acc_stderr": 0.026090162504279053, "acc_norm": 0.29411764705882354, "acc_norm_stderr": 0.026090162504279053 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.21, "acc_stderr": 0.040936018074033256, "acc_norm": 0.21, "acc_norm_stderr": 0.040936018074033256 }, "harness|ko_mmlu_international_law|5": { "acc": 0.14049586776859505, "acc_stderr": 0.031722334260021585, "acc_norm": 0.14049586776859505, "acc_norm_stderr": 0.031722334260021585 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.34210526315789475, "acc_stderr": 0.03860731599316091, "acc_norm": 0.34210526315789475, "acc_norm_stderr": 0.03860731599316091 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.22058823529411764, "acc_stderr": 0.016774672365468517, "acc_norm": 0.22058823529411764, "acc_norm_stderr": 0.016774672365468517 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.24468085106382978, "acc_stderr": 0.02564555362226673, "acc_norm": 0.24468085106382978, "acc_norm_stderr": 0.02564555362226673 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.16964285714285715, "acc_stderr": 0.0356236785009539, "acc_norm": 0.16964285714285715, "acc_norm_stderr": 0.0356236785009539 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4722222222222222, "acc_stderr": 0.0340470532865388, "acc_norm": 0.4722222222222222, "acc_norm_stderr": 0.0340470532865388 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.27262569832402234, "acc_stderr": 0.014893391735249608, "acc_norm": 0.27262569832402234, "acc_norm_stderr": 0.014893391735249608 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.23, "acc_stderr": 0.04229525846816506, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816506 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4485294117647059, "acc_stderr": 0.030211479609121593, "acc_norm": 0.4485294117647059, "acc_norm_stderr": 0.030211479609121593 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.4, "acc_stderr": 0.03136250240935892, "acc_norm": 0.4, "acc_norm_stderr": 0.03136250240935892 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.20253164556962025, "acc_stderr": 0.026160568246601464, "acc_norm": 0.20253164556962025, "acc_norm_stderr": 0.026160568246601464 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.24511082138200782, "acc_stderr": 0.010986307870045509, "acc_norm": 0.24511082138200782, "acc_norm_stderr": 0.010986307870045509 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.2549019607843137, "acc_stderr": 0.030587591351604246, "acc_norm": 0.2549019607843137, "acc_norm_stderr": 0.030587591351604246 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.2545454545454545, "acc_stderr": 0.03401506715249039, "acc_norm": 0.2545454545454545, "acc_norm_stderr": 0.03401506715249039 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2962056303549572, "mc1_stderr": 0.015983595101811392, "mc2": 0.4616568963266555, "mc2_stderr": 0.01577378737316958 }, "harness|ko_commongen_v2|2": { "acc": 0.3659976387249115, "acc_stderr": 0.01656148966489569, "acc_norm": 0.44037780401416765, "acc_norm_stderr": 0.017067699774312987 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "mncai/Mistral-7B-v0.1-combine-1k", "model_sha": "0f7abf5c07a7f3add4c89c9e3525f29ab89be562", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }