{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.3873720136518771, "acc_stderr": 0.01423587248790987, "acc_norm": 0.4274744027303754, "acc_norm_stderr": 0.014456862944650647 }, "harness|ko_hellaswag|10": { "acc": 0.392850029874527, "acc_stderr": 0.0048738583238407945, "acc_norm": 0.5120493925512846, "acc_norm_stderr": 0.004988332289642083 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.52046783625731, "acc_stderr": 0.038316105328219295, "acc_norm": 0.52046783625731, "acc_norm_stderr": 0.038316105328219295 }, "harness|ko_mmlu_management|5": { "acc": 0.6116504854368932, "acc_stderr": 0.04825729337356389, "acc_norm": 0.6116504854368932, "acc_norm_stderr": 0.04825729337356389 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.5057471264367817, "acc_stderr": 0.017878782326129224, "acc_norm": 0.5057471264367817, "acc_norm_stderr": 0.017878782326129224 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.3851851851851852, "acc_stderr": 0.042039210401562783, "acc_norm": 0.3851851851851852, "acc_norm_stderr": 0.042039210401562783 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.43829787234042555, "acc_stderr": 0.03243618636108102, "acc_norm": 0.43829787234042555, "acc_norm_stderr": 0.03243618636108102 }, "harness|ko_mmlu_virology|5": { "acc": 0.41566265060240964, "acc_stderr": 0.03836722176598052, "acc_norm": 0.41566265060240964, "acc_norm_stderr": 0.03836722176598052 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.47266881028938906, "acc_stderr": 0.02835563356832818, "acc_norm": 0.47266881028938906, "acc_norm_stderr": 0.02835563356832818 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.484304932735426, "acc_stderr": 0.0335412657542081, "acc_norm": 0.484304932735426, "acc_norm_stderr": 0.0335412657542081 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.4580152671755725, "acc_stderr": 0.04369802690578756, "acc_norm": 0.4580152671755725, "acc_norm_stderr": 0.04369802690578756 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.38, "acc_stderr": 0.048783173121456316, "acc_norm": 0.38, "acc_norm_stderr": 0.048783173121456316 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.601010101010101, "acc_stderr": 0.03488901616852731, "acc_norm": 0.601010101010101, "acc_norm_stderr": 0.03488901616852731 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.503448275862069, "acc_stderr": 0.04166567577101579, "acc_norm": 0.503448275862069, "acc_norm_stderr": 0.04166567577101579 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.28431372549019607, "acc_stderr": 0.04488482852329017, "acc_norm": 0.28431372549019607, "acc_norm_stderr": 0.04488482852329017 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.47478991596638653, "acc_stderr": 0.0324371805513741, "acc_norm": 0.47478991596638653, "acc_norm_stderr": 0.0324371805513741 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.46923076923076923, "acc_stderr": 0.025302958890850154, "acc_norm": 0.46923076923076923, "acc_norm_stderr": 0.025302958890850154 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.52, "acc_stderr": 0.05021167315686779, "acc_norm": 0.52, "acc_norm_stderr": 0.05021167315686779 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5, "acc_stderr": 0.04833682445228318, "acc_norm": 0.5, "acc_norm_stderr": 0.04833682445228318 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.42857142857142855, "acc_stderr": 0.03481904844438803, "acc_norm": 0.42857142857142855, "acc_norm_stderr": 0.03481904844438803 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.4806451612903226, "acc_stderr": 0.028422687404312107, "acc_norm": 0.4806451612903226, "acc_norm_stderr": 0.028422687404312107 }, "harness|ko_mmlu_marketing|5": { "acc": 0.7478632478632479, "acc_stderr": 0.028447965476231022, "acc_norm": 0.7478632478632479, "acc_norm_stderr": 0.028447965476231022 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.4830188679245283, "acc_stderr": 0.030755120364119905, "acc_norm": 0.4830188679245283, "acc_norm_stderr": 0.030755120364119905 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5272727272727272, "acc_stderr": 0.04782001791380061, "acc_norm": 0.5272727272727272, "acc_norm_stderr": 0.04782001791380061 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.2851851851851852, "acc_stderr": 0.027528599210340492, "acc_norm": 0.2851851851851852, "acc_norm_stderr": 0.027528599210340492 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.3509933774834437, "acc_stderr": 0.03896981964257375, "acc_norm": 0.3509933774834437, "acc_norm_stderr": 0.03896981964257375 }, "harness|ko_mmlu_sociology|5": { "acc": 0.6218905472636815, "acc_stderr": 0.034288678487786564, "acc_norm": 0.6218905472636815, "acc_norm_stderr": 0.034288678487786564 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.4393063583815029, "acc_stderr": 0.03784271932887467, "acc_norm": 0.4393063583815029, "acc_norm_stderr": 0.03784271932887467 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.38095238095238093, "acc_stderr": 0.025010749116137595, "acc_norm": 0.38095238095238093, "acc_norm_stderr": 0.025010749116137595 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3888888888888889, "acc_stderr": 0.04076663253918567, "acc_norm": 0.3888888888888889, "acc_norm_stderr": 0.04076663253918567 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.39, "acc_stderr": 0.04902071300001975, "acc_norm": 0.39, "acc_norm_stderr": 0.04902071300001975 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.63, "acc_stderr": 0.048523658709391, "acc_norm": 0.63, "acc_norm_stderr": 0.048523658709391 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5202312138728323, "acc_stderr": 0.026897049996382868, "acc_norm": 0.5202312138728323, "acc_norm_stderr": 0.026897049996382868 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.44171779141104295, "acc_stderr": 0.03901591825836184, "acc_norm": 0.44171779141104295, "acc_norm_stderr": 0.03901591825836184 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.4382716049382716, "acc_stderr": 0.027607914087400473, "acc_norm": 0.4382716049382716, "acc_norm_stderr": 0.027607914087400473 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.49740932642487046, "acc_stderr": 0.03608390745384487, "acc_norm": 0.49740932642487046, "acc_norm_stderr": 0.03608390745384487 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2631578947368421, "acc_stderr": 0.04142439719489358, "acc_norm": 0.2631578947368421, "acc_norm_stderr": 0.04142439719489358 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.48440366972477067, "acc_stderr": 0.02142689153920805, "acc_norm": 0.48440366972477067, "acc_norm_stderr": 0.02142689153920805 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.3412698412698413, "acc_stderr": 0.04240799327574924, "acc_norm": 0.3412698412698413, "acc_norm_stderr": 0.04240799327574924 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5294117647058824, "acc_stderr": 0.028580341065138296, "acc_norm": 0.5294117647058824, "acc_norm_stderr": 0.028580341065138296 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.51, "acc_stderr": 0.05024183937956912, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956912 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6776859504132231, "acc_stderr": 0.042664163633521664, "acc_norm": 0.6776859504132231, "acc_norm_stderr": 0.042664163633521664 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.506578947368421, "acc_stderr": 0.040685900502249704, "acc_norm": 0.506578947368421, "acc_norm_stderr": 0.040685900502249704 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.4068627450980392, "acc_stderr": 0.019873802005061177, "acc_norm": 0.4068627450980392, "acc_norm_stderr": 0.019873802005061177 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.375886524822695, "acc_stderr": 0.028893955412115882, "acc_norm": 0.375886524822695, "acc_norm_stderr": 0.028893955412115882 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.4107142857142857, "acc_stderr": 0.04669510663875192, "acc_norm": 0.4107142857142857, "acc_norm_stderr": 0.04669510663875192 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4166666666666667, "acc_stderr": 0.033622774366080424, "acc_norm": 0.4166666666666667, "acc_norm_stderr": 0.033622774366080424 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.27932960893854747, "acc_stderr": 0.015005762446786173, "acc_norm": 0.27932960893854747, "acc_norm_stderr": 0.015005762446786173 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.45, "acc_stderr": 0.05, "acc_norm": 0.45, "acc_norm_stderr": 0.05 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.57, "acc_stderr": 0.04975698519562428, "acc_norm": 0.57, "acc_norm_stderr": 0.04975698519562428 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4338235294117647, "acc_stderr": 0.030105636570016636, "acc_norm": 0.4338235294117647, "acc_norm_stderr": 0.030105636570016636 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.44081632653061226, "acc_stderr": 0.03178419114175363, "acc_norm": 0.44081632653061226, "acc_norm_stderr": 0.03178419114175363 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.6160337552742616, "acc_stderr": 0.031658678064106674, "acc_norm": 0.6160337552742616, "acc_norm_stderr": 0.031658678064106674 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3428943937418514, "acc_stderr": 0.012123463271585895, "acc_norm": 0.3428943937418514, "acc_norm_stderr": 0.012123463271585895 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.45588235294117646, "acc_stderr": 0.03495624522015474, "acc_norm": 0.45588235294117646, "acc_norm_stderr": 0.03495624522015474 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.4121212121212121, "acc_stderr": 0.03843566993588718, "acc_norm": 0.4121212121212121, "acc_norm_stderr": 0.03843566993588718 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2937576499388005, "mc1_stderr": 0.015945068581236614, "mc2": 0.4623692353701492, "mc2_stderr": 0.015502597273810991 }, "harness|ko_commongen_v2|2": { "acc": 0.3860684769775679, "acc_stderr": 0.016738130760321743, "acc_norm": 0.4722550177095632, "acc_norm_stderr": 0.017163867979456012 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Ja3ck/Mistral-instruct-Y24-DPO", "model_sha": "5cadddfbeeac1dd2be25ea036d8623968e987f3c", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }