{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.36689419795221845, "acc_stderr": 0.014084133118104296, "acc_norm": 0.41638225255972694, "acc_norm_stderr": 0.014405618279436172 }, "harness|ko_hellaswag|10": { "acc": 0.4124676359290978, "acc_stderr": 0.004912723848944785, "acc_norm": 0.5473013343955387, "acc_norm_stderr": 0.004967402792744853 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.5146198830409356, "acc_stderr": 0.038331852752130254, "acc_norm": 0.5146198830409356, "acc_norm_stderr": 0.038331852752130254 }, "harness|ko_mmlu_management|5": { "acc": 0.5533980582524272, "acc_stderr": 0.04922424153458933, "acc_norm": 0.5533980582524272, "acc_norm_stderr": 0.04922424153458933 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.5376756066411239, "acc_stderr": 0.017829131764287187, "acc_norm": 0.5376756066411239, "acc_norm_stderr": 0.017829131764287187 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.45185185185185184, "acc_stderr": 0.04299268905480864, "acc_norm": 0.45185185185185184, "acc_norm_stderr": 0.04299268905480864 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.23, "acc_stderr": 0.04229525846816503, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816503 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.43829787234042555, "acc_stderr": 0.03243618636108101, "acc_norm": 0.43829787234042555, "acc_norm_stderr": 0.03243618636108101 }, "harness|ko_mmlu_virology|5": { "acc": 0.3373493975903614, "acc_stderr": 0.036807836907275814, "acc_norm": 0.3373493975903614, "acc_norm_stderr": 0.036807836907275814 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5273311897106109, "acc_stderr": 0.028355633568328188, "acc_norm": 0.5273311897106109, "acc_norm_stderr": 0.028355633568328188 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.49327354260089684, "acc_stderr": 0.033554765962343545, "acc_norm": 0.49327354260089684, "acc_norm_stderr": 0.033554765962343545 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.5190839694656488, "acc_stderr": 0.043820947055509867, "acc_norm": 0.5190839694656488, "acc_norm_stderr": 0.043820947055509867 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.38, "acc_stderr": 0.048783173121456316, "acc_norm": 0.38, "acc_norm_stderr": 0.048783173121456316 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.6161616161616161, "acc_stderr": 0.034648816750163375, "acc_norm": 0.6161616161616161, "acc_norm_stderr": 0.034648816750163375 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.42758620689655175, "acc_stderr": 0.04122737111370331, "acc_norm": 0.42758620689655175, "acc_norm_stderr": 0.04122737111370331 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.3137254901960784, "acc_stderr": 0.04617034827006716, "acc_norm": 0.3137254901960784, "acc_norm_stderr": 0.04617034827006716 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.4327731092436975, "acc_stderr": 0.03218358107742613, "acc_norm": 0.4327731092436975, "acc_norm_stderr": 0.03218358107742613 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.43846153846153846, "acc_stderr": 0.025158266016868547, "acc_norm": 0.43846153846153846, "acc_norm_stderr": 0.025158266016868547 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.54, "acc_stderr": 0.05009082659620332, "acc_norm": 0.54, "acc_norm_stderr": 0.05009082659620332 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5, "acc_stderr": 0.04833682445228318, "acc_norm": 0.5, "acc_norm_stderr": 0.04833682445228318 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3891625615763547, "acc_stderr": 0.034304624161038716, "acc_norm": 0.3891625615763547, "acc_norm_stderr": 0.034304624161038716 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.4806451612903226, "acc_stderr": 0.0284226874043121, "acc_norm": 0.4806451612903226, "acc_norm_stderr": 0.0284226874043121 }, "harness|ko_mmlu_marketing|5": { "acc": 0.7136752136752137, "acc_stderr": 0.029614323690456648, "acc_norm": 0.7136752136752137, "acc_norm_stderr": 0.029614323690456648 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.5094339622641509, "acc_stderr": 0.030767394707808093, "acc_norm": 0.5094339622641509, "acc_norm_stderr": 0.030767394707808093 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.43636363636363634, "acc_stderr": 0.04750185058907297, "acc_norm": 0.43636363636363634, "acc_norm_stderr": 0.04750185058907297 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.3037037037037037, "acc_stderr": 0.028037929969114982, "acc_norm": 0.3037037037037037, "acc_norm_stderr": 0.028037929969114982 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.3443708609271523, "acc_stderr": 0.038796870240733264, "acc_norm": 0.3443708609271523, "acc_norm_stderr": 0.038796870240733264 }, "harness|ko_mmlu_sociology|5": { "acc": 0.5970149253731343, "acc_stderr": 0.03468343295111126, "acc_norm": 0.5970149253731343, "acc_norm_stderr": 0.03468343295111126 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.44508670520231214, "acc_stderr": 0.03789401760283649, "acc_norm": 0.44508670520231214, "acc_norm_stderr": 0.03789401760283649 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.3492063492063492, "acc_stderr": 0.024552292209342658, "acc_norm": 0.3492063492063492, "acc_norm_stderr": 0.024552292209342658 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.4375, "acc_stderr": 0.04148415739394154, "acc_norm": 0.4375, "acc_norm_stderr": 0.04148415739394154 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.33, "acc_stderr": 0.04725815626252605, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252605 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.59, "acc_stderr": 0.04943110704237101, "acc_norm": 0.59, "acc_norm_stderr": 0.04943110704237101 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.4884393063583815, "acc_stderr": 0.02691189868637793, "acc_norm": 0.4884393063583815, "acc_norm_stderr": 0.02691189868637793 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.4294478527607362, "acc_stderr": 0.03889066619112722, "acc_norm": 0.4294478527607362, "acc_norm_stderr": 0.03889066619112722 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.5061728395061729, "acc_stderr": 0.027818623962583295, "acc_norm": 0.5061728395061729, "acc_norm_stderr": 0.027818623962583295 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.33, "acc_stderr": 0.04725815626252606, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252606 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.6010362694300518, "acc_stderr": 0.035339990940656964, "acc_norm": 0.6010362694300518, "acc_norm_stderr": 0.035339990940656964 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2982456140350877, "acc_stderr": 0.043036840335373173, "acc_norm": 0.2982456140350877, "acc_norm_stderr": 0.043036840335373173 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.6018348623853211, "acc_stderr": 0.020987989422654257, "acc_norm": 0.6018348623853211, "acc_norm_stderr": 0.020987989422654257 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.30952380952380953, "acc_stderr": 0.04134913018303316, "acc_norm": 0.30952380952380953, "acc_norm_stderr": 0.04134913018303316 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.47058823529411764, "acc_stderr": 0.02858034106513829, "acc_norm": 0.47058823529411764, "acc_norm_stderr": 0.02858034106513829 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.51, "acc_stderr": 0.05024183937956911, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956911 }, "harness|ko_mmlu_international_law|5": { "acc": 0.5867768595041323, "acc_stderr": 0.04495087843548408, "acc_norm": 0.5867768595041323, "acc_norm_stderr": 0.04495087843548408 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.4473684210526316, "acc_stderr": 0.040463368839782514, "acc_norm": 0.4473684210526316, "acc_norm_stderr": 0.040463368839782514 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.39705882352941174, "acc_stderr": 0.019794488900024113, "acc_norm": 0.39705882352941174, "acc_norm_stderr": 0.019794488900024113 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3191489361702128, "acc_stderr": 0.027807990141320196, "acc_norm": 0.3191489361702128, "acc_norm_stderr": 0.027807990141320196 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.23214285714285715, "acc_stderr": 0.040073418097558065, "acc_norm": 0.23214285714285715, "acc_norm_stderr": 0.040073418097558065 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.3333333333333333, "acc_stderr": 0.032149521478027486, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.032149521478027486 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2547486033519553, "acc_stderr": 0.014572650383409167, "acc_norm": 0.2547486033519553, "acc_norm_stderr": 0.014572650383409167 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.5, "acc_stderr": 0.050251890762960605, "acc_norm": 0.5, "acc_norm_stderr": 0.050251890762960605 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.40808823529411764, "acc_stderr": 0.029855261393483924, "acc_norm": 0.40808823529411764, "acc_norm_stderr": 0.029855261393483924 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.42857142857142855, "acc_stderr": 0.031680911612338825, "acc_norm": 0.42857142857142855, "acc_norm_stderr": 0.031680911612338825 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.6329113924050633, "acc_stderr": 0.031376240725616185, "acc_norm": 0.6329113924050633, "acc_norm_stderr": 0.031376240725616185 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.31029986962190353, "acc_stderr": 0.011815439293469832, "acc_norm": 0.31029986962190353, "acc_norm_stderr": 0.011815439293469832 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.49019607843137253, "acc_stderr": 0.03508637358630572, "acc_norm": 0.49019607843137253, "acc_norm_stderr": 0.03508637358630572 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.5515151515151515, "acc_stderr": 0.038835659779569286, "acc_norm": 0.5515151515151515, "acc_norm_stderr": 0.038835659779569286 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.35495716034271724, "mc1_stderr": 0.0167508623813759, "mc2": 0.5208407477265208, "mc2_stderr": 0.015859200141974766 }, "harness|ko_commongen_v2|2": { "acc": 0.5017709563164109, "acc_stderr": 0.017190246276231853, "acc_norm": 0.5206611570247934, "acc_norm_stderr": 0.017175671279836446 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "GAI-LLM/Yi-Ko-6B-smash-dpo", "model_sha": "a28f3cc3bcb5b30bdfbac46949399f3119a3ffd0", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }