{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.4786689419795222, "acc_stderr": 0.014598087973127108, "acc_norm": 0.5273037542662116, "acc_norm_stderr": 0.014589589101986001 }, "harness|ko_hellaswag|10": { "acc": 0.47739494124676357, "acc_stderr": 0.004984679359375627, "acc_norm": 0.6482772356104362, "acc_norm_stderr": 0.0047653207849021086 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.6666666666666666, "acc_stderr": 0.036155076303109344, "acc_norm": 0.6666666666666666, "acc_norm_stderr": 0.036155076303109344 }, "harness|ko_mmlu_management|5": { "acc": 0.6407766990291263, "acc_stderr": 0.04750458399041696, "acc_norm": 0.6407766990291263, "acc_norm_stderr": 0.04750458399041696 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.6206896551724138, "acc_stderr": 0.01735126811754445, "acc_norm": 0.6206896551724138, "acc_norm_stderr": 0.01735126811754445 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4222222222222222, "acc_stderr": 0.04266763404099582, "acc_norm": 0.4222222222222222, "acc_norm_stderr": 0.04266763404099582 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.27, "acc_stderr": 0.044619604333847394, "acc_norm": 0.27, "acc_norm_stderr": 0.044619604333847394 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.4127659574468085, "acc_stderr": 0.03218471141400351, "acc_norm": 0.4127659574468085, "acc_norm_stderr": 0.03218471141400351 }, "harness|ko_mmlu_virology|5": { "acc": 0.4397590361445783, "acc_stderr": 0.03864139923699121, "acc_norm": 0.4397590361445783, "acc_norm_stderr": 0.03864139923699121 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.572347266881029, "acc_stderr": 0.02809924077580957, "acc_norm": 0.572347266881029, "acc_norm_stderr": 0.02809924077580957 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.547085201793722, "acc_stderr": 0.03340867501923324, "acc_norm": 0.547085201793722, "acc_norm_stderr": 0.03340867501923324 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.549618320610687, "acc_stderr": 0.04363643698524779, "acc_norm": 0.549618320610687, "acc_norm_stderr": 0.04363643698524779 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.45, "acc_stderr": 0.04999999999999999, "acc_norm": 0.45, "acc_norm_stderr": 0.04999999999999999 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.7323232323232324, "acc_stderr": 0.03154449888270286, "acc_norm": 0.7323232323232324, "acc_norm_stderr": 0.03154449888270286 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.42758620689655175, "acc_stderr": 0.041227371113703316, "acc_norm": 0.42758620689655175, "acc_norm_stderr": 0.041227371113703316 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.3235294117647059, "acc_stderr": 0.04655010411319616, "acc_norm": 0.3235294117647059, "acc_norm_stderr": 0.04655010411319616 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.6680672268907563, "acc_stderr": 0.03058869701378364, "acc_norm": 0.6680672268907563, "acc_norm_stderr": 0.03058869701378364 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.5871794871794872, "acc_stderr": 0.024962683564331793, "acc_norm": 0.5871794871794872, "acc_norm_stderr": 0.024962683564331793 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.59, "acc_stderr": 0.049431107042371025, "acc_norm": 0.59, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.38, "acc_stderr": 0.04878317312145633, "acc_norm": 0.38, "acc_norm_stderr": 0.04878317312145633 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.6296296296296297, "acc_stderr": 0.04668408033024931, "acc_norm": 0.6296296296296297, "acc_norm_stderr": 0.04668408033024931 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3793103448275862, "acc_stderr": 0.034139638059062345, "acc_norm": 0.3793103448275862, "acc_norm_stderr": 0.034139638059062345 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.5645161290322581, "acc_stderr": 0.028206225591502734, "acc_norm": 0.5645161290322581, "acc_norm_stderr": 0.028206225591502734 }, "harness|ko_mmlu_marketing|5": { "acc": 0.782051282051282, "acc_stderr": 0.027046857630716663, "acc_norm": 0.782051282051282, "acc_norm_stderr": 0.027046857630716663 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.5320754716981132, "acc_stderr": 0.030709486992556545, "acc_norm": 0.5320754716981132, "acc_norm_stderr": 0.030709486992556545 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.6, "acc_stderr": 0.0469237132203465, "acc_norm": 0.6, "acc_norm_stderr": 0.0469237132203465 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.3888888888888889, "acc_stderr": 0.029723278961476664, "acc_norm": 0.3888888888888889, "acc_norm_stderr": 0.029723278961476664 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.36423841059602646, "acc_stderr": 0.03929111781242742, "acc_norm": 0.36423841059602646, "acc_norm_stderr": 0.03929111781242742 }, "harness|ko_mmlu_sociology|5": { "acc": 0.6069651741293532, "acc_stderr": 0.0345368246603156, "acc_norm": 0.6069651741293532, "acc_norm_stderr": 0.0345368246603156 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.4797687861271676, "acc_stderr": 0.03809342081273958, "acc_norm": 0.4797687861271676, "acc_norm_stderr": 0.03809342081273958 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.4523809523809524, "acc_stderr": 0.02563425811555496, "acc_norm": 0.4523809523809524, "acc_norm_stderr": 0.02563425811555496 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.5555555555555556, "acc_stderr": 0.04155319955593146, "acc_norm": 0.5555555555555556, "acc_norm_stderr": 0.04155319955593146 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.33, "acc_stderr": 0.047258156262526045, "acc_norm": 0.33, "acc_norm_stderr": 0.047258156262526045 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.71, "acc_stderr": 0.04560480215720683, "acc_norm": 0.71, "acc_norm_stderr": 0.04560480215720683 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5809248554913294, "acc_stderr": 0.02656417811142263, "acc_norm": 0.5809248554913294, "acc_norm_stderr": 0.02656417811142263 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.4723926380368098, "acc_stderr": 0.0392237829061099, "acc_norm": 0.4723926380368098, "acc_norm_stderr": 0.0392237829061099 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.6172839506172839, "acc_stderr": 0.027044538138402602, "acc_norm": 0.6172839506172839, "acc_norm_stderr": 0.027044538138402602 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.7409326424870466, "acc_stderr": 0.03161877917935411, "acc_norm": 0.7409326424870466, "acc_norm_stderr": 0.03161877917935411 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.41228070175438597, "acc_stderr": 0.046306532033665956, "acc_norm": 0.41228070175438597, "acc_norm_stderr": 0.046306532033665956 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.671559633027523, "acc_stderr": 0.02013590279729841, "acc_norm": 0.671559633027523, "acc_norm_stderr": 0.02013590279729841 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.48412698412698413, "acc_stderr": 0.04469881854072606, "acc_norm": 0.48412698412698413, "acc_norm_stderr": 0.04469881854072606 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5686274509803921, "acc_stderr": 0.02835895631342355, "acc_norm": 0.5686274509803921, "acc_norm_stderr": 0.02835895631342355 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.59, "acc_stderr": 0.049431107042371025, "acc_norm": 0.59, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6611570247933884, "acc_stderr": 0.0432076780753667, "acc_norm": 0.6611570247933884, "acc_norm_stderr": 0.0432076780753667 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.5986842105263158, "acc_stderr": 0.039889037033362836, "acc_norm": 0.5986842105263158, "acc_norm_stderr": 0.039889037033362836 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.49836601307189543, "acc_stderr": 0.020227726838150117, "acc_norm": 0.49836601307189543, "acc_norm_stderr": 0.020227726838150117 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3829787234042553, "acc_stderr": 0.02899908090480619, "acc_norm": 0.3829787234042553, "acc_norm_stderr": 0.02899908090480619 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.375, "acc_stderr": 0.04595091388086298, "acc_norm": 0.375, "acc_norm_stderr": 0.04595091388086298 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.5277777777777778, "acc_stderr": 0.0340470532865388, "acc_norm": 0.5277777777777778, "acc_norm_stderr": 0.0340470532865388 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.25921787709497207, "acc_stderr": 0.014655780837497717, "acc_norm": 0.25921787709497207, "acc_norm_stderr": 0.014655780837497717 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.51, "acc_stderr": 0.05024183937956911, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956911 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.67, "acc_stderr": 0.04725815626252607, "acc_norm": 0.67, "acc_norm_stderr": 0.04725815626252607 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.5147058823529411, "acc_stderr": 0.03035969707904611, "acc_norm": 0.5147058823529411, "acc_norm_stderr": 0.03035969707904611 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.5102040816326531, "acc_stderr": 0.03200255347893782, "acc_norm": 0.5102040816326531, "acc_norm_stderr": 0.03200255347893782 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.759493670886076, "acc_stderr": 0.02782078198114968, "acc_norm": 0.759493670886076, "acc_norm_stderr": 0.02782078198114968 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.42046936114732725, "acc_stderr": 0.012607654553832707, "acc_norm": 0.42046936114732725, "acc_norm_stderr": 0.012607654553832707 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.6862745098039216, "acc_stderr": 0.03256685484460388, "acc_norm": 0.6862745098039216, "acc_norm_stderr": 0.03256685484460388 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.696969696969697, "acc_stderr": 0.035886248000917095, "acc_norm": 0.696969696969697, "acc_norm_stderr": 0.035886248000917095 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.41003671970624234, "mc1_stderr": 0.017217844717449318, "mc2": 0.583646136539335, "mc2_stderr": 0.016204158336690234 }, "harness|ko_commongen_v2|2": { "acc": 0.5242030696576151, "acc_stderr": 0.017170202466520748, "acc_norm": 0.5371900826446281, "acc_norm_stderr": 0.017142736117643304 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Edentns/DataVortexS-10.7B-dpo-v1.2", "model_sha": "4620f8491ea4cc4b2d47f41ee7a6f32a6b1bfa01", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }