{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.4974402730375427, "acc_stderr": 0.014611199329843784, "acc_norm": 0.5554607508532423, "acc_norm_stderr": 0.01452122640562708 }, "harness|ko_hellaswag|10": { "acc": 0.45937064329814775, "acc_stderr": 0.004973280417705513, "acc_norm": 0.6339374626568413, "acc_norm_stderr": 0.004807423343224586 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.6374269005847953, "acc_stderr": 0.0368713061556206, "acc_norm": 0.6374269005847953, "acc_norm_stderr": 0.0368713061556206 }, "harness|ko_mmlu_management|5": { "acc": 0.6601941747572816, "acc_stderr": 0.04689765937278135, "acc_norm": 0.6601941747572816, "acc_norm_stderr": 0.04689765937278135 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.70242656449553, "acc_stderr": 0.01634911191290943, "acc_norm": 0.70242656449553, "acc_norm_stderr": 0.01634911191290943 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4740740740740741, "acc_stderr": 0.04313531696750573, "acc_norm": 0.4740740740740741, "acc_norm_stderr": 0.04313531696750573 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.34, "acc_stderr": 0.04760952285695235, "acc_norm": 0.34, "acc_norm_stderr": 0.04760952285695235 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.4765957446808511, "acc_stderr": 0.03265019475033583, "acc_norm": 0.4765957446808511, "acc_norm_stderr": 0.03265019475033583 }, "harness|ko_mmlu_virology|5": { "acc": 0.4457831325301205, "acc_stderr": 0.03869543323472101, "acc_norm": 0.4457831325301205, "acc_norm_stderr": 0.03869543323472101 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5691318327974276, "acc_stderr": 0.028125340983972714, "acc_norm": 0.5691318327974276, "acc_norm_stderr": 0.028125340983972714 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.5515695067264574, "acc_stderr": 0.03337883736255098, "acc_norm": 0.5515695067264574, "acc_norm_stderr": 0.03337883736255098 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.549618320610687, "acc_stderr": 0.04363643698524779, "acc_norm": 0.549618320610687, "acc_norm_stderr": 0.04363643698524779 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.48, "acc_stderr": 0.05021167315686779, "acc_norm": 0.48, "acc_norm_stderr": 0.05021167315686779 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.6767676767676768, "acc_stderr": 0.033322999210706444, "acc_norm": 0.6767676767676768, "acc_norm_stderr": 0.033322999210706444 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.496551724137931, "acc_stderr": 0.041665675771015785, "acc_norm": 0.496551724137931, "acc_norm_stderr": 0.041665675771015785 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.2647058823529412, "acc_stderr": 0.04389869956808778, "acc_norm": 0.2647058823529412, "acc_norm_stderr": 0.04389869956808778 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.5, "acc_stderr": 0.032478490123081544, "acc_norm": 0.5, "acc_norm_stderr": 0.032478490123081544 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.5538461538461539, "acc_stderr": 0.025203571773028323, "acc_norm": 0.5538461538461539, "acc_norm_stderr": 0.025203571773028323 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.61, "acc_stderr": 0.04902071300001974, "acc_norm": 0.61, "acc_norm_stderr": 0.04902071300001974 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.34, "acc_stderr": 0.04760952285695235, "acc_norm": 0.34, "acc_norm_stderr": 0.04760952285695235 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5833333333333334, "acc_stderr": 0.04766075165356461, "acc_norm": 0.5833333333333334, "acc_norm_stderr": 0.04766075165356461 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.39901477832512317, "acc_stderr": 0.03445487686264715, "acc_norm": 0.39901477832512317, "acc_norm_stderr": 0.03445487686264715 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.6129032258064516, "acc_stderr": 0.027709359675032495, "acc_norm": 0.6129032258064516, "acc_norm_stderr": 0.027709359675032495 }, "harness|ko_mmlu_marketing|5": { "acc": 0.8034188034188035, "acc_stderr": 0.02603538609895129, "acc_norm": 0.8034188034188035, "acc_norm_stderr": 0.02603538609895129 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.5320754716981132, "acc_stderr": 0.03070948699255654, "acc_norm": 0.5320754716981132, "acc_norm_stderr": 0.03070948699255654 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5545454545454546, "acc_stderr": 0.047605488214603246, "acc_norm": 0.5545454545454546, "acc_norm_stderr": 0.047605488214603246 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.3814814814814815, "acc_stderr": 0.029616718927497596, "acc_norm": 0.3814814814814815, "acc_norm_stderr": 0.029616718927497596 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.4105960264900662, "acc_stderr": 0.04016689594849929, "acc_norm": 0.4105960264900662, "acc_norm_stderr": 0.04016689594849929 }, "harness|ko_mmlu_sociology|5": { "acc": 0.681592039800995, "acc_stderr": 0.03294118479054096, "acc_norm": 0.681592039800995, "acc_norm_stderr": 0.03294118479054096 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.4508670520231214, "acc_stderr": 0.037940126746970296, "acc_norm": 0.4508670520231214, "acc_norm_stderr": 0.037940126746970296 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.36243386243386244, "acc_stderr": 0.02475747390275206, "acc_norm": 0.36243386243386244, "acc_norm_stderr": 0.02475747390275206 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.5069444444444444, "acc_stderr": 0.04180806750294938, "acc_norm": 0.5069444444444444, "acc_norm_stderr": 0.04180806750294938 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.38, "acc_stderr": 0.04878317312145632, "acc_norm": 0.38, "acc_norm_stderr": 0.04878317312145632 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.76, "acc_stderr": 0.04292346959909283, "acc_norm": 0.76, "acc_norm_stderr": 0.04292346959909283 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5549132947976878, "acc_stderr": 0.026756255129663765, "acc_norm": 0.5549132947976878, "acc_norm_stderr": 0.026756255129663765 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.5337423312883436, "acc_stderr": 0.039194155450484096, "acc_norm": 0.5337423312883436, "acc_norm_stderr": 0.039194155450484096 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.5370370370370371, "acc_stderr": 0.027744313443376536, "acc_norm": 0.5370370370370371, "acc_norm_stderr": 0.027744313443376536 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.6580310880829016, "acc_stderr": 0.03423465100104283, "acc_norm": 0.6580310880829016, "acc_norm_stderr": 0.03423465100104283 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.37719298245614036, "acc_stderr": 0.04559522141958215, "acc_norm": 0.37719298245614036, "acc_norm_stderr": 0.04559522141958215 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.636697247706422, "acc_stderr": 0.020620603919625807, "acc_norm": 0.636697247706422, "acc_norm_stderr": 0.020620603919625807 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.373015873015873, "acc_stderr": 0.04325506042017086, "acc_norm": 0.373015873015873, "acc_norm_stderr": 0.04325506042017086 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5294117647058824, "acc_stderr": 0.02858034106513829, "acc_norm": 0.5294117647058824, "acc_norm_stderr": 0.02858034106513829 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.58, "acc_stderr": 0.049604496374885836, "acc_norm": 0.58, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_international_law|5": { "acc": 0.7024793388429752, "acc_stderr": 0.04173349148083499, "acc_norm": 0.7024793388429752, "acc_norm_stderr": 0.04173349148083499 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.5789473684210527, "acc_stderr": 0.040179012759817494, "acc_norm": 0.5789473684210527, "acc_norm_stderr": 0.040179012759817494 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.46895424836601307, "acc_stderr": 0.020188804456361887, "acc_norm": 0.46895424836601307, "acc_norm_stderr": 0.020188804456361887 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.32978723404255317, "acc_stderr": 0.0280459469420424, "acc_norm": 0.32978723404255317, "acc_norm_stderr": 0.0280459469420424 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.35714285714285715, "acc_stderr": 0.04547960999764376, "acc_norm": 0.35714285714285715, "acc_norm_stderr": 0.04547960999764376 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4074074074074074, "acc_stderr": 0.03350991604696044, "acc_norm": 0.4074074074074074, "acc_norm_stderr": 0.03350991604696044 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2346368715083799, "acc_stderr": 0.014173044098303675, "acc_norm": 0.2346368715083799, "acc_norm_stderr": 0.014173044098303675 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.52, "acc_stderr": 0.050211673156867795, "acc_norm": 0.52, "acc_norm_stderr": 0.050211673156867795 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.62, "acc_stderr": 0.048783173121456316, "acc_norm": 0.62, "acc_norm_stderr": 0.048783173121456316 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4227941176470588, "acc_stderr": 0.030008562845003466, "acc_norm": 0.4227941176470588, "acc_norm_stderr": 0.030008562845003466 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.5877551020408164, "acc_stderr": 0.031512360446742695, "acc_norm": 0.5877551020408164, "acc_norm_stderr": 0.031512360446742695 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.7046413502109705, "acc_stderr": 0.029696338713422882, "acc_norm": 0.7046413502109705, "acc_norm_stderr": 0.029696338713422882 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.35658409387222945, "acc_stderr": 0.01223364298927389, "acc_norm": 0.35658409387222945, "acc_norm_stderr": 0.01223364298927389 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.6470588235294118, "acc_stderr": 0.03354092437591519, "acc_norm": 0.6470588235294118, "acc_norm_stderr": 0.03354092437591519 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.6181818181818182, "acc_stderr": 0.03793713171165633, "acc_norm": 0.6181818181818182, "acc_norm_stderr": 0.03793713171165633 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.31701346389228885, "mc1_stderr": 0.016289203374403392, "mc2": 0.48234441684091955, "mc2_stderr": 0.015365663323313775 }, "harness|ko_commongen_v2|2": { "acc": 0.5938606847697757, "acc_stderr": 0.016884749503191396, "acc_norm": 0.6033057851239669, "acc_norm_stderr": 0.016819438642971404 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Edentns/DataVortexS-10.7B-dpo-v1.7", "model_sha": "85af13a7e6002cee79c1b0be9cd0c93fd18d723e", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }