{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.36006825938566556, "acc_stderr": 0.01402751681458519, "acc_norm": 0.4786689419795222, "acc_norm_stderr": 0.014598087973127108 }, "harness|ko_hellaswag|10": { "acc": 0.3886675960963951, "acc_stderr": 0.004864513262194307, "acc_norm": 0.5717984465245967, "acc_norm_stderr": 0.004938068627349492 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.6491228070175439, "acc_stderr": 0.03660298834049163, "acc_norm": 0.6491228070175439, "acc_norm_stderr": 0.03660298834049163 }, "harness|ko_mmlu_management|5": { "acc": 0.6116504854368932, "acc_stderr": 0.04825729337356389, "acc_norm": 0.6116504854368932, "acc_norm_stderr": 0.04825729337356389 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.6704980842911877, "acc_stderr": 0.016808322261740442, "acc_norm": 0.6704980842911877, "acc_norm_stderr": 0.016808322261740442 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4, "acc_stderr": 0.04232073695151589, "acc_norm": 0.4, "acc_norm_stderr": 0.04232073695151589 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.15, "acc_stderr": 0.03588702812826372, "acc_norm": 0.15, "acc_norm_stderr": 0.03588702812826372 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.5361702127659574, "acc_stderr": 0.03260038511835771, "acc_norm": 0.5361702127659574, "acc_norm_stderr": 0.03260038511835771 }, "harness|ko_mmlu_virology|5": { "acc": 0.4939759036144578, "acc_stderr": 0.03892212195333047, "acc_norm": 0.4939759036144578, "acc_norm_stderr": 0.03892212195333047 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.6430868167202572, "acc_stderr": 0.02721042037593402, "acc_norm": 0.6430868167202572, "acc_norm_stderr": 0.02721042037593402 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.5739910313901345, "acc_stderr": 0.03318833286217281, "acc_norm": 0.5739910313901345, "acc_norm_stderr": 0.03318833286217281 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.6641221374045801, "acc_stderr": 0.04142313771996664, "acc_norm": 0.6641221374045801, "acc_norm_stderr": 0.04142313771996664 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.53, "acc_stderr": 0.050161355804659205, "acc_norm": 0.53, "acc_norm_stderr": 0.050161355804659205 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.7070707070707071, "acc_stderr": 0.032424979581788166, "acc_norm": 0.7070707070707071, "acc_norm_stderr": 0.032424979581788166 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.5448275862068965, "acc_stderr": 0.04149886942192118, "acc_norm": 0.5448275862068965, "acc_norm_stderr": 0.04149886942192118 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.3431372549019608, "acc_stderr": 0.04724007352383888, "acc_norm": 0.3431372549019608, "acc_norm_stderr": 0.04724007352383888 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.6302521008403361, "acc_stderr": 0.03135709599613591, "acc_norm": 0.6302521008403361, "acc_norm_stderr": 0.03135709599613591 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.5666666666666667, "acc_stderr": 0.0251246535258851, "acc_norm": 0.5666666666666667, "acc_norm_stderr": 0.0251246535258851 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.58, "acc_stderr": 0.04960449637488583, "acc_norm": 0.58, "acc_norm_stderr": 0.04960449637488583 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.29, "acc_stderr": 0.045604802157206845, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206845 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.6851851851851852, "acc_stderr": 0.04489931073591312, "acc_norm": 0.6851851851851852, "acc_norm_stderr": 0.04489931073591312 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3793103448275862, "acc_stderr": 0.034139638059062345, "acc_norm": 0.3793103448275862, "acc_norm_stderr": 0.034139638059062345 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.6064516129032258, "acc_stderr": 0.02779187875313226, "acc_norm": 0.6064516129032258, "acc_norm_stderr": 0.02779187875313226 }, "harness|ko_mmlu_marketing|5": { "acc": 0.8162393162393162, "acc_stderr": 0.025372139671722933, "acc_norm": 0.8162393162393162, "acc_norm_stderr": 0.025372139671722933 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.5660377358490566, "acc_stderr": 0.03050329201334259, "acc_norm": 0.5660377358490566, "acc_norm_stderr": 0.03050329201334259 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.6272727272727273, "acc_stderr": 0.04631381319425465, "acc_norm": 0.6272727272727273, "acc_norm_stderr": 0.04631381319425465 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.3074074074074074, "acc_stderr": 0.028133252578815635, "acc_norm": 0.3074074074074074, "acc_norm_stderr": 0.028133252578815635 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.37748344370860926, "acc_stderr": 0.039580272311215706, "acc_norm": 0.37748344370860926, "acc_norm_stderr": 0.039580272311215706 }, "harness|ko_mmlu_sociology|5": { "acc": 0.7014925373134329, "acc_stderr": 0.032357437893550445, "acc_norm": 0.7014925373134329, "acc_norm_stderr": 0.032357437893550445 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.5375722543352601, "acc_stderr": 0.0380168510452446, "acc_norm": 0.5375722543352601, "acc_norm_stderr": 0.0380168510452446 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.4312169312169312, "acc_stderr": 0.025506481698138215, "acc_norm": 0.4312169312169312, "acc_norm_stderr": 0.025506481698138215 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.5138888888888888, "acc_stderr": 0.041795966175810016, "acc_norm": 0.5138888888888888, "acc_norm_stderr": 0.041795966175810016 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.7, "acc_stderr": 0.046056618647183814, "acc_norm": 0.7, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.6242774566473989, "acc_stderr": 0.02607431485165708, "acc_norm": 0.6242774566473989, "acc_norm_stderr": 0.02607431485165708 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.5276073619631901, "acc_stderr": 0.0392237829061099, "acc_norm": 0.5276073619631901, "acc_norm_stderr": 0.0392237829061099 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.5895061728395061, "acc_stderr": 0.027371350925124764, "acc_norm": 0.5895061728395061, "acc_norm_stderr": 0.027371350925124764 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.7046632124352331, "acc_stderr": 0.032922966391551414, "acc_norm": 0.7046632124352331, "acc_norm_stderr": 0.032922966391551414 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.47368421052631576, "acc_stderr": 0.046970851366478626, "acc_norm": 0.47368421052631576, "acc_norm_stderr": 0.046970851366478626 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.6880733944954128, "acc_stderr": 0.019862967976707245, "acc_norm": 0.6880733944954128, "acc_norm_stderr": 0.019862967976707245 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.4444444444444444, "acc_stderr": 0.044444444444444495, "acc_norm": 0.4444444444444444, "acc_norm_stderr": 0.044444444444444495 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5980392156862745, "acc_stderr": 0.02807415894760066, "acc_norm": 0.5980392156862745, "acc_norm_stderr": 0.02807415894760066 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.57, "acc_stderr": 0.049756985195624284, "acc_norm": 0.57, "acc_norm_stderr": 0.049756985195624284 }, "harness|ko_mmlu_international_law|5": { "acc": 0.7603305785123967, "acc_stderr": 0.03896878985070417, "acc_norm": 0.7603305785123967, "acc_norm_stderr": 0.03896878985070417 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.5723684210526315, "acc_stderr": 0.04026097083296563, "acc_norm": 0.5723684210526315, "acc_norm_stderr": 0.04026097083296563 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.5098039215686274, "acc_stderr": 0.02022394600507432, "acc_norm": 0.5098039215686274, "acc_norm_stderr": 0.02022394600507432 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.4148936170212766, "acc_stderr": 0.029392236584612493, "acc_norm": 0.4148936170212766, "acc_norm_stderr": 0.029392236584612493 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.41964285714285715, "acc_stderr": 0.04684099321077106, "acc_norm": 0.41964285714285715, "acc_norm_stderr": 0.04684099321077106 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.5694444444444444, "acc_stderr": 0.03376922151252336, "acc_norm": 0.5694444444444444, "acc_norm_stderr": 0.03376922151252336 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2871508379888268, "acc_stderr": 0.01513160884996376, "acc_norm": 0.2871508379888268, "acc_norm_stderr": 0.01513160884996376 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.49, "acc_stderr": 0.05024183937956912, "acc_norm": 0.49, "acc_norm_stderr": 0.05024183937956912 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.71, "acc_stderr": 0.045604802157206845, "acc_norm": 0.71, "acc_norm_stderr": 0.045604802157206845 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.5036764705882353, "acc_stderr": 0.0303720158854282, "acc_norm": 0.5036764705882353, "acc_norm_stderr": 0.0303720158854282 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.6040816326530613, "acc_stderr": 0.03130802899065686, "acc_norm": 0.6040816326530613, "acc_norm_stderr": 0.03130802899065686 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.7510548523206751, "acc_stderr": 0.028146970599422644, "acc_norm": 0.7510548523206751, "acc_norm_stderr": 0.028146970599422644 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.42046936114732725, "acc_stderr": 0.012607654553832705, "acc_norm": 0.42046936114732725, "acc_norm_stderr": 0.012607654553832705 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.696078431372549, "acc_stderr": 0.03228210387037891, "acc_norm": 0.696078431372549, "acc_norm_stderr": 0.03228210387037891 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.703030303030303, "acc_stderr": 0.035679697722680474, "acc_norm": 0.703030303030303, "acc_norm_stderr": 0.035679697722680474 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.3488372093023256, "mc1_stderr": 0.01668441985998688, "mc2": 0.5363520506790478, "mc2_stderr": 0.01620621230829838 }, "harness|ko_commongen_v2|2": { "acc": 0.48642266824085006, "acc_stderr": 0.01718401506040145, "acc_norm": 0.525383707201889, "acc_norm_stderr": 0.017168187201429246 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Edentns/DataVortexS-10.7B-dpo-v0.1", "model_sha": "76fafe2757d64b083d4f20b46798d9592827aacf", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }