|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.4658703071672355, |
|
"acc_stderr": 0.014577311315231097, |
|
"acc_norm": 0.5247440273037542, |
|
"acc_norm_stderr": 0.01459348769493774 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.4598685520812587, |
|
"acc_stderr": 0.00497368302620218, |
|
"acc_norm": 0.6088428599880502, |
|
"acc_norm_stderr": 0.004870121051762735 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.6198830409356725, |
|
"acc_stderr": 0.03722965741385539, |
|
"acc_norm": 0.6198830409356725, |
|
"acc_norm_stderr": 0.03722965741385539 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6601941747572816, |
|
"acc_stderr": 0.04689765937278135, |
|
"acc_norm": 0.6601941747572816, |
|
"acc_norm_stderr": 0.04689765937278135 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.6500638569604087, |
|
"acc_stderr": 0.01705567979715042, |
|
"acc_norm": 0.6500638569604087, |
|
"acc_norm_stderr": 0.01705567979715042 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.5185185185185185, |
|
"acc_stderr": 0.043163785995113245, |
|
"acc_norm": 0.5185185185185185, |
|
"acc_norm_stderr": 0.043163785995113245 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.04512608598542127, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.04512608598542127 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.425531914893617, |
|
"acc_stderr": 0.03232146916224467, |
|
"acc_norm": 0.425531914893617, |
|
"acc_norm_stderr": 0.03232146916224467 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.39156626506024095, |
|
"acc_stderr": 0.03799857454479636, |
|
"acc_norm": 0.39156626506024095, |
|
"acc_norm_stderr": 0.03799857454479636 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.5466237942122186, |
|
"acc_stderr": 0.02827435985489424, |
|
"acc_norm": 0.5466237942122186, |
|
"acc_norm_stderr": 0.02827435985489424 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5426008968609866, |
|
"acc_stderr": 0.033435777055830646, |
|
"acc_norm": 0.5426008968609866, |
|
"acc_norm_stderr": 0.033435777055830646 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.48091603053435117, |
|
"acc_stderr": 0.04382094705550988, |
|
"acc_norm": 0.48091603053435117, |
|
"acc_norm_stderr": 0.04382094705550988 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.050251890762960605, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.050251890762960605 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.6414141414141414, |
|
"acc_stderr": 0.03416903640391521, |
|
"acc_norm": 0.6414141414141414, |
|
"acc_norm_stderr": 0.03416903640391521 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.46206896551724136, |
|
"acc_stderr": 0.041546596717075474, |
|
"acc_norm": 0.46206896551724136, |
|
"acc_norm_stderr": 0.041546596717075474 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.27450980392156865, |
|
"acc_stderr": 0.04440521906179327, |
|
"acc_norm": 0.27450980392156865, |
|
"acc_norm_stderr": 0.04440521906179327 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.4831932773109244, |
|
"acc_stderr": 0.03246013680375308, |
|
"acc_norm": 0.4831932773109244, |
|
"acc_norm_stderr": 0.03246013680375308 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.45384615384615384, |
|
"acc_stderr": 0.025242770987126163, |
|
"acc_norm": 0.45384615384615384, |
|
"acc_norm_stderr": 0.025242770987126163 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.6, |
|
"acc_stderr": 0.04923659639173309, |
|
"acc_norm": 0.6, |
|
"acc_norm_stderr": 0.04923659639173309 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.047937248544110196, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.047937248544110196 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6296296296296297, |
|
"acc_stderr": 0.04668408033024931, |
|
"acc_norm": 0.6296296296296297, |
|
"acc_norm_stderr": 0.04668408033024931 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.37438423645320196, |
|
"acc_stderr": 0.03405155380561952, |
|
"acc_norm": 0.37438423645320196, |
|
"acc_norm_stderr": 0.03405155380561952 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.5096774193548387, |
|
"acc_stderr": 0.02843867799890955, |
|
"acc_norm": 0.5096774193548387, |
|
"acc_norm_stderr": 0.02843867799890955 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7521367521367521, |
|
"acc_stderr": 0.028286324075564397, |
|
"acc_norm": 0.7521367521367521, |
|
"acc_norm_stderr": 0.028286324075564397 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.4679245283018868, |
|
"acc_stderr": 0.030709486992556555, |
|
"acc_norm": 0.4679245283018868, |
|
"acc_norm_stderr": 0.030709486992556555 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5363636363636364, |
|
"acc_stderr": 0.04776449162396197, |
|
"acc_norm": 0.5363636363636364, |
|
"acc_norm_stderr": 0.04776449162396197 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.29259259259259257, |
|
"acc_stderr": 0.027738969632176095, |
|
"acc_norm": 0.29259259259259257, |
|
"acc_norm_stderr": 0.027738969632176095 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.32450331125827814, |
|
"acc_stderr": 0.038227469376587525, |
|
"acc_norm": 0.32450331125827814, |
|
"acc_norm_stderr": 0.038227469376587525 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6666666666666666, |
|
"acc_stderr": 0.03333333333333334, |
|
"acc_norm": 0.6666666666666666, |
|
"acc_norm_stderr": 0.03333333333333334 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.41040462427745666, |
|
"acc_stderr": 0.03750757044895538, |
|
"acc_norm": 0.41040462427745666, |
|
"acc_norm_stderr": 0.03750757044895538 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.335978835978836, |
|
"acc_stderr": 0.024326310529149138, |
|
"acc_norm": 0.335978835978836, |
|
"acc_norm_stderr": 0.024326310529149138 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4513888888888889, |
|
"acc_stderr": 0.04161402398403279, |
|
"acc_norm": 0.4513888888888889, |
|
"acc_norm_stderr": 0.04161402398403279 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.04923659639173309, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.04923659639173309 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.65, |
|
"acc_stderr": 0.0479372485441102, |
|
"acc_norm": 0.65, |
|
"acc_norm_stderr": 0.0479372485441102 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5173410404624278, |
|
"acc_stderr": 0.02690290045866664, |
|
"acc_norm": 0.5173410404624278, |
|
"acc_norm_stderr": 0.02690290045866664 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.37423312883435583, |
|
"acc_stderr": 0.03802068102899615, |
|
"acc_norm": 0.37423312883435583, |
|
"acc_norm_stderr": 0.03802068102899615 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.5432098765432098, |
|
"acc_stderr": 0.02771666165019404, |
|
"acc_norm": 0.5432098765432098, |
|
"acc_norm_stderr": 0.02771666165019404 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.6113989637305699, |
|
"acc_stderr": 0.03517739796373131, |
|
"acc_norm": 0.6113989637305699, |
|
"acc_norm_stderr": 0.03517739796373131 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.32456140350877194, |
|
"acc_stderr": 0.04404556157374767, |
|
"acc_norm": 0.32456140350877194, |
|
"acc_norm_stderr": 0.04404556157374767 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.5853211009174312, |
|
"acc_stderr": 0.02112290320860259, |
|
"acc_norm": 0.5853211009174312, |
|
"acc_norm_stderr": 0.02112290320860259 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.35714285714285715, |
|
"acc_stderr": 0.04285714285714281, |
|
"acc_norm": 0.35714285714285715, |
|
"acc_norm_stderr": 0.04285714285714281 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.4934640522875817, |
|
"acc_stderr": 0.028627470550556047, |
|
"acc_norm": 0.4934640522875817, |
|
"acc_norm_stderr": 0.028627470550556047 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.55, |
|
"acc_stderr": 0.04999999999999999, |
|
"acc_norm": 0.55, |
|
"acc_norm_stderr": 0.04999999999999999 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.5950413223140496, |
|
"acc_stderr": 0.04481137755942469, |
|
"acc_norm": 0.5950413223140496, |
|
"acc_norm_stderr": 0.04481137755942469 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5394736842105263, |
|
"acc_stderr": 0.04056242252249034, |
|
"acc_norm": 0.5394736842105263, |
|
"acc_norm_stderr": 0.04056242252249034 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.44607843137254904, |
|
"acc_stderr": 0.020109864547181357, |
|
"acc_norm": 0.44607843137254904, |
|
"acc_norm_stderr": 0.020109864547181357 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3617021276595745, |
|
"acc_stderr": 0.028663820147199495, |
|
"acc_norm": 0.3617021276595745, |
|
"acc_norm_stderr": 0.028663820147199495 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.3125, |
|
"acc_stderr": 0.043994650575715215, |
|
"acc_norm": 0.3125, |
|
"acc_norm_stderr": 0.043994650575715215 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.3425925925925926, |
|
"acc_stderr": 0.032365852526021574, |
|
"acc_norm": 0.3425925925925926, |
|
"acc_norm_stderr": 0.032365852526021574 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.29608938547486036, |
|
"acc_stderr": 0.0152686773176023, |
|
"acc_norm": 0.29608938547486036, |
|
"acc_norm_stderr": 0.0152686773176023 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.047258156262526045, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.56, |
|
"acc_stderr": 0.04988876515698589, |
|
"acc_norm": 0.56, |
|
"acc_norm_stderr": 0.04988876515698589 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.3860294117647059, |
|
"acc_stderr": 0.029573269134411124, |
|
"acc_norm": 0.3860294117647059, |
|
"acc_norm_stderr": 0.029573269134411124 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.46530612244897956, |
|
"acc_stderr": 0.03193207024425314, |
|
"acc_norm": 0.46530612244897956, |
|
"acc_norm_stderr": 0.03193207024425314 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.6286919831223629, |
|
"acc_stderr": 0.0314506860074486, |
|
"acc_norm": 0.6286919831223629, |
|
"acc_norm_stderr": 0.0314506860074486 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3318122555410691, |
|
"acc_stderr": 0.012026088259897637, |
|
"acc_norm": 0.3318122555410691, |
|
"acc_norm_stderr": 0.012026088259897637 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.5588235294117647, |
|
"acc_stderr": 0.034849415144292316, |
|
"acc_norm": 0.5588235294117647, |
|
"acc_norm_stderr": 0.034849415144292316 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.5515151515151515, |
|
"acc_stderr": 0.038835659779569286, |
|
"acc_norm": 0.5515151515151515, |
|
"acc_norm_stderr": 0.038835659779569286 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.37576499388004897, |
|
"mc1_stderr": 0.016954584060214307, |
|
"mc2": 0.5536468577288562, |
|
"mc2_stderr": 0.016032830714481056 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5903187721369539, |
|
"acc_stderr": 0.016907568192219478, |
|
"acc_norm": 0.6092089728453365, |
|
"acc_norm_stderr": 0.01677529846510826 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "GAI-LLM/OPEN-SOLAR-KO-10.7B-dpo-v1", |
|
"model_sha": "66f56a656d085c813ad58226a21a3b951df8543c", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |