|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.27559726962457337, |
|
"acc_stderr": 0.01305716965576184, |
|
"acc_norm": 0.3216723549488055, |
|
"acc_norm_stderr": 0.013650488084494164 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3536148177653854, |
|
"acc_stderr": 0.004771143074426132, |
|
"acc_norm": 0.45120493925512845, |
|
"acc_norm_stderr": 0.004965963647210318 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.34502923976608185, |
|
"acc_stderr": 0.036459813773888065, |
|
"acc_norm": 0.34502923976608185, |
|
"acc_norm_stderr": 0.036459813773888065 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.1553398058252427, |
|
"acc_stderr": 0.03586594738573974, |
|
"acc_norm": 0.1553398058252427, |
|
"acc_norm_stderr": 0.03586594738573974 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.2388250319284802, |
|
"acc_stderr": 0.015246803197398698, |
|
"acc_norm": 0.2388250319284802, |
|
"acc_norm_stderr": 0.015246803197398698 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.25925925925925924, |
|
"acc_stderr": 0.03785714465066654, |
|
"acc_norm": 0.25925925925925924, |
|
"acc_norm_stderr": 0.03785714465066654 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.047258156262526045, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.3021276595744681, |
|
"acc_stderr": 0.030017554471880554, |
|
"acc_norm": 0.3021276595744681, |
|
"acc_norm_stderr": 0.030017554471880554 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.30120481927710846, |
|
"acc_stderr": 0.035716092300534796, |
|
"acc_norm": 0.30120481927710846, |
|
"acc_norm_stderr": 0.035716092300534796 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.21864951768488747, |
|
"acc_stderr": 0.023475581417861106, |
|
"acc_norm": 0.21864951768488747, |
|
"acc_norm_stderr": 0.023475581417861106 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.28699551569506726, |
|
"acc_stderr": 0.03036037971029196, |
|
"acc_norm": 0.28699551569506726, |
|
"acc_norm_stderr": 0.03036037971029196 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.26717557251908397, |
|
"acc_stderr": 0.03880848301082396, |
|
"acc_norm": 0.26717557251908397, |
|
"acc_norm_stderr": 0.03880848301082396 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.04461960433384741, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.04461960433384741 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.24242424242424243, |
|
"acc_stderr": 0.03053289223393202, |
|
"acc_norm": 0.24242424242424243, |
|
"acc_norm_stderr": 0.03053289223393202 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.23448275862068965, |
|
"acc_stderr": 0.035306258743465914, |
|
"acc_norm": 0.23448275862068965, |
|
"acc_norm_stderr": 0.035306258743465914 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.21568627450980393, |
|
"acc_stderr": 0.04092563958237653, |
|
"acc_norm": 0.21568627450980393, |
|
"acc_norm_stderr": 0.04092563958237653 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.23949579831932774, |
|
"acc_stderr": 0.027722065493361255, |
|
"acc_norm": 0.23949579831932774, |
|
"acc_norm_stderr": 0.027722065493361255 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.23076923076923078, |
|
"acc_stderr": 0.021362027725222717, |
|
"acc_norm": 0.23076923076923078, |
|
"acc_norm_stderr": 0.021362027725222717 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.044619604333847394, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.044619604333847394 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.047258156262526045, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.2222222222222222, |
|
"acc_stderr": 0.040191074725573483, |
|
"acc_norm": 0.2222222222222222, |
|
"acc_norm_stderr": 0.040191074725573483 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.21182266009852216, |
|
"acc_stderr": 0.028748983689941065, |
|
"acc_norm": 0.21182266009852216, |
|
"acc_norm_stderr": 0.028748983689941065 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.25483870967741934, |
|
"acc_stderr": 0.024790118459332208, |
|
"acc_norm": 0.25483870967741934, |
|
"acc_norm_stderr": 0.024790118459332208 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.2905982905982906, |
|
"acc_stderr": 0.029745048572674054, |
|
"acc_norm": 0.2905982905982906, |
|
"acc_norm_stderr": 0.029745048572674054 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.22641509433962265, |
|
"acc_stderr": 0.025757559893106727, |
|
"acc_norm": 0.22641509433962265, |
|
"acc_norm_stderr": 0.025757559893106727 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.2818181818181818, |
|
"acc_stderr": 0.04309118709946459, |
|
"acc_norm": 0.2818181818181818, |
|
"acc_norm_stderr": 0.04309118709946459 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.27037037037037037, |
|
"acc_stderr": 0.02708037281514566, |
|
"acc_norm": 0.27037037037037037, |
|
"acc_norm_stderr": 0.02708037281514566 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.33112582781456956, |
|
"acc_stderr": 0.038425817186598696, |
|
"acc_norm": 0.33112582781456956, |
|
"acc_norm_stderr": 0.038425817186598696 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.24378109452736318, |
|
"acc_stderr": 0.030360490154014645, |
|
"acc_norm": 0.24378109452736318, |
|
"acc_norm_stderr": 0.030360490154014645 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.21965317919075145, |
|
"acc_stderr": 0.031568093627031744, |
|
"acc_norm": 0.21965317919075145, |
|
"acc_norm_stderr": 0.031568093627031744 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.20634920634920634, |
|
"acc_stderr": 0.020842290930114662, |
|
"acc_norm": 0.20634920634920634, |
|
"acc_norm_stderr": 0.020842290930114662 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.2847222222222222, |
|
"acc_stderr": 0.037738099906869334, |
|
"acc_norm": 0.2847222222222222, |
|
"acc_norm_stderr": 0.037738099906869334 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.23, |
|
"acc_stderr": 0.04229525846816503, |
|
"acc_norm": 0.23, |
|
"acc_norm_stderr": 0.04229525846816503 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.048523658709391, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.048523658709391 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.28034682080924855, |
|
"acc_stderr": 0.024182427496577615, |
|
"acc_norm": 0.28034682080924855, |
|
"acc_norm_stderr": 0.024182427496577615 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.3067484662576687, |
|
"acc_stderr": 0.036230899157241474, |
|
"acc_norm": 0.3067484662576687, |
|
"acc_norm_stderr": 0.036230899157241474 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.25617283950617287, |
|
"acc_stderr": 0.024288533637726095, |
|
"acc_norm": 0.25617283950617287, |
|
"acc_norm_stderr": 0.024288533637726095 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.04512608598542129, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.04512608598542129 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.23316062176165803, |
|
"acc_stderr": 0.030516111371476005, |
|
"acc_norm": 0.23316062176165803, |
|
"acc_norm_stderr": 0.030516111371476005 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2543859649122807, |
|
"acc_stderr": 0.040969851398436716, |
|
"acc_norm": 0.2543859649122807, |
|
"acc_norm_stderr": 0.040969851398436716 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.23486238532110093, |
|
"acc_stderr": 0.018175110510343585, |
|
"acc_norm": 0.23486238532110093, |
|
"acc_norm_stderr": 0.018175110510343585 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.20634920634920634, |
|
"acc_stderr": 0.036196045241242494, |
|
"acc_norm": 0.20634920634920634, |
|
"acc_norm_stderr": 0.036196045241242494 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.28104575163398693, |
|
"acc_stderr": 0.025738854797818716, |
|
"acc_norm": 0.28104575163398693, |
|
"acc_norm_stderr": 0.025738854797818716 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.045604802157206845, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.045604802157206845 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.34710743801652894, |
|
"acc_stderr": 0.043457245702925335, |
|
"acc_norm": 0.34710743801652894, |
|
"acc_norm_stderr": 0.043457245702925335 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.19078947368421054, |
|
"acc_stderr": 0.031975658210325, |
|
"acc_norm": 0.19078947368421054, |
|
"acc_norm_stderr": 0.031975658210325 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.27124183006535946, |
|
"acc_stderr": 0.01798661530403032, |
|
"acc_norm": 0.27124183006535946, |
|
"acc_norm_stderr": 0.01798661530403032 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.24113475177304963, |
|
"acc_stderr": 0.025518731049537755, |
|
"acc_norm": 0.24113475177304963, |
|
"acc_norm_stderr": 0.025518731049537755 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.23214285714285715, |
|
"acc_stderr": 0.04007341809755808, |
|
"acc_norm": 0.23214285714285715, |
|
"acc_norm_stderr": 0.04007341809755808 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.3611111111111111, |
|
"acc_stderr": 0.03275773486100999, |
|
"acc_norm": 0.3611111111111111, |
|
"acc_norm_stderr": 0.03275773486100999 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.27150837988826815, |
|
"acc_stderr": 0.014874252168095278, |
|
"acc_norm": 0.27150837988826815, |
|
"acc_norm_stderr": 0.014874252168095278 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.044619604333847394, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.044619604333847394 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.43014705882352944, |
|
"acc_stderr": 0.030074971917302875, |
|
"acc_norm": 0.43014705882352944, |
|
"acc_norm_stderr": 0.030074971917302875 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.24081632653061225, |
|
"acc_stderr": 0.027372942201788153, |
|
"acc_norm": 0.24081632653061225, |
|
"acc_norm_stderr": 0.027372942201788153 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.2616033755274262, |
|
"acc_stderr": 0.028609516716994934, |
|
"acc_norm": 0.2616033755274262, |
|
"acc_norm_stderr": 0.028609516716994934 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.2692307692307692, |
|
"acc_stderr": 0.011328734403140313, |
|
"acc_norm": 0.2692307692307692, |
|
"acc_norm_stderr": 0.011328734403140313 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.29901960784313725, |
|
"acc_stderr": 0.03213325717373617, |
|
"acc_norm": 0.29901960784313725, |
|
"acc_norm_stderr": 0.03213325717373617 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.23636363636363636, |
|
"acc_stderr": 0.03317505930009179, |
|
"acc_norm": 0.23636363636363636, |
|
"acc_norm_stderr": 0.03317505930009179 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.29253365973072215, |
|
"mc1_stderr": 0.015925597445286165, |
|
"mc2": 0.43509454295381983, |
|
"mc2_stderr": 0.0148878635424864 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.23730814639905548, |
|
"acc_stderr": 0.014626677837186226, |
|
"acc_norm": 0.35064935064935066, |
|
"acc_norm_stderr": 0.016405556903893295 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "ITT-AF/ITT-42dot_LLM-PLM-1.3B-v3.0", |
|
"model_sha": "ccdfc183000dc328c5f64d6d9f70e2086de71cbf", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |