|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.20733788395904437, |
|
"acc_stderr": 0.011846905782971368, |
|
"acc_norm": 0.2440273037542662, |
|
"acc_norm_stderr": 0.012551447627856259 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.2900816570404302, |
|
"acc_stderr": 0.004528723951878253, |
|
"acc_norm": 0.32772356104361683, |
|
"acc_norm_stderr": 0.004684241685200313 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.30409356725146197, |
|
"acc_stderr": 0.03528211258245232, |
|
"acc_norm": 0.30409356725146197, |
|
"acc_norm_stderr": 0.03528211258245232 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.17475728155339806, |
|
"acc_stderr": 0.037601780060266196, |
|
"acc_norm": 0.17475728155339806, |
|
"acc_norm_stderr": 0.037601780060266196 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.280970625798212, |
|
"acc_stderr": 0.01607312785122125, |
|
"acc_norm": 0.280970625798212, |
|
"acc_norm_stderr": 0.01607312785122125 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.34074074074074073, |
|
"acc_stderr": 0.040943762699967946, |
|
"acc_norm": 0.34074074074074073, |
|
"acc_norm_stderr": 0.040943762699967946 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.22, |
|
"acc_stderr": 0.04163331998932268, |
|
"acc_norm": 0.22, |
|
"acc_norm_stderr": 0.04163331998932268 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.26382978723404255, |
|
"acc_stderr": 0.02880998985410299, |
|
"acc_norm": 0.26382978723404255, |
|
"acc_norm_stderr": 0.02880998985410299 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.28313253012048195, |
|
"acc_stderr": 0.03507295431370518, |
|
"acc_norm": 0.28313253012048195, |
|
"acc_norm_stderr": 0.03507295431370518 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.26366559485530544, |
|
"acc_stderr": 0.02502553850053234, |
|
"acc_norm": 0.26366559485530544, |
|
"acc_norm_stderr": 0.02502553850053234 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.2062780269058296, |
|
"acc_stderr": 0.027157150479563824, |
|
"acc_norm": 0.2062780269058296, |
|
"acc_norm_stderr": 0.027157150479563824 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.2595419847328244, |
|
"acc_stderr": 0.03844876139785271, |
|
"acc_norm": 0.2595419847328244, |
|
"acc_norm_stderr": 0.03844876139785271 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.045604802157206845, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.045604802157206845 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.18686868686868688, |
|
"acc_stderr": 0.027772533334218967, |
|
"acc_norm": 0.18686868686868688, |
|
"acc_norm_stderr": 0.027772533334218967 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.22758620689655173, |
|
"acc_stderr": 0.03493950380131184, |
|
"acc_norm": 0.22758620689655173, |
|
"acc_norm_stderr": 0.03493950380131184 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.18627450980392157, |
|
"acc_stderr": 0.038739587141493524, |
|
"acc_norm": 0.18627450980392157, |
|
"acc_norm_stderr": 0.038739587141493524 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.25630252100840334, |
|
"acc_stderr": 0.028359620870533946, |
|
"acc_norm": 0.25630252100840334, |
|
"acc_norm_stderr": 0.028359620870533946 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.24358974358974358, |
|
"acc_stderr": 0.021763733684173933, |
|
"acc_norm": 0.24358974358974358, |
|
"acc_norm_stderr": 0.021763733684173933 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.19, |
|
"acc_stderr": 0.03942772444036625, |
|
"acc_norm": 0.19, |
|
"acc_norm_stderr": 0.03942772444036625 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.044619604333847394, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.044619604333847394 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04186091791394607, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04186091791394607 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.18719211822660098, |
|
"acc_stderr": 0.027444924966882618, |
|
"acc_norm": 0.18719211822660098, |
|
"acc_norm_stderr": 0.027444924966882618 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.25806451612903225, |
|
"acc_stderr": 0.024892469172462833, |
|
"acc_norm": 0.25806451612903225, |
|
"acc_norm_stderr": 0.024892469172462833 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.23931623931623933, |
|
"acc_stderr": 0.02795182680892433, |
|
"acc_norm": 0.23931623931623933, |
|
"acc_norm_stderr": 0.02795182680892433 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.2188679245283019, |
|
"acc_stderr": 0.025447863825108608, |
|
"acc_norm": 0.2188679245283019, |
|
"acc_norm_stderr": 0.025447863825108608 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.20909090909090908, |
|
"acc_stderr": 0.038950910157241364, |
|
"acc_norm": 0.20909090909090908, |
|
"acc_norm_stderr": 0.038950910157241364 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.2740740740740741, |
|
"acc_stderr": 0.027195934804085622, |
|
"acc_norm": 0.2740740740740741, |
|
"acc_norm_stderr": 0.027195934804085622 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.271523178807947, |
|
"acc_stderr": 0.03631329803969653, |
|
"acc_norm": 0.271523178807947, |
|
"acc_norm_stderr": 0.03631329803969653 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.263681592039801, |
|
"acc_stderr": 0.03115715086935558, |
|
"acc_norm": 0.263681592039801, |
|
"acc_norm_stderr": 0.03115715086935558 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.2254335260115607, |
|
"acc_stderr": 0.03186209851641145, |
|
"acc_norm": 0.2254335260115607, |
|
"acc_norm_stderr": 0.03186209851641145 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.2724867724867725, |
|
"acc_stderr": 0.022930973071633345, |
|
"acc_norm": 0.2724867724867725, |
|
"acc_norm_stderr": 0.022930973071633345 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.20833333333333334, |
|
"acc_stderr": 0.03396116205845335, |
|
"acc_norm": 0.20833333333333334, |
|
"acc_norm_stderr": 0.03396116205845335 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.040201512610368445, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.040201512610368445 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.04605661864718381, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.04605661864718381 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.24855491329479767, |
|
"acc_stderr": 0.023267528432100174, |
|
"acc_norm": 0.24855491329479767, |
|
"acc_norm_stderr": 0.023267528432100174 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.2883435582822086, |
|
"acc_stderr": 0.035590395316173425, |
|
"acc_norm": 0.2883435582822086, |
|
"acc_norm_stderr": 0.035590395316173425 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.21604938271604937, |
|
"acc_stderr": 0.022899162918445796, |
|
"acc_norm": 0.21604938271604937, |
|
"acc_norm_stderr": 0.022899162918445796 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.3160621761658031, |
|
"acc_stderr": 0.033553973696861736, |
|
"acc_norm": 0.3160621761658031, |
|
"acc_norm_stderr": 0.033553973696861736 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2543859649122807, |
|
"acc_stderr": 0.040969851398436716, |
|
"acc_norm": 0.2543859649122807, |
|
"acc_norm_stderr": 0.040969851398436716 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.22201834862385322, |
|
"acc_stderr": 0.017818849564796627, |
|
"acc_norm": 0.22201834862385322, |
|
"acc_norm_stderr": 0.017818849564796627 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.21428571428571427, |
|
"acc_stderr": 0.03670066451047181, |
|
"acc_norm": 0.21428571428571427, |
|
"acc_norm_stderr": 0.03670066451047181 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.21241830065359477, |
|
"acc_stderr": 0.023420375478296136, |
|
"acc_norm": 0.21241830065359477, |
|
"acc_norm_stderr": 0.023420375478296136 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.24, |
|
"acc_stderr": 0.04292346959909282, |
|
"acc_norm": 0.24, |
|
"acc_norm_stderr": 0.04292346959909282 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.2396694214876033, |
|
"acc_stderr": 0.038968789850704164, |
|
"acc_norm": 0.2396694214876033, |
|
"acc_norm_stderr": 0.038968789850704164 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.17763157894736842, |
|
"acc_stderr": 0.03110318238312338, |
|
"acc_norm": 0.17763157894736842, |
|
"acc_norm_stderr": 0.03110318238312338 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.25163398692810457, |
|
"acc_stderr": 0.01755581809132227, |
|
"acc_norm": 0.25163398692810457, |
|
"acc_norm_stderr": 0.01755581809132227 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.23049645390070922, |
|
"acc_stderr": 0.02512373922687241, |
|
"acc_norm": 0.23049645390070922, |
|
"acc_norm_stderr": 0.02512373922687241 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.19642857142857142, |
|
"acc_stderr": 0.03770970049347018, |
|
"acc_norm": 0.19642857142857142, |
|
"acc_norm_stderr": 0.03770970049347018 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.39351851851851855, |
|
"acc_stderr": 0.03331747876370312, |
|
"acc_norm": 0.39351851851851855, |
|
"acc_norm_stderr": 0.03331747876370312 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2435754189944134, |
|
"acc_stderr": 0.014355911964767867, |
|
"acc_norm": 0.2435754189944134, |
|
"acc_norm_stderr": 0.014355911964767867 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252604, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252604 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.4375, |
|
"acc_stderr": 0.030134614954403924, |
|
"acc_norm": 0.4375, |
|
"acc_norm_stderr": 0.030134614954403924 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.19591836734693877, |
|
"acc_stderr": 0.025409301953225678, |
|
"acc_norm": 0.19591836734693877, |
|
"acc_norm_stderr": 0.025409301953225678 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.2869198312236287, |
|
"acc_stderr": 0.029443773022594693, |
|
"acc_norm": 0.2869198312236287, |
|
"acc_norm_stderr": 0.029443773022594693 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.24511082138200782, |
|
"acc_stderr": 0.010986307870045519, |
|
"acc_norm": 0.24511082138200782, |
|
"acc_norm_stderr": 0.010986307870045519 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.25980392156862747, |
|
"acc_stderr": 0.030778554678693257, |
|
"acc_norm": 0.25980392156862747, |
|
"acc_norm_stderr": 0.030778554678693257 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.24242424242424243, |
|
"acc_stderr": 0.033464098810559534, |
|
"acc_norm": 0.24242424242424243, |
|
"acc_norm_stderr": 0.033464098810559534 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2533659730722154, |
|
"mc1_stderr": 0.015225899340826824, |
|
"mc2": 0.44746742387165916, |
|
"mc2_stderr": 0.015687397901643654 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.3435655253837072, |
|
"acc_stderr": 0.016327334806429134, |
|
"acc_norm": 0.4510035419126328, |
|
"acc_norm_stderr": 0.017107618859549357 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "blueapple8259/ANHSY_test2", |
|
"model_sha": "bb2cfbebcdf945d14ec8f53215d9f9b5bbc96742", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |