|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.22098976109215018, |
|
"acc_stderr": 0.01212492920681826, |
|
"acc_norm": 0.2790102389078498, |
|
"acc_norm_stderr": 0.013106784883601346 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.27703644692292373, |
|
"acc_stderr": 0.004466200055292544, |
|
"acc_norm": 0.3209520015933081, |
|
"acc_norm_stderr": 0.004658882929099516 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.18128654970760233, |
|
"acc_stderr": 0.029547741687640024, |
|
"acc_norm": 0.18128654970760233, |
|
"acc_norm_stderr": 0.029547741687640024 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.3786407766990291, |
|
"acc_stderr": 0.04802694698258973, |
|
"acc_norm": 0.3786407766990291, |
|
"acc_norm_stderr": 0.04802694698258973 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.20945083014048532, |
|
"acc_stderr": 0.014551310568143698, |
|
"acc_norm": 0.20945083014048532, |
|
"acc_norm_stderr": 0.014551310568143698 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.22962962962962963, |
|
"acc_stderr": 0.036333844140734636, |
|
"acc_norm": 0.22962962962962963, |
|
"acc_norm_stderr": 0.036333844140734636 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.04512608598542128, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.04512608598542128 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.23404255319148937, |
|
"acc_stderr": 0.027678452578212377, |
|
"acc_norm": 0.23404255319148937, |
|
"acc_norm_stderr": 0.027678452578212377 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.19879518072289157, |
|
"acc_stderr": 0.031069390260789396, |
|
"acc_norm": 0.19879518072289157, |
|
"acc_norm_stderr": 0.031069390260789396 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.24115755627009647, |
|
"acc_stderr": 0.024296594034763426, |
|
"acc_norm": 0.24115755627009647, |
|
"acc_norm_stderr": 0.024296594034763426 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.11210762331838565, |
|
"acc_stderr": 0.0211748942063461, |
|
"acc_norm": 0.11210762331838565, |
|
"acc_norm_stderr": 0.0211748942063461 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.2748091603053435, |
|
"acc_stderr": 0.03915345408847836, |
|
"acc_norm": 0.2748091603053435, |
|
"acc_norm_stderr": 0.03915345408847836 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.04408440022768079, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.04408440022768079 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.35353535353535354, |
|
"acc_stderr": 0.03406086723547153, |
|
"acc_norm": 0.35353535353535354, |
|
"acc_norm_stderr": 0.03406086723547153 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.2413793103448276, |
|
"acc_stderr": 0.03565998174135302, |
|
"acc_norm": 0.2413793103448276, |
|
"acc_norm_stderr": 0.03565998174135302 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.37254901960784315, |
|
"acc_stderr": 0.048108401480826346, |
|
"acc_norm": 0.37254901960784315, |
|
"acc_norm_stderr": 0.048108401480826346 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.3445378151260504, |
|
"acc_stderr": 0.030868682604121622, |
|
"acc_norm": 0.3445378151260504, |
|
"acc_norm_stderr": 0.030868682604121622 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.36923076923076925, |
|
"acc_stderr": 0.024468615241478912, |
|
"acc_norm": 0.36923076923076925, |
|
"acc_norm_stderr": 0.024468615241478912 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.22, |
|
"acc_stderr": 0.0416333199893227, |
|
"acc_norm": 0.22, |
|
"acc_norm_stderr": 0.0416333199893227 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.19, |
|
"acc_stderr": 0.039427724440366234, |
|
"acc_norm": 0.19, |
|
"acc_norm_stderr": 0.039427724440366234 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.2222222222222222, |
|
"acc_stderr": 0.040191074725573483, |
|
"acc_norm": 0.2222222222222222, |
|
"acc_norm_stderr": 0.040191074725573483 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.28078817733990147, |
|
"acc_stderr": 0.0316185633535861, |
|
"acc_norm": 0.28078817733990147, |
|
"acc_norm_stderr": 0.0316185633535861 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.3096774193548387, |
|
"acc_stderr": 0.026302774983517418, |
|
"acc_norm": 0.3096774193548387, |
|
"acc_norm_stderr": 0.026302774983517418 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.20512820512820512, |
|
"acc_stderr": 0.02645350805404033, |
|
"acc_norm": 0.20512820512820512, |
|
"acc_norm_stderr": 0.02645350805404033 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.3169811320754717, |
|
"acc_stderr": 0.02863723563980092, |
|
"acc_norm": 0.3169811320754717, |
|
"acc_norm_stderr": 0.02863723563980092 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.22727272727272727, |
|
"acc_stderr": 0.040139645540727735, |
|
"acc_norm": 0.22727272727272727, |
|
"acc_norm_stderr": 0.040139645540727735 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.26296296296296295, |
|
"acc_stderr": 0.02684205787383371, |
|
"acc_norm": 0.26296296296296295, |
|
"acc_norm_stderr": 0.02684205787383371 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.32450331125827814, |
|
"acc_stderr": 0.03822746937658754, |
|
"acc_norm": 0.32450331125827814, |
|
"acc_norm_stderr": 0.03822746937658754 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.26865671641791045, |
|
"acc_stderr": 0.03134328358208954, |
|
"acc_norm": 0.26865671641791045, |
|
"acc_norm_stderr": 0.03134328358208954 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.28901734104046245, |
|
"acc_stderr": 0.034564257450869995, |
|
"acc_norm": 0.28901734104046245, |
|
"acc_norm_stderr": 0.034564257450869995 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.2724867724867725, |
|
"acc_stderr": 0.022930973071633335, |
|
"acc_norm": 0.2724867724867725, |
|
"acc_norm_stderr": 0.022930973071633335 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.2569444444444444, |
|
"acc_stderr": 0.03653946969442099, |
|
"acc_norm": 0.2569444444444444, |
|
"acc_norm_stderr": 0.03653946969442099 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.044084400227680794, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.044084400227680794 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.2745664739884393, |
|
"acc_stderr": 0.02402774515526501, |
|
"acc_norm": 0.2745664739884393, |
|
"acc_norm_stderr": 0.02402774515526501 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.3006134969325153, |
|
"acc_stderr": 0.03602511318806771, |
|
"acc_norm": 0.3006134969325153, |
|
"acc_norm_stderr": 0.03602511318806771 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.2345679012345679, |
|
"acc_stderr": 0.02357688174400572, |
|
"acc_norm": 0.2345679012345679, |
|
"acc_norm_stderr": 0.02357688174400572 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.37823834196891193, |
|
"acc_stderr": 0.03499807276193338, |
|
"acc_norm": 0.37823834196891193, |
|
"acc_norm_stderr": 0.03499807276193338 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.23684210526315788, |
|
"acc_stderr": 0.039994238792813365, |
|
"acc_norm": 0.23684210526315788, |
|
"acc_norm_stderr": 0.039994238792813365 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.3486238532110092, |
|
"acc_stderr": 0.020431254090714328, |
|
"acc_norm": 0.3486238532110092, |
|
"acc_norm_stderr": 0.020431254090714328 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.373015873015873, |
|
"acc_stderr": 0.04325506042017086, |
|
"acc_norm": 0.373015873015873, |
|
"acc_norm_stderr": 0.04325506042017086 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.3006535947712418, |
|
"acc_stderr": 0.02625605383571896, |
|
"acc_norm": 0.3006535947712418, |
|
"acc_norm_stderr": 0.02625605383571896 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.22, |
|
"acc_stderr": 0.0416333199893227, |
|
"acc_norm": 0.22, |
|
"acc_norm_stderr": 0.0416333199893227 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.18181818181818182, |
|
"acc_stderr": 0.03520893951097653, |
|
"acc_norm": 0.18181818181818182, |
|
"acc_norm_stderr": 0.03520893951097653 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.32894736842105265, |
|
"acc_stderr": 0.03823428969926605, |
|
"acc_norm": 0.32894736842105265, |
|
"acc_norm_stderr": 0.03823428969926605 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.22058823529411764, |
|
"acc_stderr": 0.016774672365468517, |
|
"acc_norm": 0.22058823529411764, |
|
"acc_norm_stderr": 0.016774672365468517 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.24113475177304963, |
|
"acc_stderr": 0.02551873104953776, |
|
"acc_norm": 0.24113475177304963, |
|
"acc_norm_stderr": 0.02551873104953776 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.16071428571428573, |
|
"acc_stderr": 0.034859460964757394, |
|
"acc_norm": 0.16071428571428573, |
|
"acc_norm_stderr": 0.034859460964757394 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.4722222222222222, |
|
"acc_stderr": 0.0340470532865388, |
|
"acc_norm": 0.4722222222222222, |
|
"acc_norm_stderr": 0.0340470532865388 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.27262569832402234, |
|
"acc_stderr": 0.014893391735249608, |
|
"acc_norm": 0.27262569832402234, |
|
"acc_norm_stderr": 0.014893391735249608 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.048241815132442176, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.048241815132442176 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.04020151261036846, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.04020151261036846 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.4485294117647059, |
|
"acc_stderr": 0.030211479609121593, |
|
"acc_norm": 0.4485294117647059, |
|
"acc_norm_stderr": 0.030211479609121593 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.40408163265306124, |
|
"acc_stderr": 0.0314147080258659, |
|
"acc_norm": 0.40408163265306124, |
|
"acc_norm_stderr": 0.0314147080258659 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.20253164556962025, |
|
"acc_stderr": 0.026160568246601464, |
|
"acc_norm": 0.20253164556962025, |
|
"acc_norm_stderr": 0.026160568246601464 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.2457627118644068, |
|
"acc_stderr": 0.01099615663514269, |
|
"acc_norm": 0.2457627118644068, |
|
"acc_norm_stderr": 0.01099615663514269 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.23039215686274508, |
|
"acc_stderr": 0.029554292605695053, |
|
"acc_norm": 0.23039215686274508, |
|
"acc_norm_stderr": 0.029554292605695053 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.2545454545454545, |
|
"acc_stderr": 0.03401506715249039, |
|
"acc_norm": 0.2545454545454545, |
|
"acc_norm_stderr": 0.03401506715249039 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2594859241126071, |
|
"mc1_stderr": 0.015345409485557961, |
|
"mc2": 0.45977172073584577, |
|
"mc2_stderr": 0.01635863808501988 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.23494687131050768, |
|
"acc_stderr": 0.014576237948550175, |
|
"acc_norm": 0.32585596221959856, |
|
"acc_norm_stderr": 0.016114023894800326 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "Korabbit/llama-2-ko-7b-pru", |
|
"model_sha": "cc3233e29e5358863df06a27bc23e1f07ed994e5", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |