|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3387372013651877, |
|
"acc_stderr": 0.013830568927974332, |
|
"acc_norm": 0.40017064846416384, |
|
"acc_norm_stderr": 0.014317197787809174 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.4023102967536347, |
|
"acc_stderr": 0.004893617014975314, |
|
"acc_norm": 0.5194184425413264, |
|
"acc_norm_stderr": 0.004986016938678532 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.43859649122807015, |
|
"acc_stderr": 0.038057975055904594, |
|
"acc_norm": 0.43859649122807015, |
|
"acc_norm_stderr": 0.038057975055904594 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.34951456310679613, |
|
"acc_stderr": 0.047211885060971716, |
|
"acc_norm": 0.34951456310679613, |
|
"acc_norm_stderr": 0.047211885060971716 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.421455938697318, |
|
"acc_stderr": 0.017657976412654857, |
|
"acc_norm": 0.421455938697318, |
|
"acc_norm_stderr": 0.017657976412654857 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.3851851851851852, |
|
"acc_stderr": 0.042039210401562783, |
|
"acc_norm": 0.3851851851851852, |
|
"acc_norm_stderr": 0.042039210401562783 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.3276595744680851, |
|
"acc_stderr": 0.030683020843231004, |
|
"acc_norm": 0.3276595744680851, |
|
"acc_norm_stderr": 0.030683020843231004 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.3373493975903614, |
|
"acc_stderr": 0.03680783690727581, |
|
"acc_norm": 0.3373493975903614, |
|
"acc_norm_stderr": 0.03680783690727581 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.3987138263665595, |
|
"acc_stderr": 0.0278093225857745, |
|
"acc_norm": 0.3987138263665595, |
|
"acc_norm_stderr": 0.0278093225857745 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.37668161434977576, |
|
"acc_stderr": 0.032521134899291884, |
|
"acc_norm": 0.37668161434977576, |
|
"acc_norm_stderr": 0.032521134899291884 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.4122137404580153, |
|
"acc_stderr": 0.043171711948702535, |
|
"acc_norm": 0.4122137404580153, |
|
"acc_norm_stderr": 0.043171711948702535 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.044619604333847394, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.044619604333847394 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.3787878787878788, |
|
"acc_stderr": 0.03456088731993747, |
|
"acc_norm": 0.3787878787878788, |
|
"acc_norm_stderr": 0.03456088731993747 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.3586206896551724, |
|
"acc_stderr": 0.039966295748767186, |
|
"acc_norm": 0.3586206896551724, |
|
"acc_norm_stderr": 0.039966295748767186 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.18627450980392157, |
|
"acc_stderr": 0.038739587141493524, |
|
"acc_norm": 0.18627450980392157, |
|
"acc_norm_stderr": 0.038739587141493524 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.31512605042016806, |
|
"acc_stderr": 0.03017680828897434, |
|
"acc_norm": 0.31512605042016806, |
|
"acc_norm_stderr": 0.03017680828897434 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.30256410256410254, |
|
"acc_stderr": 0.023290888053772725, |
|
"acc_norm": 0.30256410256410254, |
|
"acc_norm_stderr": 0.023290888053772725 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.39, |
|
"acc_stderr": 0.04902071300001974, |
|
"acc_norm": 0.39, |
|
"acc_norm_stderr": 0.04902071300001974 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.3888888888888889, |
|
"acc_stderr": 0.0471282125742677, |
|
"acc_norm": 0.3888888888888889, |
|
"acc_norm_stderr": 0.0471282125742677 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.30049261083743845, |
|
"acc_stderr": 0.03225799476233484, |
|
"acc_norm": 0.30049261083743845, |
|
"acc_norm_stderr": 0.03225799476233484 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.34516129032258064, |
|
"acc_stderr": 0.027045746573534327, |
|
"acc_norm": 0.34516129032258064, |
|
"acc_norm_stderr": 0.027045746573534327 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.5427350427350427, |
|
"acc_stderr": 0.03263622596380688, |
|
"acc_norm": 0.5427350427350427, |
|
"acc_norm_stderr": 0.03263622596380688 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.35471698113207545, |
|
"acc_stderr": 0.02944517532819959, |
|
"acc_norm": 0.35471698113207545, |
|
"acc_norm_stderr": 0.02944517532819959 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.34545454545454546, |
|
"acc_stderr": 0.04554619617541054, |
|
"acc_norm": 0.34545454545454546, |
|
"acc_norm_stderr": 0.04554619617541054 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.24444444444444444, |
|
"acc_stderr": 0.02620276653465215, |
|
"acc_norm": 0.24444444444444444, |
|
"acc_norm_stderr": 0.02620276653465215 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.26490066225165565, |
|
"acc_stderr": 0.03603038545360383, |
|
"acc_norm": 0.26490066225165565, |
|
"acc_norm_stderr": 0.03603038545360383 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.3383084577114428, |
|
"acc_stderr": 0.03345563070339193, |
|
"acc_norm": 0.3383084577114428, |
|
"acc_norm_stderr": 0.03345563070339193 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.2774566473988439, |
|
"acc_stderr": 0.034140140070440354, |
|
"acc_norm": 0.2774566473988439, |
|
"acc_norm_stderr": 0.034140140070440354 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.28835978835978837, |
|
"acc_stderr": 0.0233306540545359, |
|
"acc_norm": 0.28835978835978837, |
|
"acc_norm_stderr": 0.0233306540545359 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.2569444444444444, |
|
"acc_stderr": 0.03653946969442099, |
|
"acc_norm": 0.2569444444444444, |
|
"acc_norm_stderr": 0.03653946969442099 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.0440844002276808, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.0440844002276808 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.47, |
|
"acc_stderr": 0.05016135580465919, |
|
"acc_norm": 0.47, |
|
"acc_norm_stderr": 0.05016135580465919 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.3439306358381503, |
|
"acc_stderr": 0.025574123786546672, |
|
"acc_norm": 0.3439306358381503, |
|
"acc_norm_stderr": 0.025574123786546672 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.3374233128834356, |
|
"acc_stderr": 0.03714908409935575, |
|
"acc_norm": 0.3374233128834356, |
|
"acc_norm_stderr": 0.03714908409935575 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.3395061728395062, |
|
"acc_stderr": 0.026348564412011628, |
|
"acc_norm": 0.3395061728395062, |
|
"acc_norm_stderr": 0.026348564412011628 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.37305699481865284, |
|
"acc_stderr": 0.03490205592048574, |
|
"acc_norm": 0.37305699481865284, |
|
"acc_norm_stderr": 0.03490205592048574 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2631578947368421, |
|
"acc_stderr": 0.04142439719489362, |
|
"acc_norm": 0.2631578947368421, |
|
"acc_norm_stderr": 0.04142439719489362 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.3706422018348624, |
|
"acc_stderr": 0.02070745816435298, |
|
"acc_norm": 0.3706422018348624, |
|
"acc_norm_stderr": 0.02070745816435298 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.2222222222222222, |
|
"acc_stderr": 0.037184890068181146, |
|
"acc_norm": 0.2222222222222222, |
|
"acc_norm_stderr": 0.037184890068181146 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.3464052287581699, |
|
"acc_stderr": 0.027245613047215355, |
|
"acc_norm": 0.3464052287581699, |
|
"acc_norm_stderr": 0.027245613047215355 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252606, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252606 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.5041322314049587, |
|
"acc_stderr": 0.04564198767432754, |
|
"acc_norm": 0.5041322314049587, |
|
"acc_norm_stderr": 0.04564198767432754 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.28289473684210525, |
|
"acc_stderr": 0.03665349695640767, |
|
"acc_norm": 0.28289473684210525, |
|
"acc_norm_stderr": 0.03665349695640767 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.32679738562091504, |
|
"acc_stderr": 0.018975427920507222, |
|
"acc_norm": 0.32679738562091504, |
|
"acc_norm_stderr": 0.018975427920507222 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.33687943262411346, |
|
"acc_stderr": 0.028195534873966727, |
|
"acc_norm": 0.33687943262411346, |
|
"acc_norm_stderr": 0.028195534873966727 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.3125, |
|
"acc_stderr": 0.043994650575715215, |
|
"acc_norm": 0.3125, |
|
"acc_norm_stderr": 0.043994650575715215 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.26851851851851855, |
|
"acc_stderr": 0.030225226160012393, |
|
"acc_norm": 0.26851851851851855, |
|
"acc_norm_stderr": 0.030225226160012393 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.24134078212290502, |
|
"acc_stderr": 0.014310999547961443, |
|
"acc_norm": 0.24134078212290502, |
|
"acc_norm_stderr": 0.014310999547961443 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252604, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252604 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.2647058823529412, |
|
"acc_stderr": 0.026799562024887678, |
|
"acc_norm": 0.2647058823529412, |
|
"acc_norm_stderr": 0.026799562024887678 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.2897959183673469, |
|
"acc_stderr": 0.02904308868330433, |
|
"acc_norm": 0.2897959183673469, |
|
"acc_norm_stderr": 0.02904308868330433 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.45147679324894513, |
|
"acc_stderr": 0.032393600173974704, |
|
"acc_norm": 0.45147679324894513, |
|
"acc_norm_stderr": 0.032393600173974704 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.2940026075619296, |
|
"acc_stderr": 0.011636062953698602, |
|
"acc_norm": 0.2940026075619296, |
|
"acc_norm_stderr": 0.011636062953698602 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.3382352941176471, |
|
"acc_stderr": 0.03320574612945431, |
|
"acc_norm": 0.3382352941176471, |
|
"acc_norm_stderr": 0.03320574612945431 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.37575757575757573, |
|
"acc_stderr": 0.03781887353205982, |
|
"acc_norm": 0.37575757575757573, |
|
"acc_norm_stderr": 0.03781887353205982 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.25091799265605874, |
|
"mc1_stderr": 0.01517698502770769, |
|
"mc2": 0.41272169126715796, |
|
"mc2_stderr": 0.015689006867142138 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.4332939787485242, |
|
"acc_stderr": 0.0170366836418931, |
|
"acc_norm": 0.5112160566706021, |
|
"acc_norm_stderr": 0.017186028469489294 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "Korabbit/llama-2-ko-7b-bilingual", |
|
"model_sha": "3182f7a997a431a53f9157384c6fb742619f8fbc", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |