|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3387372013651877, |
|
"acc_stderr": 0.013830568927974334, |
|
"acc_norm": 0.3924914675767918, |
|
"acc_norm_stderr": 0.01426963463567071 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3910575582553276, |
|
"acc_stderr": 0.004869899297734548, |
|
"acc_norm": 0.5143397729535949, |
|
"acc_norm_stderr": 0.004987728900897584 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.3684210526315789, |
|
"acc_stderr": 0.036996580176568775, |
|
"acc_norm": 0.3684210526315789, |
|
"acc_norm_stderr": 0.036996580176568775 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.3106796116504854, |
|
"acc_stderr": 0.04582124160161549, |
|
"acc_norm": 0.3106796116504854, |
|
"acc_norm_stderr": 0.04582124160161549 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.39208173690932313, |
|
"acc_stderr": 0.01745852405014764, |
|
"acc_norm": 0.39208173690932313, |
|
"acc_norm_stderr": 0.01745852405014764 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4074074074074074, |
|
"acc_stderr": 0.042446332383532286, |
|
"acc_norm": 0.4074074074074074, |
|
"acc_norm_stderr": 0.042446332383532286 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.2680851063829787, |
|
"acc_stderr": 0.028957342788342343, |
|
"acc_norm": 0.2680851063829787, |
|
"acc_norm_stderr": 0.028957342788342343 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.3132530120481928, |
|
"acc_stderr": 0.03610805018031024, |
|
"acc_norm": 0.3132530120481928, |
|
"acc_norm_stderr": 0.03610805018031024 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.43086816720257237, |
|
"acc_stderr": 0.028125340983972714, |
|
"acc_norm": 0.43086816720257237, |
|
"acc_norm_stderr": 0.028125340983972714 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.36771300448430494, |
|
"acc_stderr": 0.03236198350928275, |
|
"acc_norm": 0.36771300448430494, |
|
"acc_norm_stderr": 0.03236198350928275 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.3893129770992366, |
|
"acc_stderr": 0.04276486542814591, |
|
"acc_norm": 0.3893129770992366, |
|
"acc_norm_stderr": 0.04276486542814591 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.047258156262526045, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.35858585858585856, |
|
"acc_stderr": 0.0341690364039152, |
|
"acc_norm": 0.35858585858585856, |
|
"acc_norm_stderr": 0.0341690364039152 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.3793103448275862, |
|
"acc_stderr": 0.04043461861916747, |
|
"acc_norm": 0.3793103448275862, |
|
"acc_norm_stderr": 0.04043461861916747 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.10784313725490197, |
|
"acc_stderr": 0.03086428212206014, |
|
"acc_norm": 0.10784313725490197, |
|
"acc_norm_stderr": 0.03086428212206014 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.3235294117647059, |
|
"acc_stderr": 0.030388353551886845, |
|
"acc_norm": 0.3235294117647059, |
|
"acc_norm_stderr": 0.030388353551886845 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.26666666666666666, |
|
"acc_stderr": 0.0224212736129237, |
|
"acc_norm": 0.26666666666666666, |
|
"acc_norm_stderr": 0.0224212736129237 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.45, |
|
"acc_stderr": 0.05, |
|
"acc_norm": 0.45, |
|
"acc_norm_stderr": 0.05 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.22, |
|
"acc_stderr": 0.04163331998932269, |
|
"acc_norm": 0.22, |
|
"acc_norm_stderr": 0.04163331998932269 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.4074074074074074, |
|
"acc_stderr": 0.047500773411999854, |
|
"acc_norm": 0.4074074074074074, |
|
"acc_norm_stderr": 0.047500773411999854 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.28078817733990147, |
|
"acc_stderr": 0.03161856335358611, |
|
"acc_norm": 0.28078817733990147, |
|
"acc_norm_stderr": 0.03161856335358611 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.3161290322580645, |
|
"acc_stderr": 0.02645087448904277, |
|
"acc_norm": 0.3161290322580645, |
|
"acc_norm_stderr": 0.02645087448904277 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.4700854700854701, |
|
"acc_stderr": 0.03269741106812443, |
|
"acc_norm": 0.4700854700854701, |
|
"acc_norm_stderr": 0.03269741106812443 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.33962264150943394, |
|
"acc_stderr": 0.029146904747798345, |
|
"acc_norm": 0.33962264150943394, |
|
"acc_norm_stderr": 0.029146904747798345 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.3181818181818182, |
|
"acc_stderr": 0.04461272175910507, |
|
"acc_norm": 0.3181818181818182, |
|
"acc_norm_stderr": 0.04461272175910507 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.2814814814814815, |
|
"acc_stderr": 0.027420019350945277, |
|
"acc_norm": 0.2814814814814815, |
|
"acc_norm_stderr": 0.027420019350945277 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2781456953642384, |
|
"acc_stderr": 0.03658603262763743, |
|
"acc_norm": 0.2781456953642384, |
|
"acc_norm_stderr": 0.03658603262763743 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.39800995024875624, |
|
"acc_stderr": 0.034611994290400135, |
|
"acc_norm": 0.39800995024875624, |
|
"acc_norm_stderr": 0.034611994290400135 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.3063583815028902, |
|
"acc_stderr": 0.035149425512674394, |
|
"acc_norm": 0.3063583815028902, |
|
"acc_norm_stderr": 0.035149425512674394 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.2830687830687831, |
|
"acc_stderr": 0.023201392938194978, |
|
"acc_norm": 0.2830687830687831, |
|
"acc_norm_stderr": 0.023201392938194978 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.2638888888888889, |
|
"acc_stderr": 0.03685651095897532, |
|
"acc_norm": 0.2638888888888889, |
|
"acc_norm_stderr": 0.03685651095897532 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.43641618497109824, |
|
"acc_stderr": 0.026700545424943687, |
|
"acc_norm": 0.43641618497109824, |
|
"acc_norm_stderr": 0.026700545424943687 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.2883435582822086, |
|
"acc_stderr": 0.03559039531617342, |
|
"acc_norm": 0.2883435582822086, |
|
"acc_norm_stderr": 0.03559039531617342 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.3950617283950617, |
|
"acc_stderr": 0.027201117666925657, |
|
"acc_norm": 0.3950617283950617, |
|
"acc_norm_stderr": 0.027201117666925657 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.3160621761658031, |
|
"acc_stderr": 0.033553973696861736, |
|
"acc_norm": 0.3160621761658031, |
|
"acc_norm_stderr": 0.033553973696861736 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2719298245614035, |
|
"acc_stderr": 0.04185774424022057, |
|
"acc_norm": 0.2719298245614035, |
|
"acc_norm_stderr": 0.04185774424022057 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.3779816513761468, |
|
"acc_stderr": 0.020789187066728113, |
|
"acc_norm": 0.3779816513761468, |
|
"acc_norm_stderr": 0.020789187066728113 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.21428571428571427, |
|
"acc_stderr": 0.03670066451047182, |
|
"acc_norm": 0.21428571428571427, |
|
"acc_norm_stderr": 0.03670066451047182 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.39869281045751637, |
|
"acc_stderr": 0.02803609227389176, |
|
"acc_norm": 0.39869281045751637, |
|
"acc_norm_stderr": 0.02803609227389176 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252605, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252605 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.5537190082644629, |
|
"acc_stderr": 0.0453793517794788, |
|
"acc_norm": 0.5537190082644629, |
|
"acc_norm_stderr": 0.0453793517794788 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.3815789473684211, |
|
"acc_stderr": 0.03953173377749194, |
|
"acc_norm": 0.3815789473684211, |
|
"acc_norm_stderr": 0.03953173377749194 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.3104575163398693, |
|
"acc_stderr": 0.01871806705262322, |
|
"acc_norm": 0.3104575163398693, |
|
"acc_norm_stderr": 0.01871806705262322 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3191489361702128, |
|
"acc_stderr": 0.027807990141320203, |
|
"acc_norm": 0.3191489361702128, |
|
"acc_norm_stderr": 0.027807990141320203 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.29464285714285715, |
|
"acc_stderr": 0.0432704093257873, |
|
"acc_norm": 0.29464285714285715, |
|
"acc_norm_stderr": 0.0432704093257873 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.3425925925925926, |
|
"acc_stderr": 0.032365852526021574, |
|
"acc_norm": 0.3425925925925926, |
|
"acc_norm_stderr": 0.032365852526021574 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.24134078212290502, |
|
"acc_stderr": 0.014310999547961441, |
|
"acc_norm": 0.24134078212290502, |
|
"acc_norm_stderr": 0.014310999547961441 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.39, |
|
"acc_stderr": 0.04902071300001975, |
|
"acc_norm": 0.39, |
|
"acc_norm_stderr": 0.04902071300001975 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.3235294117647059, |
|
"acc_stderr": 0.028418208619406794, |
|
"acc_norm": 0.3235294117647059, |
|
"acc_norm_stderr": 0.028418208619406794 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.31020408163265306, |
|
"acc_stderr": 0.029613459872484378, |
|
"acc_norm": 0.31020408163265306, |
|
"acc_norm_stderr": 0.029613459872484378 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.47257383966244726, |
|
"acc_stderr": 0.032498227183013026, |
|
"acc_norm": 0.47257383966244726, |
|
"acc_norm_stderr": 0.032498227183013026 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.2900912646675359, |
|
"acc_stderr": 0.0115903755547331, |
|
"acc_norm": 0.2900912646675359, |
|
"acc_norm_stderr": 0.0115903755547331 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.29411764705882354, |
|
"acc_stderr": 0.03198001660115072, |
|
"acc_norm": 0.29411764705882354, |
|
"acc_norm_stderr": 0.03198001660115072 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.3696969696969697, |
|
"acc_stderr": 0.037694303145125674, |
|
"acc_norm": 0.3696969696969697, |
|
"acc_norm_stderr": 0.037694303145125674 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.26560587515299877, |
|
"mc1_stderr": 0.015461027627253597, |
|
"mc2": 0.40727214174838056, |
|
"mc2_stderr": 0.014940202090745085 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.3204225352112676, |
|
"acc_stderr": 0.015996178088626918, |
|
"acc_norm": 0.4307511737089202, |
|
"acc_norm_stderr": 0.016974599121731444 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "FINDA-FIT/xllama-instruct", |
|
"model_sha": "1e5ee340d5f4558a1bc451ba7942fa5f3a1c8d80", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |