|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.4539249146757679, |
|
"acc_stderr": 0.014549221105171865, |
|
"acc_norm": 0.5085324232081911, |
|
"acc_norm_stderr": 0.014609263165632182 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.4373630750846445, |
|
"acc_stderr": 0.004950472918523313, |
|
"acc_norm": 0.5981876120294762, |
|
"acc_norm_stderr": 0.004892624490937216 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.6549707602339181, |
|
"acc_stderr": 0.036459813773888065, |
|
"acc_norm": 0.6549707602339181, |
|
"acc_norm_stderr": 0.036459813773888065 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6601941747572816, |
|
"acc_stderr": 0.046897659372781335, |
|
"acc_norm": 0.6601941747572816, |
|
"acc_norm_stderr": 0.046897659372781335 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.6819923371647509, |
|
"acc_stderr": 0.01665348627561539, |
|
"acc_norm": 0.6819923371647509, |
|
"acc_norm_stderr": 0.01665348627561539 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.5185185185185185, |
|
"acc_stderr": 0.043163785995113245, |
|
"acc_norm": 0.5185185185185185, |
|
"acc_norm_stderr": 0.043163785995113245 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.047937248544110196, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.047937248544110196 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.46808510638297873, |
|
"acc_stderr": 0.03261936918467381, |
|
"acc_norm": 0.46808510638297873, |
|
"acc_norm_stderr": 0.03261936918467381 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.42771084337349397, |
|
"acc_stderr": 0.038515976837185335, |
|
"acc_norm": 0.42771084337349397, |
|
"acc_norm_stderr": 0.038515976837185335 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.5691318327974276, |
|
"acc_stderr": 0.028125340983972714, |
|
"acc_norm": 0.5691318327974276, |
|
"acc_norm_stderr": 0.028125340983972714 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5381165919282511, |
|
"acc_stderr": 0.03346015011973228, |
|
"acc_norm": 0.5381165919282511, |
|
"acc_norm_stderr": 0.03346015011973228 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.5954198473282443, |
|
"acc_stderr": 0.043046937953806645, |
|
"acc_norm": 0.5954198473282443, |
|
"acc_norm_stderr": 0.043046937953806645 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.53, |
|
"acc_stderr": 0.050161355804659205, |
|
"acc_norm": 0.53, |
|
"acc_norm_stderr": 0.050161355804659205 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.6767676767676768, |
|
"acc_stderr": 0.03332299921070645, |
|
"acc_norm": 0.6767676767676768, |
|
"acc_norm_stderr": 0.03332299921070645 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.4689655172413793, |
|
"acc_stderr": 0.04158632762097828, |
|
"acc_norm": 0.4689655172413793, |
|
"acc_norm_stderr": 0.04158632762097828 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.23529411764705882, |
|
"acc_stderr": 0.04220773659171453, |
|
"acc_norm": 0.23529411764705882, |
|
"acc_norm_stderr": 0.04220773659171453 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.48739495798319327, |
|
"acc_stderr": 0.03246816765752173, |
|
"acc_norm": 0.48739495798319327, |
|
"acc_norm_stderr": 0.03246816765752173 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.5102564102564102, |
|
"acc_stderr": 0.025345672221942374, |
|
"acc_norm": 0.5102564102564102, |
|
"acc_norm_stderr": 0.025345672221942374 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.62, |
|
"acc_stderr": 0.04878317312145631, |
|
"acc_norm": 0.62, |
|
"acc_norm_stderr": 0.04878317312145631 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621504, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621504 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6111111111111112, |
|
"acc_stderr": 0.04712821257426769, |
|
"acc_norm": 0.6111111111111112, |
|
"acc_norm_stderr": 0.04712821257426769 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.3793103448275862, |
|
"acc_stderr": 0.034139638059062345, |
|
"acc_norm": 0.3793103448275862, |
|
"acc_norm_stderr": 0.034139638059062345 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.5741935483870968, |
|
"acc_stderr": 0.02812911270916591, |
|
"acc_norm": 0.5741935483870968, |
|
"acc_norm_stderr": 0.02812911270916591 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7905982905982906, |
|
"acc_stderr": 0.026655699653922754, |
|
"acc_norm": 0.7905982905982906, |
|
"acc_norm_stderr": 0.026655699653922754 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.5207547169811321, |
|
"acc_stderr": 0.03074634997572347, |
|
"acc_norm": 0.5207547169811321, |
|
"acc_norm_stderr": 0.03074634997572347 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5727272727272728, |
|
"acc_stderr": 0.04738198703545483, |
|
"acc_norm": 0.5727272727272728, |
|
"acc_norm_stderr": 0.04738198703545483 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.362962962962963, |
|
"acc_stderr": 0.029318203645206865, |
|
"acc_norm": 0.362962962962963, |
|
"acc_norm_stderr": 0.029318203645206865 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.31125827814569534, |
|
"acc_stderr": 0.03780445850526732, |
|
"acc_norm": 0.31125827814569534, |
|
"acc_norm_stderr": 0.03780445850526732 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.7114427860696517, |
|
"acc_stderr": 0.03203841040213321, |
|
"acc_norm": 0.7114427860696517, |
|
"acc_norm_stderr": 0.03203841040213321 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.4393063583815029, |
|
"acc_stderr": 0.03784271932887467, |
|
"acc_norm": 0.4393063583815029, |
|
"acc_norm_stderr": 0.03784271932887467 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.3306878306878307, |
|
"acc_stderr": 0.02422996529842507, |
|
"acc_norm": 0.3306878306878307, |
|
"acc_norm_stderr": 0.02422996529842507 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4861111111111111, |
|
"acc_stderr": 0.04179596617581, |
|
"acc_norm": 0.4861111111111111, |
|
"acc_norm_stderr": 0.04179596617581 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.76, |
|
"acc_stderr": 0.04292346959909283, |
|
"acc_norm": 0.76, |
|
"acc_norm_stderr": 0.04292346959909283 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.523121387283237, |
|
"acc_stderr": 0.026890297881303125, |
|
"acc_norm": 0.523121387283237, |
|
"acc_norm_stderr": 0.026890297881303125 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.5153374233128835, |
|
"acc_stderr": 0.039265223787088424, |
|
"acc_norm": 0.5153374233128835, |
|
"acc_norm_stderr": 0.039265223787088424 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.5679012345679012, |
|
"acc_stderr": 0.027563010971606676, |
|
"acc_norm": 0.5679012345679012, |
|
"acc_norm_stderr": 0.027563010971606676 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.047937248544110196, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.047937248544110196 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.689119170984456, |
|
"acc_stderr": 0.03340361906276586, |
|
"acc_norm": 0.689119170984456, |
|
"acc_norm_stderr": 0.03340361906276586 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.35964912280701755, |
|
"acc_stderr": 0.04514496132873633, |
|
"acc_norm": 0.35964912280701755, |
|
"acc_norm_stderr": 0.04514496132873633 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.6293577981651376, |
|
"acc_stderr": 0.020707458164352984, |
|
"acc_norm": 0.6293577981651376, |
|
"acc_norm_stderr": 0.020707458164352984 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.30952380952380953, |
|
"acc_stderr": 0.04134913018303316, |
|
"acc_norm": 0.30952380952380953, |
|
"acc_norm_stderr": 0.04134913018303316 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5196078431372549, |
|
"acc_stderr": 0.028607893699576063, |
|
"acc_norm": 0.5196078431372549, |
|
"acc_norm_stderr": 0.028607893699576063 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.57, |
|
"acc_stderr": 0.049756985195624284, |
|
"acc_norm": 0.57, |
|
"acc_norm_stderr": 0.049756985195624284 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.71900826446281, |
|
"acc_stderr": 0.04103203830514511, |
|
"acc_norm": 0.71900826446281, |
|
"acc_norm_stderr": 0.04103203830514511 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5263157894736842, |
|
"acc_stderr": 0.04063302731486671, |
|
"acc_norm": 0.5263157894736842, |
|
"acc_norm_stderr": 0.04063302731486671 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.45751633986928103, |
|
"acc_stderr": 0.0201546857125909, |
|
"acc_norm": 0.45751633986928103, |
|
"acc_norm_stderr": 0.0201546857125909 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.34397163120567376, |
|
"acc_stderr": 0.02833801742861132, |
|
"acc_norm": 0.34397163120567376, |
|
"acc_norm_stderr": 0.02833801742861132 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.25892857142857145, |
|
"acc_stderr": 0.041577515398656284, |
|
"acc_norm": 0.25892857142857145, |
|
"acc_norm_stderr": 0.041577515398656284 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.38425925925925924, |
|
"acc_stderr": 0.03317354514310742, |
|
"acc_norm": 0.38425925925925924, |
|
"acc_norm_stderr": 0.03317354514310742 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.28044692737430166, |
|
"acc_stderr": 0.015024083883322872, |
|
"acc_norm": 0.28044692737430166, |
|
"acc_norm_stderr": 0.015024083883322872 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.048783173121456316, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.048783173121456316 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.63, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.63, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.41544117647058826, |
|
"acc_stderr": 0.029935342707877753, |
|
"acc_norm": 0.41544117647058826, |
|
"acc_norm_stderr": 0.029935342707877753 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.6122448979591837, |
|
"acc_stderr": 0.031192230726795656, |
|
"acc_norm": 0.6122448979591837, |
|
"acc_norm_stderr": 0.031192230726795656 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.6962025316455697, |
|
"acc_stderr": 0.02993669638713861, |
|
"acc_norm": 0.6962025316455697, |
|
"acc_norm_stderr": 0.02993669638713861 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.34159061277705344, |
|
"acc_stderr": 0.012112391320842854, |
|
"acc_norm": 0.34159061277705344, |
|
"acc_norm_stderr": 0.012112391320842854 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.6274509803921569, |
|
"acc_stderr": 0.03393388584958404, |
|
"acc_norm": 0.6274509803921569, |
|
"acc_norm_stderr": 0.03393388584958404 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.6181818181818182, |
|
"acc_stderr": 0.03793713171165634, |
|
"acc_norm": 0.6181818181818182, |
|
"acc_norm_stderr": 0.03793713171165634 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2937576499388005, |
|
"mc1_stderr": 0.015945068581236614, |
|
"mc2": 0.46102353979455557, |
|
"mc2_stderr": 0.01511169348606206 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5974025974025974, |
|
"acc_stderr": 0.016861020486407773, |
|
"acc_norm": 0.6375442739079102, |
|
"acc_norm_stderr": 0.016527131240453692 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "GAI-LLM/OPEN-SOLAR-KO-10.7B-mixed-v15", |
|
"model_sha": "da2d5692181ad63e0cd1ae2eb4bcd3860e183a45", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |