{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.26706484641638223, "acc_stderr": 0.012928933196496337, "acc_norm": 0.3310580204778157, "acc_norm_stderr": 0.013752062419817836 }, "harness|ko_hellaswag|10": { "acc": 0.37134037044413465, "acc_stderr": 0.004821757734156723, "acc_norm": 0.47470623381796456, "acc_norm_stderr": 0.004983392650570962 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.27485380116959063, "acc_stderr": 0.03424042924691584, "acc_norm": 0.27485380116959063, "acc_norm_stderr": 0.03424042924691584 }, "harness|ko_mmlu_management|5": { "acc": 0.22330097087378642, "acc_stderr": 0.04123553189891431, "acc_norm": 0.22330097087378642, "acc_norm_stderr": 0.04123553189891431 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.2413793103448276, "acc_stderr": 0.0153023801235421, "acc_norm": 0.2413793103448276, "acc_norm_stderr": 0.0153023801235421 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.25925925925925924, "acc_stderr": 0.03785714465066654, "acc_norm": 0.25925925925925924, "acc_norm_stderr": 0.03785714465066654 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.23, "acc_stderr": 0.042295258468165044, "acc_norm": 0.23, "acc_norm_stderr": 0.042295258468165044 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.2, "acc_stderr": 0.026148818018424513, "acc_norm": 0.2, "acc_norm_stderr": 0.026148818018424513 }, "harness|ko_mmlu_virology|5": { "acc": 0.24096385542168675, "acc_stderr": 0.03329394119073529, "acc_norm": 0.24096385542168675, "acc_norm_stderr": 0.03329394119073529 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.21864951768488747, "acc_stderr": 0.02347558141786111, "acc_norm": 0.21864951768488747, "acc_norm_stderr": 0.02347558141786111 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.3004484304932735, "acc_stderr": 0.030769352008229136, "acc_norm": 0.3004484304932735, "acc_norm_stderr": 0.030769352008229136 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.24427480916030533, "acc_stderr": 0.037683359597287434, "acc_norm": 0.24427480916030533, "acc_norm_stderr": 0.037683359597287434 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.29, "acc_stderr": 0.045604802157206845, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206845 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.29292929292929293, "acc_stderr": 0.03242497958178817, "acc_norm": 0.29292929292929293, "acc_norm_stderr": 0.03242497958178817 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.25517241379310346, "acc_stderr": 0.03632984052707842, "acc_norm": 0.25517241379310346, "acc_norm_stderr": 0.03632984052707842 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.27450980392156865, "acc_stderr": 0.044405219061793275, "acc_norm": 0.27450980392156865, "acc_norm_stderr": 0.044405219061793275 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.25630252100840334, "acc_stderr": 0.028359620870533953, "acc_norm": 0.25630252100840334, "acc_norm_stderr": 0.028359620870533953 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.2128205128205128, "acc_stderr": 0.020752423722128002, "acc_norm": 0.2128205128205128, "acc_norm_stderr": 0.020752423722128002 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.24, "acc_stderr": 0.042923469599092816, "acc_norm": 0.24, "acc_norm_stderr": 0.042923469599092816 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.19, "acc_stderr": 0.039427724440366234, "acc_norm": 0.19, "acc_norm_stderr": 0.039427724440366234 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.2222222222222222, "acc_stderr": 0.040191074725573483, "acc_norm": 0.2222222222222222, "acc_norm_stderr": 0.040191074725573483 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.30049261083743845, "acc_stderr": 0.03225799476233483, "acc_norm": 0.30049261083743845, "acc_norm_stderr": 0.03225799476233483 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.24193548387096775, "acc_stderr": 0.024362599693031093, "acc_norm": 0.24193548387096775, "acc_norm_stderr": 0.024362599693031093 }, "harness|ko_mmlu_marketing|5": { "acc": 0.2777777777777778, "acc_stderr": 0.02934311479809445, "acc_norm": 0.2777777777777778, "acc_norm_stderr": 0.02934311479809445 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.2490566037735849, "acc_stderr": 0.026616482980501704, "acc_norm": 0.2490566037735849, "acc_norm_stderr": 0.026616482980501704 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.2636363636363636, "acc_stderr": 0.04220224692971987, "acc_norm": 0.2636363636363636, "acc_norm_stderr": 0.04220224692971987 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.23703703703703705, "acc_stderr": 0.025928876132766114, "acc_norm": 0.23703703703703705, "acc_norm_stderr": 0.025928876132766114 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.25165562913907286, "acc_stderr": 0.03543304234389985, "acc_norm": 0.25165562913907286, "acc_norm_stderr": 0.03543304234389985 }, "harness|ko_mmlu_sociology|5": { "acc": 0.21393034825870647, "acc_stderr": 0.028996909693328927, "acc_norm": 0.21393034825870647, "acc_norm_stderr": 0.028996909693328927 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.2023121387283237, "acc_stderr": 0.03063114553919882, "acc_norm": 0.2023121387283237, "acc_norm_stderr": 0.03063114553919882 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.23015873015873015, "acc_stderr": 0.021679219663693138, "acc_norm": 0.23015873015873015, "acc_norm_stderr": 0.021679219663693138 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.25, "acc_stderr": 0.03621034121889507, "acc_norm": 0.25, "acc_norm_stderr": 0.03621034121889507 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.24, "acc_stderr": 0.04292346959909284, "acc_norm": 0.24, "acc_norm_stderr": 0.04292346959909284 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.28, "acc_stderr": 0.045126085985421296, "acc_norm": 0.28, "acc_norm_stderr": 0.045126085985421296 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.26011560693641617, "acc_stderr": 0.023618678310069363, "acc_norm": 0.26011560693641617, "acc_norm_stderr": 0.023618678310069363 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.24539877300613497, "acc_stderr": 0.03380939813943354, "acc_norm": 0.24539877300613497, "acc_norm_stderr": 0.03380939813943354 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.26851851851851855, "acc_stderr": 0.02465968518596729, "acc_norm": 0.26851851851851855, "acc_norm_stderr": 0.02465968518596729 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.24, "acc_stderr": 0.04292346959909282, "acc_norm": 0.24, "acc_norm_stderr": 0.04292346959909282 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.21243523316062177, "acc_stderr": 0.029519282616817244, "acc_norm": 0.21243523316062177, "acc_norm_stderr": 0.029519282616817244 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2543859649122807, "acc_stderr": 0.040969851398436716, "acc_norm": 0.2543859649122807, "acc_norm_stderr": 0.040969851398436716 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.25688073394495414, "acc_stderr": 0.01873249292834247, "acc_norm": 0.25688073394495414, "acc_norm_stderr": 0.01873249292834247 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.16666666666666666, "acc_stderr": 0.03333333333333338, "acc_norm": 0.16666666666666666, "acc_norm_stderr": 0.03333333333333338 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.2549019607843137, "acc_stderr": 0.02495418432487991, "acc_norm": 0.2549019607843137, "acc_norm_stderr": 0.02495418432487991 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_international_law|5": { "acc": 0.2396694214876033, "acc_stderr": 0.038968789850704164, "acc_norm": 0.2396694214876033, "acc_norm_stderr": 0.038968789850704164 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.18421052631578946, "acc_stderr": 0.0315469804508223, "acc_norm": 0.18421052631578946, "acc_norm_stderr": 0.0315469804508223 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.24836601307189543, "acc_stderr": 0.017479487001364764, "acc_norm": 0.24836601307189543, "acc_norm_stderr": 0.017479487001364764 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.24113475177304963, "acc_stderr": 0.02551873104953777, "acc_norm": 0.24113475177304963, "acc_norm_stderr": 0.02551873104953777 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.25892857142857145, "acc_stderr": 0.04157751539865629, "acc_norm": 0.25892857142857145, "acc_norm_stderr": 0.04157751539865629 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.19907407407407407, "acc_stderr": 0.027232298462690218, "acc_norm": 0.19907407407407407, "acc_norm_stderr": 0.027232298462690218 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2547486033519553, "acc_stderr": 0.014572650383409162, "acc_norm": 0.2547486033519553, "acc_norm_stderr": 0.014572650383409162 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.27, "acc_stderr": 0.044619604333847394, "acc_norm": 0.27, "acc_norm_stderr": 0.044619604333847394 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.28, "acc_stderr": 0.04512608598542128, "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542128 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.34191176470588236, "acc_stderr": 0.028814722422254184, "acc_norm": 0.34191176470588236, "acc_norm_stderr": 0.028814722422254184 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.21224489795918366, "acc_stderr": 0.026176967197866767, "acc_norm": 0.21224489795918366, "acc_norm_stderr": 0.026176967197866767 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.2616033755274262, "acc_stderr": 0.028609516716994934, "acc_norm": 0.2616033755274262, "acc_norm_stderr": 0.028609516716994934 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.25097783572359844, "acc_stderr": 0.011073730299187234, "acc_norm": 0.25097783572359844, "acc_norm_stderr": 0.011073730299187234 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.2696078431372549, "acc_stderr": 0.031145570659486782, "acc_norm": 0.2696078431372549, "acc_norm_stderr": 0.031145570659486782 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.2909090909090909, "acc_stderr": 0.03546563019624335, "acc_norm": 0.2909090909090909, "acc_norm_stderr": 0.03546563019624335 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.27539779681762544, "mc1_stderr": 0.015638135667775523, "mc2": 0.4336773026110262, "mc2_stderr": 0.01517918566270363 }, "harness|ko_commongen_v2|2": { "acc": 0.45774647887323944, "acc_stderr": 0.01707846824220106, "acc_norm": 0.5469483568075117, "acc_norm_stderr": 0.01706405474980085 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "iknow-lab/AULM-12.8b-v0", "model_sha": "daeca40346ba44b1fbb6939cc635adf467fa6cab", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }