{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.3455631399317406, "acc_stderr": 0.013896938461145687, "acc_norm": 0.38139931740614336, "acc_norm_stderr": 0.014194389086685265 }, "harness|ko_hellaswag|10": { "acc": 0.3619796853216491, "acc_stderr": 0.004795908282584555, "acc_norm": 0.44761999601672975, "acc_norm_stderr": 0.0049623252978409915 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.47368421052631576, "acc_stderr": 0.03829509868994727, "acc_norm": 0.47368421052631576, "acc_norm_stderr": 0.03829509868994727 }, "harness|ko_mmlu_management|5": { "acc": 0.5436893203883495, "acc_stderr": 0.049318019942204146, "acc_norm": 0.5436893203883495, "acc_norm_stderr": 0.049318019942204146 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.4725415070242657, "acc_stderr": 0.01785298126663395, "acc_norm": 0.4725415070242657, "acc_norm_stderr": 0.01785298126663395 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.37777777777777777, "acc_stderr": 0.04188307537595853, "acc_norm": 0.37777777777777777, "acc_norm_stderr": 0.04188307537595853 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.29, "acc_stderr": 0.04560480215720684, "acc_norm": 0.29, "acc_norm_stderr": 0.04560480215720684 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.39148936170212767, "acc_stderr": 0.03190701242326812, "acc_norm": 0.39148936170212767, "acc_norm_stderr": 0.03190701242326812 }, "harness|ko_mmlu_virology|5": { "acc": 0.4397590361445783, "acc_stderr": 0.03864139923699122, "acc_norm": 0.4397590361445783, "acc_norm_stderr": 0.03864139923699122 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.4855305466237942, "acc_stderr": 0.028386198084177673, "acc_norm": 0.4855305466237942, "acc_norm_stderr": 0.028386198084177673 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.4170403587443946, "acc_stderr": 0.03309266936071722, "acc_norm": 0.4170403587443946, "acc_norm_stderr": 0.03309266936071722 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.4580152671755725, "acc_stderr": 0.04369802690578756, "acc_norm": 0.4580152671755725, "acc_norm_stderr": 0.04369802690578756 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5151515151515151, "acc_stderr": 0.03560716516531061, "acc_norm": 0.5151515151515151, "acc_norm_stderr": 0.03560716516531061 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.4068965517241379, "acc_stderr": 0.040937939812662374, "acc_norm": 0.4068965517241379, "acc_norm_stderr": 0.040937939812662374 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.23529411764705882, "acc_stderr": 0.04220773659171452, "acc_norm": 0.23529411764705882, "acc_norm_stderr": 0.04220773659171452 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.4579831932773109, "acc_stderr": 0.03236361111951941, "acc_norm": 0.4579831932773109, "acc_norm_stderr": 0.03236361111951941 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.37435897435897436, "acc_stderr": 0.024537591572830524, "acc_norm": 0.37435897435897436, "acc_norm_stderr": 0.024537591572830524 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.52, "acc_stderr": 0.05021167315686779, "acc_norm": 0.52, "acc_norm_stderr": 0.05021167315686779 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.32, "acc_stderr": 0.04688261722621504, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621504 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.4444444444444444, "acc_stderr": 0.04803752235190193, "acc_norm": 0.4444444444444444, "acc_norm_stderr": 0.04803752235190193 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.43842364532019706, "acc_stderr": 0.03491207857486518, "acc_norm": 0.43842364532019706, "acc_norm_stderr": 0.03491207857486518 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.45161290322580644, "acc_stderr": 0.028310500348568385, "acc_norm": 0.45161290322580644, "acc_norm_stderr": 0.028310500348568385 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6752136752136753, "acc_stderr": 0.03067902276549883, "acc_norm": 0.6752136752136753, "acc_norm_stderr": 0.03067902276549883 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.43018867924528303, "acc_stderr": 0.030471445867183235, "acc_norm": 0.43018867924528303, "acc_norm_stderr": 0.030471445867183235 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5, "acc_stderr": 0.04789131426105757, "acc_norm": 0.5, "acc_norm_stderr": 0.04789131426105757 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.32592592592592595, "acc_stderr": 0.02857834836547308, "acc_norm": 0.32592592592592595, "acc_norm_stderr": 0.02857834836547308 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.23178807947019867, "acc_stderr": 0.03445406271987053, "acc_norm": 0.23178807947019867, "acc_norm_stderr": 0.03445406271987053 }, "harness|ko_mmlu_sociology|5": { "acc": 0.5522388059701493, "acc_stderr": 0.03516184772952167, "acc_norm": 0.5522388059701493, "acc_norm_stderr": 0.03516184772952167 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.3468208092485549, "acc_stderr": 0.03629146670159665, "acc_norm": 0.3468208092485549, "acc_norm_stderr": 0.03629146670159665 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.37566137566137564, "acc_stderr": 0.02494236893115979, "acc_norm": 0.37566137566137564, "acc_norm_stderr": 0.02494236893115979 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3194444444444444, "acc_stderr": 0.038990736873573344, "acc_norm": 0.3194444444444444, "acc_norm_stderr": 0.038990736873573344 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.36, "acc_stderr": 0.04824181513244218, "acc_norm": 0.36, "acc_norm_stderr": 0.04824181513244218 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.58, "acc_stderr": 0.049604496374885836, "acc_norm": 0.58, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5, "acc_stderr": 0.026919095102908273, "acc_norm": 0.5, "acc_norm_stderr": 0.026919095102908273 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.4233128834355828, "acc_stderr": 0.03881891213334383, "acc_norm": 0.4233128834355828, "acc_norm_stderr": 0.03881891213334383 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.4228395061728395, "acc_stderr": 0.02748747298087159, "acc_norm": 0.4228395061728395, "acc_norm_stderr": 0.02748747298087159 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.29, "acc_stderr": 0.045604802157206845, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206845 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.46113989637305697, "acc_stderr": 0.03597524411734577, "acc_norm": 0.46113989637305697, "acc_norm_stderr": 0.03597524411734577 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2807017543859649, "acc_stderr": 0.042270544512321984, "acc_norm": 0.2807017543859649, "acc_norm_stderr": 0.042270544512321984 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.45871559633027525, "acc_stderr": 0.021364122533881695, "acc_norm": 0.45871559633027525, "acc_norm_stderr": 0.021364122533881695 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.2857142857142857, "acc_stderr": 0.0404061017820884, "acc_norm": 0.2857142857142857, "acc_norm_stderr": 0.0404061017820884 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.46405228758169936, "acc_stderr": 0.028555827516528787, "acc_norm": 0.46405228758169936, "acc_norm_stderr": 0.028555827516528787 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_international_law|5": { "acc": 0.5785123966942148, "acc_stderr": 0.04507732278775087, "acc_norm": 0.5785123966942148, "acc_norm_stderr": 0.04507732278775087 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.40131578947368424, "acc_stderr": 0.03988903703336285, "acc_norm": 0.40131578947368424, "acc_norm_stderr": 0.03988903703336285 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.3627450980392157, "acc_stderr": 0.01945076843250551, "acc_norm": 0.3627450980392157, "acc_norm_stderr": 0.01945076843250551 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3404255319148936, "acc_stderr": 0.02826765748265014, "acc_norm": 0.3404255319148936, "acc_norm_stderr": 0.02826765748265014 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.35714285714285715, "acc_stderr": 0.04547960999764376, "acc_norm": 0.35714285714285715, "acc_norm_stderr": 0.04547960999764376 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4027777777777778, "acc_stderr": 0.033448873829978666, "acc_norm": 0.4027777777777778, "acc_norm_stderr": 0.033448873829978666 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2335195530726257, "acc_stderr": 0.014149575348976276, "acc_norm": 0.2335195530726257, "acc_norm_stderr": 0.014149575348976276 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.37, "acc_stderr": 0.048523658709391, "acc_norm": 0.37, "acc_norm_stderr": 0.048523658709391 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.52, "acc_stderr": 0.050211673156867795, "acc_norm": 0.52, "acc_norm_stderr": 0.050211673156867795 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4007352941176471, "acc_stderr": 0.029768263528933105, "acc_norm": 0.4007352941176471, "acc_norm_stderr": 0.029768263528933105 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.49387755102040815, "acc_stderr": 0.03200682020163908, "acc_norm": 0.49387755102040815, "acc_norm_stderr": 0.03200682020163908 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.5443037974683544, "acc_stderr": 0.032419206846933335, "acc_norm": 0.5443037974683544, "acc_norm_stderr": 0.032419206846933335 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3155149934810952, "acc_stderr": 0.011869184843058642, "acc_norm": 0.3155149934810952, "acc_norm_stderr": 0.011869184843058642 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.44607843137254904, "acc_stderr": 0.034888454513049734, "acc_norm": 0.44607843137254904, "acc_norm_stderr": 0.034888454513049734 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.41818181818181815, "acc_stderr": 0.038517163193983954, "acc_norm": 0.41818181818181815, "acc_norm_stderr": 0.038517163193983954 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.29498164014687883, "mc1_stderr": 0.015964400965589667, "mc2": 0.46872875951621523, "mc2_stderr": 0.01631020915826667 }, "harness|ko_commongen_v2|2": { "acc": 0.3624557260920897, "acc_stderr": 0.016527131240453716, "acc_norm": 0.3825265643447462, "acc_norm_stderr": 0.016709165387228817 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "MNC-Jihun/Mistral-7B-OP-u1k-ver0.6", "model_sha": "23c7a5ec9de97c7c729fb2d9dc76bba8f6cb3406", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }