{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.30887372013651876, "acc_stderr": 0.013501770929344003, "acc_norm": 0.3515358361774744, "acc_norm_stderr": 0.013952413699600933 }, "harness|ko_hellaswag|10": { "acc": 0.391256721768572, "acc_stderr": 0.004870342592915049, "acc_norm": 0.5038836885082653, "acc_norm_stderr": 0.004989630887066195 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.28654970760233917, "acc_stderr": 0.034678266857038245, "acc_norm": 0.28654970760233917, "acc_norm_stderr": 0.034678266857038245 }, "harness|ko_mmlu_management|5": { "acc": 0.1941747572815534, "acc_stderr": 0.03916667762822584, "acc_norm": 0.1941747572815534, "acc_norm_stderr": 0.03916667762822584 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.26309067688378035, "acc_stderr": 0.015745497169049057, "acc_norm": 0.26309067688378035, "acc_norm_stderr": 0.015745497169049057 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.3333333333333333, "acc_stderr": 0.04072314811876837, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.04072314811876837 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.3, "acc_stderr": 0.04605661864718381, "acc_norm": 0.3, "acc_norm_stderr": 0.04605661864718381 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.20851063829787234, "acc_stderr": 0.026556982117838752, "acc_norm": 0.20851063829787234, "acc_norm_stderr": 0.026556982117838752 }, "harness|ko_mmlu_virology|5": { "acc": 0.20481927710843373, "acc_stderr": 0.03141784291663926, "acc_norm": 0.20481927710843373, "acc_norm_stderr": 0.03141784291663926 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.2990353697749196, "acc_stderr": 0.02600330111788514, "acc_norm": 0.2990353697749196, "acc_norm_stderr": 0.02600330111788514 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.2062780269058296, "acc_stderr": 0.027157150479563824, "acc_norm": 0.2062780269058296, "acc_norm_stderr": 0.027157150479563824 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.2366412213740458, "acc_stderr": 0.03727673575596916, "acc_norm": 0.2366412213740458, "acc_norm_stderr": 0.03727673575596916 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.2, "acc_stderr": 0.04020151261036845, "acc_norm": 0.2, "acc_norm_stderr": 0.04020151261036845 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.25252525252525254, "acc_stderr": 0.030954055470365907, "acc_norm": 0.25252525252525254, "acc_norm_stderr": 0.030954055470365907 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.296551724137931, "acc_stderr": 0.03806142687309994, "acc_norm": 0.296551724137931, "acc_norm_stderr": 0.03806142687309994 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.21568627450980393, "acc_stderr": 0.04092563958237657, "acc_norm": 0.21568627450980393, "acc_norm_stderr": 0.04092563958237657 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.21008403361344538, "acc_stderr": 0.026461398717471874, "acc_norm": 0.21008403361344538, "acc_norm_stderr": 0.026461398717471874 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.21025641025641026, "acc_stderr": 0.020660597485026928, "acc_norm": 0.21025641025641026, "acc_norm_stderr": 0.020660597485026928 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.23148148148148148, "acc_stderr": 0.04077494709252627, "acc_norm": 0.23148148148148148, "acc_norm_stderr": 0.04077494709252627 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.27586206896551724, "acc_stderr": 0.03144712581678242, "acc_norm": 0.27586206896551724, "acc_norm_stderr": 0.03144712581678242 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.25161290322580643, "acc_stderr": 0.024685979286239956, "acc_norm": 0.25161290322580643, "acc_norm_stderr": 0.024685979286239956 }, "harness|ko_mmlu_marketing|5": { "acc": 0.2564102564102564, "acc_stderr": 0.028605953702004253, "acc_norm": 0.2564102564102564, "acc_norm_stderr": 0.028605953702004253 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.22264150943396227, "acc_stderr": 0.025604233470899098, "acc_norm": 0.22264150943396227, "acc_norm_stderr": 0.025604233470899098 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.20909090909090908, "acc_stderr": 0.03895091015724135, "acc_norm": 0.20909090909090908, "acc_norm_stderr": 0.03895091015724135 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.26666666666666666, "acc_stderr": 0.026962424325073838, "acc_norm": 0.26666666666666666, "acc_norm_stderr": 0.026962424325073838 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2847682119205298, "acc_stderr": 0.03684881521389023, "acc_norm": 0.2847682119205298, "acc_norm_stderr": 0.03684881521389023 }, "harness|ko_mmlu_sociology|5": { "acc": 0.25870646766169153, "acc_stderr": 0.03096590312357301, "acc_norm": 0.25870646766169153, "acc_norm_stderr": 0.03096590312357301 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.23699421965317918, "acc_stderr": 0.03242414757483098, "acc_norm": 0.23699421965317918, "acc_norm_stderr": 0.03242414757483098 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2751322751322751, "acc_stderr": 0.02300008685906865, "acc_norm": 0.2751322751322751, "acc_norm_stderr": 0.02300008685906865 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2569444444444444, "acc_stderr": 0.03653946969442099, "acc_norm": 0.2569444444444444, "acc_norm_stderr": 0.03653946969442099 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.2, "acc_stderr": 0.04020151261036846, "acc_norm": 0.2, "acc_norm_stderr": 0.04020151261036846 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.2947976878612717, "acc_stderr": 0.02454761779480383, "acc_norm": 0.2947976878612717, "acc_norm_stderr": 0.02454761779480383 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.3006134969325153, "acc_stderr": 0.03602511318806771, "acc_norm": 0.3006134969325153, "acc_norm_stderr": 0.03602511318806771 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.2932098765432099, "acc_stderr": 0.025329888171900926, "acc_norm": 0.2932098765432099, "acc_norm_stderr": 0.025329888171900926 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.27, "acc_stderr": 0.044619604333847394, "acc_norm": 0.27, "acc_norm_stderr": 0.044619604333847394 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.22797927461139897, "acc_stderr": 0.030276909945178256, "acc_norm": 0.22797927461139897, "acc_norm_stderr": 0.030276909945178256 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.24561403508771928, "acc_stderr": 0.04049339297748141, "acc_norm": 0.24561403508771928, "acc_norm_stderr": 0.04049339297748141 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.22752293577981653, "acc_stderr": 0.0179744635787765, "acc_norm": 0.22752293577981653, "acc_norm_stderr": 0.0179744635787765 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.1984126984126984, "acc_stderr": 0.03567016675276863, "acc_norm": 0.1984126984126984, "acc_norm_stderr": 0.03567016675276863 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.2647058823529412, "acc_stderr": 0.025261691219729487, "acc_norm": 0.2647058823529412, "acc_norm_stderr": 0.025261691219729487 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.23, "acc_stderr": 0.04229525846816506, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816506 }, "harness|ko_mmlu_international_law|5": { "acc": 0.371900826446281, "acc_stderr": 0.04412015806624503, "acc_norm": 0.371900826446281, "acc_norm_stderr": 0.04412015806624503 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.3092105263157895, "acc_stderr": 0.03761070869867479, "acc_norm": 0.3092105263157895, "acc_norm_stderr": 0.03761070869867479 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.2761437908496732, "acc_stderr": 0.018087276935663133, "acc_norm": 0.2761437908496732, "acc_norm_stderr": 0.018087276935663133 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.25886524822695034, "acc_stderr": 0.026129572527180848, "acc_norm": 0.25886524822695034, "acc_norm_stderr": 0.026129572527180848 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.24107142857142858, "acc_stderr": 0.04059867246952689, "acc_norm": 0.24107142857142858, "acc_norm_stderr": 0.04059867246952689 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.2175925925925926, "acc_stderr": 0.028139689444859676, "acc_norm": 0.2175925925925926, "acc_norm_stderr": 0.028139689444859676 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24692737430167597, "acc_stderr": 0.014422292204808852, "acc_norm": 0.24692737430167597, "acc_norm_stderr": 0.014422292204808852 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.26, "acc_stderr": 0.044084400227680794, "acc_norm": 0.26, "acc_norm_stderr": 0.044084400227680794 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.16544117647058823, "acc_stderr": 0.022571771025494757, "acc_norm": 0.16544117647058823, "acc_norm_stderr": 0.022571771025494757 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.23265306122448978, "acc_stderr": 0.02704925791589618, "acc_norm": 0.23265306122448978, "acc_norm_stderr": 0.02704925791589618 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.29957805907172996, "acc_stderr": 0.0298180247497531, "acc_norm": 0.29957805907172996, "acc_norm_stderr": 0.0298180247497531 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.25945241199478486, "acc_stderr": 0.011195262076350309, "acc_norm": 0.25945241199478486, "acc_norm_stderr": 0.011195262076350309 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.27450980392156865, "acc_stderr": 0.031321798030832904, "acc_norm": 0.27450980392156865, "acc_norm_stderr": 0.031321798030832904 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.2909090909090909, "acc_stderr": 0.03546563019624337, "acc_norm": 0.2909090909090909, "acc_norm_stderr": 0.03546563019624337 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.23011015911872704, "mc1_stderr": 0.01473455795980776, "mc2": 0.38739814063055383, "mc2_stderr": 0.01474443864761987 }, "harness|ko_commongen_v2|2": { "acc": 0.2939787485242031, "acc_stderr": 0.015663242569091115, "acc_norm": 0.33884297520661155, "acc_norm_stderr": 0.016272952997019124 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "MarkrAI/kyujin-Poly-platypus-ko-12.8b", "model_sha": "cc48d722e28e785ef32b05f4ef0246df177af942", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }