{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.2645051194539249, "acc_stderr": 0.012889272949313368, "acc_norm": 0.32849829351535836, "acc_norm_stderr": 0.013724978465537378 }, "harness|ko_hellaswag|10": { "acc": 0.32682732523401714, "acc_stderr": 0.004680949283855315, "acc_norm": 0.3868751244771958, "acc_norm_stderr": 0.004860393011974685 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.38596491228070173, "acc_stderr": 0.03733756969066163, "acc_norm": 0.38596491228070173, "acc_norm_stderr": 0.03733756969066163 }, "harness|ko_mmlu_management|5": { "acc": 0.4563106796116505, "acc_stderr": 0.04931801994220414, "acc_norm": 0.4563106796116505, "acc_norm_stderr": 0.04931801994220414 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.38569604086845466, "acc_stderr": 0.01740647661921291, "acc_norm": 0.38569604086845466, "acc_norm_stderr": 0.01740647661921291 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.3037037037037037, "acc_stderr": 0.039725528847851375, "acc_norm": 0.3037037037037037, "acc_norm_stderr": 0.039725528847851375 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.29, "acc_stderr": 0.04560480215720683, "acc_norm": 0.29, "acc_norm_stderr": 0.04560480215720683 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.32340425531914896, "acc_stderr": 0.030579442773610337, "acc_norm": 0.32340425531914896, "acc_norm_stderr": 0.030579442773610337 }, "harness|ko_mmlu_virology|5": { "acc": 0.3192771084337349, "acc_stderr": 0.036293353299478595, "acc_norm": 0.3192771084337349, "acc_norm_stderr": 0.036293353299478595 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.41479099678456594, "acc_stderr": 0.027982680459759563, "acc_norm": 0.41479099678456594, "acc_norm_stderr": 0.027982680459759563 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.36771300448430494, "acc_stderr": 0.03236198350928276, "acc_norm": 0.36771300448430494, "acc_norm_stderr": 0.03236198350928276 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.3893129770992366, "acc_stderr": 0.04276486542814591, "acc_norm": 0.3893129770992366, "acc_norm_stderr": 0.04276486542814591 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.41, "acc_stderr": 0.04943110704237102, "acc_norm": 0.41, "acc_norm_stderr": 0.04943110704237102 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.4444444444444444, "acc_stderr": 0.03540294377095367, "acc_norm": 0.4444444444444444, "acc_norm_stderr": 0.03540294377095367 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.45517241379310347, "acc_stderr": 0.04149886942192117, "acc_norm": 0.45517241379310347, "acc_norm_stderr": 0.04149886942192117 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.27450980392156865, "acc_stderr": 0.044405219061793254, "acc_norm": 0.27450980392156865, "acc_norm_stderr": 0.044405219061793254 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.39915966386554624, "acc_stderr": 0.03181110032413926, "acc_norm": 0.39915966386554624, "acc_norm_stderr": 0.03181110032413926 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.37435897435897436, "acc_stderr": 0.024537591572830524, "acc_norm": 0.37435897435897436, "acc_norm_stderr": 0.024537591572830524 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.51, "acc_stderr": 0.05024183937956912, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956912 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.4074074074074074, "acc_stderr": 0.04750077341199985, "acc_norm": 0.4074074074074074, "acc_norm_stderr": 0.04750077341199985 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3448275862068966, "acc_stderr": 0.033442837442804574, "acc_norm": 0.3448275862068966, "acc_norm_stderr": 0.033442837442804574 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.3774193548387097, "acc_stderr": 0.027575960723278253, "acc_norm": 0.3774193548387097, "acc_norm_stderr": 0.027575960723278253 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6111111111111112, "acc_stderr": 0.031937057262002924, "acc_norm": 0.6111111111111112, "acc_norm_stderr": 0.031937057262002924 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.43018867924528303, "acc_stderr": 0.03047144586718323, "acc_norm": 0.43018867924528303, "acc_norm_stderr": 0.03047144586718323 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.37272727272727274, "acc_stderr": 0.04631381319425463, "acc_norm": 0.37272727272727274, "acc_norm_stderr": 0.04631381319425463 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.3037037037037037, "acc_stderr": 0.02803792996911499, "acc_norm": 0.3037037037037037, "acc_norm_stderr": 0.02803792996911499 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2980132450331126, "acc_stderr": 0.037345356767871984, "acc_norm": 0.2980132450331126, "acc_norm_stderr": 0.037345356767871984 }, "harness|ko_mmlu_sociology|5": { "acc": 0.47761194029850745, "acc_stderr": 0.035319879302087305, "acc_norm": 0.47761194029850745, "acc_norm_stderr": 0.035319879302087305 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.3699421965317919, "acc_stderr": 0.03681229633394319, "acc_norm": 0.3699421965317919, "acc_norm_stderr": 0.03681229633394319 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.31216931216931215, "acc_stderr": 0.023865206836972602, "acc_norm": 0.31216931216931215, "acc_norm_stderr": 0.023865206836972602 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3125, "acc_stderr": 0.038760854559127644, "acc_norm": 0.3125, "acc_norm_stderr": 0.038760854559127644 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.35, "acc_stderr": 0.047937248544110196, "acc_norm": 0.35, "acc_norm_stderr": 0.047937248544110196 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.51, "acc_stderr": 0.05024183937956911, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956911 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.4046242774566474, "acc_stderr": 0.02642481659400985, "acc_norm": 0.4046242774566474, "acc_norm_stderr": 0.02642481659400985 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.37423312883435583, "acc_stderr": 0.038020681028996146, "acc_norm": 0.37423312883435583, "acc_norm_stderr": 0.038020681028996146 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.3271604938271605, "acc_stderr": 0.026105673861409814, "acc_norm": 0.3271604938271605, "acc_norm_stderr": 0.026105673861409814 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.41968911917098445, "acc_stderr": 0.03561587327685884, "acc_norm": 0.41968911917098445, "acc_norm_stderr": 0.03561587327685884 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.22807017543859648, "acc_stderr": 0.03947152782669415, "acc_norm": 0.22807017543859648, "acc_norm_stderr": 0.03947152782669415 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.3834862385321101, "acc_stderr": 0.020847156641915984, "acc_norm": 0.3834862385321101, "acc_norm_stderr": 0.020847156641915984 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.42857142857142855, "acc_stderr": 0.0442626668137991, "acc_norm": 0.42857142857142855, "acc_norm_stderr": 0.0442626668137991 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.4477124183006536, "acc_stderr": 0.028472938478033526, "acc_norm": 0.4477124183006536, "acc_norm_stderr": 0.028472938478033526 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.4, "acc_stderr": 0.049236596391733084, "acc_norm": 0.4, "acc_norm_stderr": 0.049236596391733084 }, "harness|ko_mmlu_international_law|5": { "acc": 0.5371900826446281, "acc_stderr": 0.04551711196104218, "acc_norm": 0.5371900826446281, "acc_norm_stderr": 0.04551711196104218 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.3223684210526316, "acc_stderr": 0.03803510248351585, "acc_norm": 0.3223684210526316, "acc_norm_stderr": 0.03803510248351585 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.3055555555555556, "acc_stderr": 0.01863559403442397, "acc_norm": 0.3055555555555556, "acc_norm_stderr": 0.01863559403442397 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3262411347517731, "acc_stderr": 0.027968453043563168, "acc_norm": 0.3262411347517731, "acc_norm_stderr": 0.027968453043563168 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.30357142857142855, "acc_stderr": 0.043642261558410445, "acc_norm": 0.30357142857142855, "acc_norm_stderr": 0.043642261558410445 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.41203703703703703, "acc_stderr": 0.03356787758160835, "acc_norm": 0.41203703703703703, "acc_norm_stderr": 0.03356787758160835 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.31843575418994413, "acc_stderr": 0.015581008080360274, "acc_norm": 0.31843575418994413, "acc_norm_stderr": 0.015581008080360274 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.42, "acc_stderr": 0.049604496374885836, "acc_norm": 0.42, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.44, "acc_stderr": 0.04988876515698589, "acc_norm": 0.44, "acc_norm_stderr": 0.04988876515698589 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.38235294117647056, "acc_stderr": 0.029520095697687758, "acc_norm": 0.38235294117647056, "acc_norm_stderr": 0.029520095697687758 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.44081632653061226, "acc_stderr": 0.03178419114175363, "acc_norm": 0.44081632653061226, "acc_norm_stderr": 0.03178419114175363 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.47257383966244726, "acc_stderr": 0.032498227183013026, "acc_norm": 0.47257383966244726, "acc_norm_stderr": 0.032498227183013026 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.2966101694915254, "acc_stderr": 0.011665946586082868, "acc_norm": 0.2966101694915254, "acc_norm_stderr": 0.011665946586082868 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.3235294117647059, "acc_stderr": 0.03283472056108567, "acc_norm": 0.3235294117647059, "acc_norm_stderr": 0.03283472056108567 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.3515151515151515, "acc_stderr": 0.0372820699868265, "acc_norm": 0.3515151515151515, "acc_norm_stderr": 0.0372820699868265 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.3243574051407589, "mc1_stderr": 0.01638797677964793, "mc2": 0.49917419306073907, "mc2_stderr": 0.016202138687957245 }, "harness|ko_commongen_v2|2": { "acc": 0.2857142857142857, "acc_stderr": 0.015531620786986743, "acc_norm": 0.3565525383707202, "acc_norm_stderr": 0.016467706981527445 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "mistralai/Mistral-7B-Instruct-v0.1", "model_sha": "7ad5799710574ba1c1d953eba3077af582f3a773", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }