{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.28668941979522183, "acc_stderr": 0.013214986329274762, "acc_norm": 0.35665529010238906, "acc_norm_stderr": 0.013998056902620196 }, "harness|ko_hellaswag|10": { "acc": 0.35570603465445133, "acc_stderr": 0.00477748315963403, "acc_norm": 0.4393547102170882, "acc_norm_stderr": 0.004952942072999276 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.38596491228070173, "acc_stderr": 0.03733756969066165, "acc_norm": 0.38596491228070173, "acc_norm_stderr": 0.03733756969066165 }, "harness|ko_mmlu_management|5": { "acc": 0.24271844660194175, "acc_stderr": 0.042450224863844956, "acc_norm": 0.24271844660194175, "acc_norm_stderr": 0.042450224863844956 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.3652618135376756, "acc_stderr": 0.01721853002883864, "acc_norm": 0.3652618135376756, "acc_norm_stderr": 0.01721853002883864 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.34814814814814815, "acc_stderr": 0.041153246103369526, "acc_norm": 0.34814814814814815, "acc_norm_stderr": 0.041153246103369526 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.2851063829787234, "acc_stderr": 0.029513196625539345, "acc_norm": 0.2851063829787234, "acc_norm_stderr": 0.029513196625539345 }, "harness|ko_mmlu_virology|5": { "acc": 0.23493975903614459, "acc_stderr": 0.03300533186128922, "acc_norm": 0.23493975903614459, "acc_norm_stderr": 0.03300533186128922 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.33440514469453375, "acc_stderr": 0.026795422327893947, "acc_norm": 0.33440514469453375, "acc_norm_stderr": 0.026795422327893947 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.32286995515695066, "acc_stderr": 0.031381476375754995, "acc_norm": 0.32286995515695066, "acc_norm_stderr": 0.031381476375754995 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.24427480916030533, "acc_stderr": 0.03768335959728743, "acc_norm": 0.24427480916030533, "acc_norm_stderr": 0.03768335959728743 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.26, "acc_stderr": 0.044084400227680794, "acc_norm": 0.26, "acc_norm_stderr": 0.044084400227680794 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.2828282828282828, "acc_stderr": 0.032087795587867514, "acc_norm": 0.2828282828282828, "acc_norm_stderr": 0.032087795587867514 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.30344827586206896, "acc_stderr": 0.03831226048850333, "acc_norm": 0.30344827586206896, "acc_norm_stderr": 0.03831226048850333 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.16666666666666666, "acc_stderr": 0.03708284662416544, "acc_norm": 0.16666666666666666, "acc_norm_stderr": 0.03708284662416544 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.24369747899159663, "acc_stderr": 0.027886828078380565, "acc_norm": 0.24369747899159663, "acc_norm_stderr": 0.027886828078380565 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.25384615384615383, "acc_stderr": 0.022066054378726257, "acc_norm": 0.25384615384615383, "acc_norm_stderr": 0.022066054378726257 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.39, "acc_stderr": 0.04902071300001974, "acc_norm": 0.39, "acc_norm_stderr": 0.04902071300001974 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.37037037037037035, "acc_stderr": 0.04668408033024931, "acc_norm": 0.37037037037037035, "acc_norm_stderr": 0.04668408033024931 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3103448275862069, "acc_stderr": 0.03255086769970103, "acc_norm": 0.3103448275862069, "acc_norm_stderr": 0.03255086769970103 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.2709677419354839, "acc_stderr": 0.02528441611490016, "acc_norm": 0.2709677419354839, "acc_norm_stderr": 0.02528441611490016 }, "harness|ko_mmlu_marketing|5": { "acc": 0.4358974358974359, "acc_stderr": 0.032485775115783995, "acc_norm": 0.4358974358974359, "acc_norm_stderr": 0.032485775115783995 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.24150943396226415, "acc_stderr": 0.026341480371118345, "acc_norm": 0.24150943396226415, "acc_norm_stderr": 0.026341480371118345 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.33636363636363636, "acc_stderr": 0.04525393596302506, "acc_norm": 0.33636363636363636, "acc_norm_stderr": 0.04525393596302506 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.26296296296296295, "acc_stderr": 0.026842057873833706, "acc_norm": 0.26296296296296295, "acc_norm_stderr": 0.026842057873833706 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.24503311258278146, "acc_stderr": 0.03511807571804725, "acc_norm": 0.24503311258278146, "acc_norm_stderr": 0.03511807571804725 }, "harness|ko_mmlu_sociology|5": { "acc": 0.373134328358209, "acc_stderr": 0.034198326081760065, "acc_norm": 0.373134328358209, "acc_norm_stderr": 0.034198326081760065 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.2832369942196532, "acc_stderr": 0.03435568056047875, "acc_norm": 0.2832369942196532, "acc_norm_stderr": 0.03435568056047875 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.291005291005291, "acc_stderr": 0.023393826500484875, "acc_norm": 0.291005291005291, "acc_norm_stderr": 0.023393826500484875 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2569444444444444, "acc_stderr": 0.03653946969442099, "acc_norm": 0.2569444444444444, "acc_norm_stderr": 0.03653946969442099 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.18, "acc_stderr": 0.03861229196653695, "acc_norm": 0.18, "acc_norm_stderr": 0.03861229196653695 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.39, "acc_stderr": 0.04902071300001974, "acc_norm": 0.39, "acc_norm_stderr": 0.04902071300001974 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.32947976878612717, "acc_stderr": 0.0253052581318797, "acc_norm": 0.32947976878612717, "acc_norm_stderr": 0.0253052581318797 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.32515337423312884, "acc_stderr": 0.03680350371286462, "acc_norm": 0.32515337423312884, "acc_norm_stderr": 0.03680350371286462 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.3271604938271605, "acc_stderr": 0.02610567386140981, "acc_norm": 0.3271604938271605, "acc_norm_stderr": 0.02610567386140981 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.26, "acc_stderr": 0.04408440022768078, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768078 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.30569948186528495, "acc_stderr": 0.03324837939758159, "acc_norm": 0.30569948186528495, "acc_norm_stderr": 0.03324837939758159 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2894736842105263, "acc_stderr": 0.04266339443159394, "acc_norm": 0.2894736842105263, "acc_norm_stderr": 0.04266339443159394 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.26788990825688075, "acc_stderr": 0.018987462257978652, "acc_norm": 0.26788990825688075, "acc_norm_stderr": 0.018987462257978652 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.18253968253968253, "acc_stderr": 0.034550710191021496, "acc_norm": 0.18253968253968253, "acc_norm_stderr": 0.034550710191021496 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.3006535947712418, "acc_stderr": 0.026256053835718968, "acc_norm": 0.3006535947712418, "acc_norm_stderr": 0.026256053835718968 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_international_law|5": { "acc": 0.4462809917355372, "acc_stderr": 0.0453793517794788, "acc_norm": 0.4462809917355372, "acc_norm_stderr": 0.0453793517794788 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.28289473684210525, "acc_stderr": 0.03665349695640767, "acc_norm": 0.28289473684210525, "acc_norm_stderr": 0.03665349695640767 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.2875816993464052, "acc_stderr": 0.018311653053648222, "acc_norm": 0.2875816993464052, "acc_norm_stderr": 0.018311653053648222 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.2978723404255319, "acc_stderr": 0.027281608344469414, "acc_norm": 0.2978723404255319, "acc_norm_stderr": 0.027281608344469414 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.25, "acc_stderr": 0.04109974682633932, "acc_norm": 0.25, "acc_norm_stderr": 0.04109974682633932 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.21296296296296297, "acc_stderr": 0.027920963147993666, "acc_norm": 0.21296296296296297, "acc_norm_stderr": 0.027920963147993666 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24692737430167597, "acc_stderr": 0.014422292204808852, "acc_norm": 0.24692737430167597, "acc_norm_stderr": 0.014422292204808852 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.26, "acc_stderr": 0.04408440022768079, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768079 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.37, "acc_stderr": 0.048523658709391, "acc_norm": 0.37, "acc_norm_stderr": 0.048523658709391 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.19852941176470587, "acc_stderr": 0.024231013370541097, "acc_norm": 0.19852941176470587, "acc_norm_stderr": 0.024231013370541097 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.24081632653061225, "acc_stderr": 0.02737294220178816, "acc_norm": 0.24081632653061225, "acc_norm_stderr": 0.02737294220178816 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.3628691983122363, "acc_stderr": 0.03129920825530213, "acc_norm": 0.3628691983122363, "acc_norm_stderr": 0.03129920825530213 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.27249022164276404, "acc_stderr": 0.01137165829431153, "acc_norm": 0.27249022164276404, "acc_norm_stderr": 0.01137165829431153 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.3333333333333333, "acc_stderr": 0.03308611113236436, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.03308611113236436 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.3393939393939394, "acc_stderr": 0.03697442205031596, "acc_norm": 0.3393939393939394, "acc_norm_stderr": 0.03697442205031596 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2839657282741738, "mc1_stderr": 0.015785370858396708, "mc2": 0.4469469691662156, "mc2_stderr": 0.015668694918169947 }, "harness|ko_commongen_v2|2": { "acc": 0.2939787485242031, "acc_stderr": 0.015663242569091115, "acc_norm": 0.40613931523022434, "acc_norm_stderr": 0.016884749503191392 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "maum-ai/llamaum-13b-instruct-s", "model_sha": "d9a9f9c019908c2d302da856473891095ad81940", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }