{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.31569965870307165, "acc_stderr": 0.013582571095815291, "acc_norm": 0.3720136518771331, "acc_norm_stderr": 0.014124597881844453 }, "harness|ko_hellaswag|10": { "acc": 0.36317466640111534, "acc_stderr": 0.004799317209902019, "acc_norm": 0.46574387572196774, "acc_norm_stderr": 0.004978056798794869 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.47953216374269003, "acc_stderr": 0.038316105328219316, "acc_norm": 0.47953216374269003, "acc_norm_stderr": 0.038316105328219316 }, "harness|ko_mmlu_management|5": { "acc": 0.5242718446601942, "acc_stderr": 0.049449010929737795, "acc_norm": 0.5242718446601942, "acc_norm_stderr": 0.049449010929737795 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.39208173690932313, "acc_stderr": 0.017458524050147643, "acc_norm": 0.39208173690932313, "acc_norm_stderr": 0.017458524050147643 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.3925925925925926, "acc_stderr": 0.04218506215368879, "acc_norm": 0.3925925925925926, "acc_norm_stderr": 0.04218506215368879 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.2, "acc_stderr": 0.04020151261036844, "acc_norm": 0.2, "acc_norm_stderr": 0.04020151261036844 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.3829787234042553, "acc_stderr": 0.03177821250236922, "acc_norm": 0.3829787234042553, "acc_norm_stderr": 0.03177821250236922 }, "harness|ko_mmlu_virology|5": { "acc": 0.39156626506024095, "acc_stderr": 0.037998574544796354, "acc_norm": 0.39156626506024095, "acc_norm_stderr": 0.037998574544796354 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.47266881028938906, "acc_stderr": 0.028355633568328188, "acc_norm": 0.47266881028938906, "acc_norm_stderr": 0.028355633568328188 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.4349775784753363, "acc_stderr": 0.033272833702713445, "acc_norm": 0.4349775784753363, "acc_norm_stderr": 0.033272833702713445 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.48091603053435117, "acc_stderr": 0.04382094705550988, "acc_norm": 0.48091603053435117, "acc_norm_stderr": 0.04382094705550988 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.42, "acc_stderr": 0.04960449637488583, "acc_norm": 0.42, "acc_norm_stderr": 0.04960449637488583 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.43434343434343436, "acc_stderr": 0.03531505879359183, "acc_norm": 0.43434343434343436, "acc_norm_stderr": 0.03531505879359183 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.5241379310344828, "acc_stderr": 0.041618085035015295, "acc_norm": 0.5241379310344828, "acc_norm_stderr": 0.041618085035015295 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.3431372549019608, "acc_stderr": 0.04724007352383888, "acc_norm": 0.3431372549019608, "acc_norm_stderr": 0.04724007352383888 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.4579831932773109, "acc_stderr": 0.032363611119519416, "acc_norm": 0.4579831932773109, "acc_norm_stderr": 0.032363611119519416 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.4717948717948718, "acc_stderr": 0.025310639254933927, "acc_norm": 0.4717948717948718, "acc_norm_stderr": 0.025310639254933927 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.56, "acc_stderr": 0.0498887651569859, "acc_norm": 0.56, "acc_norm_stderr": 0.0498887651569859 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5, "acc_stderr": 0.04833682445228318, "acc_norm": 0.5, "acc_norm_stderr": 0.04833682445228318 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.4187192118226601, "acc_stderr": 0.034711928605184676, "acc_norm": 0.4187192118226601, "acc_norm_stderr": 0.034711928605184676 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.47419354838709676, "acc_stderr": 0.02840609505765332, "acc_norm": 0.47419354838709676, "acc_norm_stderr": 0.02840609505765332 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6324786324786325, "acc_stderr": 0.03158539157745636, "acc_norm": 0.6324786324786325, "acc_norm_stderr": 0.03158539157745636 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.42641509433962266, "acc_stderr": 0.030437794342983045, "acc_norm": 0.42641509433962266, "acc_norm_stderr": 0.030437794342983045 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.44545454545454544, "acc_stderr": 0.047605488214603246, "acc_norm": 0.44545454545454544, "acc_norm_stderr": 0.047605488214603246 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.35555555555555557, "acc_stderr": 0.029185714949857392, "acc_norm": 0.35555555555555557, "acc_norm_stderr": 0.029185714949857392 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.3576158940397351, "acc_stderr": 0.03913453431177258, "acc_norm": 0.3576158940397351, "acc_norm_stderr": 0.03913453431177258 }, "harness|ko_mmlu_sociology|5": { "acc": 0.5572139303482587, "acc_stderr": 0.03512310964123935, "acc_norm": 0.5572139303482587, "acc_norm_stderr": 0.03512310964123935 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.4508670520231214, "acc_stderr": 0.03794012674697029, "acc_norm": 0.4508670520231214, "acc_norm_stderr": 0.03794012674697029 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.36772486772486773, "acc_stderr": 0.02483383982556242, "acc_norm": 0.36772486772486773, "acc_norm_stderr": 0.02483383982556242 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3819444444444444, "acc_stderr": 0.040629907841466674, "acc_norm": 0.3819444444444444, "acc_norm_stderr": 0.040629907841466674 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.62, "acc_stderr": 0.04878317312145633, "acc_norm": 0.62, "acc_norm_stderr": 0.04878317312145633 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.476878612716763, "acc_stderr": 0.026890297881303125, "acc_norm": 0.476878612716763, "acc_norm_stderr": 0.026890297881303125 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.34355828220858897, "acc_stderr": 0.037311335196738925, "acc_norm": 0.34355828220858897, "acc_norm_stderr": 0.037311335196738925 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.4537037037037037, "acc_stderr": 0.027701228468542595, "acc_norm": 0.4537037037037037, "acc_norm_stderr": 0.027701228468542595 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.38, "acc_stderr": 0.048783173121456316, "acc_norm": 0.38, "acc_norm_stderr": 0.048783173121456316 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.5284974093264249, "acc_stderr": 0.03602573571288441, "acc_norm": 0.5284974093264249, "acc_norm_stderr": 0.03602573571288441 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.3684210526315789, "acc_stderr": 0.04537815354939391, "acc_norm": 0.3684210526315789, "acc_norm_stderr": 0.04537815354939391 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.5155963302752293, "acc_stderr": 0.02142689153920805, "acc_norm": 0.5155963302752293, "acc_norm_stderr": 0.02142689153920805 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.30952380952380953, "acc_stderr": 0.04134913018303316, "acc_norm": 0.30952380952380953, "acc_norm_stderr": 0.04134913018303316 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5098039215686274, "acc_stderr": 0.028624412550167958, "acc_norm": 0.5098039215686274, "acc_norm_stderr": 0.028624412550167958 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.51, "acc_stderr": 0.05024183937956914, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956914 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6446280991735537, "acc_stderr": 0.0436923632657398, "acc_norm": 0.6446280991735537, "acc_norm_stderr": 0.0436923632657398 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.4868421052631579, "acc_stderr": 0.04067533136309173, "acc_norm": 0.4868421052631579, "acc_norm_stderr": 0.04067533136309173 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.4068627450980392, "acc_stderr": 0.019873802005061177, "acc_norm": 0.4068627450980392, "acc_norm_stderr": 0.019873802005061177 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.33687943262411346, "acc_stderr": 0.02819553487396673, "acc_norm": 0.33687943262411346, "acc_norm_stderr": 0.02819553487396673 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.375, "acc_stderr": 0.04595091388086298, "acc_norm": 0.375, "acc_norm_stderr": 0.04595091388086298 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.39351851851851855, "acc_stderr": 0.03331747876370312, "acc_norm": 0.39351851851851855, "acc_norm_stderr": 0.03331747876370312 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.26145251396648045, "acc_stderr": 0.014696599650364555, "acc_norm": 0.26145251396648045, "acc_norm_stderr": 0.014696599650364555 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.34, "acc_stderr": 0.04760952285695235, "acc_norm": 0.34, "acc_norm_stderr": 0.04760952285695235 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.61, "acc_stderr": 0.04902071300001974, "acc_norm": 0.61, "acc_norm_stderr": 0.04902071300001974 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.35661764705882354, "acc_stderr": 0.02909720956841196, "acc_norm": 0.35661764705882354, "acc_norm_stderr": 0.02909720956841196 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.563265306122449, "acc_stderr": 0.031751952375833226, "acc_norm": 0.563265306122449, "acc_norm_stderr": 0.031751952375833226 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.540084388185654, "acc_stderr": 0.03244246810187913, "acc_norm": 0.540084388185654, "acc_norm_stderr": 0.03244246810187913 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.31290743155149936, "acc_stderr": 0.011842529823062997, "acc_norm": 0.31290743155149936, "acc_norm_stderr": 0.011842529823062997 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.4362745098039216, "acc_stderr": 0.03480693138457039, "acc_norm": 0.4362745098039216, "acc_norm_stderr": 0.03480693138457039 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.5212121212121212, "acc_stderr": 0.03900828913737302, "acc_norm": 0.5212121212121212, "acc_norm_stderr": 0.03900828913737302 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2729498164014688, "mc1_stderr": 0.015594753632006516, "mc2": 0.4494211990695322, "mc2_stderr": 0.016190690301781833 }, "harness|ko_commongen_v2|2": { "acc": 0.3990554899645809, "acc_stderr": 0.0168363772928493, "acc_norm": 0.4498229043683589, "acc_norm_stderr": 0.01710357334382571 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "4yo1/llama3-eng-ko-8b-sl", "model_sha": "5f82a1a6e36c50db9a2ee4b815d742b27cdb6023", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }