|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.31569965870307165, |
|
"acc_stderr": 0.013582571095815291, |
|
"acc_norm": 0.3720136518771331, |
|
"acc_norm_stderr": 0.014124597881844453 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.36317466640111534, |
|
"acc_stderr": 0.004799317209902019, |
|
"acc_norm": 0.46574387572196774, |
|
"acc_norm_stderr": 0.004978056798794869 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.47953216374269003, |
|
"acc_stderr": 0.038316105328219316, |
|
"acc_norm": 0.47953216374269003, |
|
"acc_norm_stderr": 0.038316105328219316 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5242718446601942, |
|
"acc_stderr": 0.049449010929737795, |
|
"acc_norm": 0.5242718446601942, |
|
"acc_norm_stderr": 0.049449010929737795 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.39208173690932313, |
|
"acc_stderr": 0.017458524050147643, |
|
"acc_norm": 0.39208173690932313, |
|
"acc_norm_stderr": 0.017458524050147643 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.3925925925925926, |
|
"acc_stderr": 0.04218506215368879, |
|
"acc_norm": 0.3925925925925926, |
|
"acc_norm_stderr": 0.04218506215368879 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.04020151261036844, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.04020151261036844 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.3829787234042553, |
|
"acc_stderr": 0.03177821250236922, |
|
"acc_norm": 0.3829787234042553, |
|
"acc_norm_stderr": 0.03177821250236922 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.39156626506024095, |
|
"acc_stderr": 0.037998574544796354, |
|
"acc_norm": 0.39156626506024095, |
|
"acc_norm_stderr": 0.037998574544796354 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.47266881028938906, |
|
"acc_stderr": 0.028355633568328188, |
|
"acc_norm": 0.47266881028938906, |
|
"acc_norm_stderr": 0.028355633568328188 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.4349775784753363, |
|
"acc_stderr": 0.033272833702713445, |
|
"acc_norm": 0.4349775784753363, |
|
"acc_norm_stderr": 0.033272833702713445 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.48091603053435117, |
|
"acc_stderr": 0.04382094705550988, |
|
"acc_norm": 0.48091603053435117, |
|
"acc_norm_stderr": 0.04382094705550988 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.04960449637488583, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.04960449637488583 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.43434343434343436, |
|
"acc_stderr": 0.03531505879359183, |
|
"acc_norm": 0.43434343434343436, |
|
"acc_norm_stderr": 0.03531505879359183 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.5241379310344828, |
|
"acc_stderr": 0.041618085035015295, |
|
"acc_norm": 0.5241379310344828, |
|
"acc_norm_stderr": 0.041618085035015295 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.3431372549019608, |
|
"acc_stderr": 0.04724007352383888, |
|
"acc_norm": 0.3431372549019608, |
|
"acc_norm_stderr": 0.04724007352383888 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.4579831932773109, |
|
"acc_stderr": 0.032363611119519416, |
|
"acc_norm": 0.4579831932773109, |
|
"acc_norm_stderr": 0.032363611119519416 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4717948717948718, |
|
"acc_stderr": 0.025310639254933927, |
|
"acc_norm": 0.4717948717948718, |
|
"acc_norm_stderr": 0.025310639254933927 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.56, |
|
"acc_stderr": 0.0498887651569859, |
|
"acc_norm": 0.56, |
|
"acc_norm_stderr": 0.0498887651569859 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.04833682445228318, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.04833682445228318 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.4187192118226601, |
|
"acc_stderr": 0.034711928605184676, |
|
"acc_norm": 0.4187192118226601, |
|
"acc_norm_stderr": 0.034711928605184676 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.47419354838709676, |
|
"acc_stderr": 0.02840609505765332, |
|
"acc_norm": 0.47419354838709676, |
|
"acc_norm_stderr": 0.02840609505765332 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6324786324786325, |
|
"acc_stderr": 0.03158539157745636, |
|
"acc_norm": 0.6324786324786325, |
|
"acc_norm_stderr": 0.03158539157745636 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.42641509433962266, |
|
"acc_stderr": 0.030437794342983045, |
|
"acc_norm": 0.42641509433962266, |
|
"acc_norm_stderr": 0.030437794342983045 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.44545454545454544, |
|
"acc_stderr": 0.047605488214603246, |
|
"acc_norm": 0.44545454545454544, |
|
"acc_norm_stderr": 0.047605488214603246 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.35555555555555557, |
|
"acc_stderr": 0.029185714949857392, |
|
"acc_norm": 0.35555555555555557, |
|
"acc_norm_stderr": 0.029185714949857392 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3576158940397351, |
|
"acc_stderr": 0.03913453431177258, |
|
"acc_norm": 0.3576158940397351, |
|
"acc_norm_stderr": 0.03913453431177258 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.5572139303482587, |
|
"acc_stderr": 0.03512310964123935, |
|
"acc_norm": 0.5572139303482587, |
|
"acc_norm_stderr": 0.03512310964123935 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.4508670520231214, |
|
"acc_stderr": 0.03794012674697029, |
|
"acc_norm": 0.4508670520231214, |
|
"acc_norm_stderr": 0.03794012674697029 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.36772486772486773, |
|
"acc_stderr": 0.02483383982556242, |
|
"acc_norm": 0.36772486772486773, |
|
"acc_norm_stderr": 0.02483383982556242 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3819444444444444, |
|
"acc_stderr": 0.040629907841466674, |
|
"acc_norm": 0.3819444444444444, |
|
"acc_norm_stderr": 0.040629907841466674 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.62, |
|
"acc_stderr": 0.04878317312145633, |
|
"acc_norm": 0.62, |
|
"acc_norm_stderr": 0.04878317312145633 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.476878612716763, |
|
"acc_stderr": 0.026890297881303125, |
|
"acc_norm": 0.476878612716763, |
|
"acc_norm_stderr": 0.026890297881303125 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.34355828220858897, |
|
"acc_stderr": 0.037311335196738925, |
|
"acc_norm": 0.34355828220858897, |
|
"acc_norm_stderr": 0.037311335196738925 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.4537037037037037, |
|
"acc_stderr": 0.027701228468542595, |
|
"acc_norm": 0.4537037037037037, |
|
"acc_norm_stderr": 0.027701228468542595 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.048783173121456316, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.048783173121456316 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.5284974093264249, |
|
"acc_stderr": 0.03602573571288441, |
|
"acc_norm": 0.5284974093264249, |
|
"acc_norm_stderr": 0.03602573571288441 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.3684210526315789, |
|
"acc_stderr": 0.04537815354939391, |
|
"acc_norm": 0.3684210526315789, |
|
"acc_norm_stderr": 0.04537815354939391 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.5155963302752293, |
|
"acc_stderr": 0.02142689153920805, |
|
"acc_norm": 0.5155963302752293, |
|
"acc_norm_stderr": 0.02142689153920805 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.30952380952380953, |
|
"acc_stderr": 0.04134913018303316, |
|
"acc_norm": 0.30952380952380953, |
|
"acc_norm_stderr": 0.04134913018303316 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5098039215686274, |
|
"acc_stderr": 0.028624412550167958, |
|
"acc_norm": 0.5098039215686274, |
|
"acc_norm_stderr": 0.028624412550167958 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.51, |
|
"acc_stderr": 0.05024183937956914, |
|
"acc_norm": 0.51, |
|
"acc_norm_stderr": 0.05024183937956914 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6446280991735537, |
|
"acc_stderr": 0.0436923632657398, |
|
"acc_norm": 0.6446280991735537, |
|
"acc_norm_stderr": 0.0436923632657398 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.4868421052631579, |
|
"acc_stderr": 0.04067533136309173, |
|
"acc_norm": 0.4868421052631579, |
|
"acc_norm_stderr": 0.04067533136309173 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.4068627450980392, |
|
"acc_stderr": 0.019873802005061177, |
|
"acc_norm": 0.4068627450980392, |
|
"acc_norm_stderr": 0.019873802005061177 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.33687943262411346, |
|
"acc_stderr": 0.02819553487396673, |
|
"acc_norm": 0.33687943262411346, |
|
"acc_norm_stderr": 0.02819553487396673 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.375, |
|
"acc_stderr": 0.04595091388086298, |
|
"acc_norm": 0.375, |
|
"acc_norm_stderr": 0.04595091388086298 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.39351851851851855, |
|
"acc_stderr": 0.03331747876370312, |
|
"acc_norm": 0.39351851851851855, |
|
"acc_norm_stderr": 0.03331747876370312 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.26145251396648045, |
|
"acc_stderr": 0.014696599650364555, |
|
"acc_norm": 0.26145251396648045, |
|
"acc_norm_stderr": 0.014696599650364555 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.34, |
|
"acc_stderr": 0.04760952285695235, |
|
"acc_norm": 0.34, |
|
"acc_norm_stderr": 0.04760952285695235 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.61, |
|
"acc_stderr": 0.04902071300001974, |
|
"acc_norm": 0.61, |
|
"acc_norm_stderr": 0.04902071300001974 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.35661764705882354, |
|
"acc_stderr": 0.02909720956841196, |
|
"acc_norm": 0.35661764705882354, |
|
"acc_norm_stderr": 0.02909720956841196 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.563265306122449, |
|
"acc_stderr": 0.031751952375833226, |
|
"acc_norm": 0.563265306122449, |
|
"acc_norm_stderr": 0.031751952375833226 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.540084388185654, |
|
"acc_stderr": 0.03244246810187913, |
|
"acc_norm": 0.540084388185654, |
|
"acc_norm_stderr": 0.03244246810187913 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.31290743155149936, |
|
"acc_stderr": 0.011842529823062997, |
|
"acc_norm": 0.31290743155149936, |
|
"acc_norm_stderr": 0.011842529823062997 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.4362745098039216, |
|
"acc_stderr": 0.03480693138457039, |
|
"acc_norm": 0.4362745098039216, |
|
"acc_norm_stderr": 0.03480693138457039 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.5212121212121212, |
|
"acc_stderr": 0.03900828913737302, |
|
"acc_norm": 0.5212121212121212, |
|
"acc_norm_stderr": 0.03900828913737302 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2729498164014688, |
|
"mc1_stderr": 0.015594753632006516, |
|
"mc2": 0.4494211990695322, |
|
"mc2_stderr": 0.016190690301781833 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.3990554899645809, |
|
"acc_stderr": 0.0168363772928493, |
|
"acc_norm": 0.4498229043683589, |
|
"acc_norm_stderr": 0.01710357334382571 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "4yo1/llama3-eng-ko-8b-sl", |
|
"model_sha": "5f82a1a6e36c50db9a2ee4b815d742b27cdb6023", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |