|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3916382252559727, |
|
"acc_stderr": 0.014264122124938213, |
|
"acc_norm": 0.4564846416382253, |
|
"acc_norm_stderr": 0.014555949760496435 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.42272455686118304, |
|
"acc_stderr": 0.00492982833760698, |
|
"acc_norm": 0.5592511451902011, |
|
"acc_norm_stderr": 0.004954622308739005 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.47953216374269003, |
|
"acc_stderr": 0.0383161053282193, |
|
"acc_norm": 0.47953216374269003, |
|
"acc_norm_stderr": 0.0383161053282193 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.46601941747572817, |
|
"acc_stderr": 0.04939291447273481, |
|
"acc_norm": 0.46601941747572817, |
|
"acc_norm_stderr": 0.04939291447273481 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5146871008939975, |
|
"acc_stderr": 0.017872248024429122, |
|
"acc_norm": 0.5146871008939975, |
|
"acc_norm_stderr": 0.017872248024429122 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.45185185185185184, |
|
"acc_stderr": 0.04299268905480863, |
|
"acc_norm": 0.45185185185185184, |
|
"acc_norm_stderr": 0.04299268905480863 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.045604802157206824, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.045604802157206824 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.4297872340425532, |
|
"acc_stderr": 0.03236214467715564, |
|
"acc_norm": 0.4297872340425532, |
|
"acc_norm_stderr": 0.03236214467715564 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.40963855421686746, |
|
"acc_stderr": 0.038284011150790206, |
|
"acc_norm": 0.40963855421686746, |
|
"acc_norm_stderr": 0.038284011150790206 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.4887459807073955, |
|
"acc_stderr": 0.028390897396863537, |
|
"acc_norm": 0.4887459807073955, |
|
"acc_norm_stderr": 0.028390897396863537 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.515695067264574, |
|
"acc_stderr": 0.0335412657542081, |
|
"acc_norm": 0.515695067264574, |
|
"acc_norm_stderr": 0.0335412657542081 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.4351145038167939, |
|
"acc_stderr": 0.04348208051644858, |
|
"acc_norm": 0.4351145038167939, |
|
"acc_norm_stderr": 0.04348208051644858 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.39, |
|
"acc_stderr": 0.04902071300001974, |
|
"acc_norm": 0.39, |
|
"acc_norm_stderr": 0.04902071300001974 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5050505050505051, |
|
"acc_stderr": 0.035621707606254015, |
|
"acc_norm": 0.5050505050505051, |
|
"acc_norm_stderr": 0.035621707606254015 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.35172413793103446, |
|
"acc_stderr": 0.03979236637497411, |
|
"acc_norm": 0.35172413793103446, |
|
"acc_norm_stderr": 0.03979236637497411 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.21568627450980393, |
|
"acc_stderr": 0.04092563958237655, |
|
"acc_norm": 0.21568627450980393, |
|
"acc_norm_stderr": 0.04092563958237655 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.3697478991596639, |
|
"acc_stderr": 0.031357095996135904, |
|
"acc_norm": 0.3697478991596639, |
|
"acc_norm_stderr": 0.031357095996135904 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.38974358974358975, |
|
"acc_stderr": 0.024726967886647074, |
|
"acc_norm": 0.38974358974358975, |
|
"acc_norm_stderr": 0.024726967886647074 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.52, |
|
"acc_stderr": 0.050211673156867795, |
|
"acc_norm": 0.52, |
|
"acc_norm_stderr": 0.050211673156867795 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.04824181513244218, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.04824181513244218 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5185185185185185, |
|
"acc_stderr": 0.04830366024635331, |
|
"acc_norm": 0.5185185185185185, |
|
"acc_norm_stderr": 0.04830366024635331 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.3842364532019704, |
|
"acc_stderr": 0.0342239856565755, |
|
"acc_norm": 0.3842364532019704, |
|
"acc_norm_stderr": 0.0342239856565755 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.43548387096774194, |
|
"acc_stderr": 0.028206225591502737, |
|
"acc_norm": 0.43548387096774194, |
|
"acc_norm_stderr": 0.028206225591502737 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6410256410256411, |
|
"acc_stderr": 0.03142616993791924, |
|
"acc_norm": 0.6410256410256411, |
|
"acc_norm_stderr": 0.03142616993791924 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.43018867924528303, |
|
"acc_stderr": 0.030471445867183238, |
|
"acc_norm": 0.43018867924528303, |
|
"acc_norm_stderr": 0.030471445867183238 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.44545454545454544, |
|
"acc_stderr": 0.047605488214603246, |
|
"acc_norm": 0.44545454545454544, |
|
"acc_norm_stderr": 0.047605488214603246 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.24444444444444444, |
|
"acc_stderr": 0.026202766534652148, |
|
"acc_norm": 0.24444444444444444, |
|
"acc_norm_stderr": 0.026202766534652148 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2582781456953642, |
|
"acc_stderr": 0.035737053147634576, |
|
"acc_norm": 0.2582781456953642, |
|
"acc_norm_stderr": 0.035737053147634576 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.527363184079602, |
|
"acc_stderr": 0.03530235517334682, |
|
"acc_norm": 0.527363184079602, |
|
"acc_norm_stderr": 0.03530235517334682 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.3815028901734104, |
|
"acc_stderr": 0.03703851193099522, |
|
"acc_norm": 0.3815028901734104, |
|
"acc_norm_stderr": 0.03703851193099522 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.2698412698412698, |
|
"acc_stderr": 0.022860838309232072, |
|
"acc_norm": 0.2698412698412698, |
|
"acc_norm_stderr": 0.022860838309232072 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.039420826399272135, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.039420826399272135 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252606, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252606 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.61, |
|
"acc_stderr": 0.04902071300001975, |
|
"acc_norm": 0.61, |
|
"acc_norm_stderr": 0.04902071300001975 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.48265895953757226, |
|
"acc_stderr": 0.026902900458666647, |
|
"acc_norm": 0.48265895953757226, |
|
"acc_norm_stderr": 0.026902900458666647 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.44171779141104295, |
|
"acc_stderr": 0.03901591825836184, |
|
"acc_norm": 0.44171779141104295, |
|
"acc_norm_stderr": 0.03901591825836184 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.44753086419753085, |
|
"acc_stderr": 0.027667138569422708, |
|
"acc_norm": 0.44753086419753085, |
|
"acc_norm_stderr": 0.027667138569422708 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.045604802157206845, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.045604802157206845 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.49740932642487046, |
|
"acc_stderr": 0.03608390745384486, |
|
"acc_norm": 0.49740932642487046, |
|
"acc_norm_stderr": 0.03608390745384486 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2719298245614035, |
|
"acc_stderr": 0.041857744240220575, |
|
"acc_norm": 0.2719298245614035, |
|
"acc_norm_stderr": 0.041857744240220575 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.48807339449541287, |
|
"acc_stderr": 0.021431223617362233, |
|
"acc_norm": 0.48807339449541287, |
|
"acc_norm_stderr": 0.021431223617362233 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.23809523809523808, |
|
"acc_stderr": 0.038095238095238106, |
|
"acc_norm": 0.23809523809523808, |
|
"acc_norm_stderr": 0.038095238095238106 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.4019607843137255, |
|
"acc_stderr": 0.028074158947600663, |
|
"acc_norm": 0.4019607843137255, |
|
"acc_norm_stderr": 0.028074158947600663 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.43, |
|
"acc_stderr": 0.049756985195624284, |
|
"acc_norm": 0.43, |
|
"acc_norm_stderr": 0.049756985195624284 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.5371900826446281, |
|
"acc_stderr": 0.04551711196104218, |
|
"acc_norm": 0.5371900826446281, |
|
"acc_norm_stderr": 0.04551711196104218 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.3355263157894737, |
|
"acc_stderr": 0.038424985593952694, |
|
"acc_norm": 0.3355263157894737, |
|
"acc_norm_stderr": 0.038424985593952694 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.37254901960784315, |
|
"acc_stderr": 0.019559646809215923, |
|
"acc_norm": 0.37254901960784315, |
|
"acc_norm_stderr": 0.019559646809215923 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.30141843971631205, |
|
"acc_stderr": 0.02737412888263115, |
|
"acc_norm": 0.30141843971631205, |
|
"acc_norm_stderr": 0.02737412888263115 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.25892857142857145, |
|
"acc_stderr": 0.04157751539865629, |
|
"acc_norm": 0.25892857142857145, |
|
"acc_norm_stderr": 0.04157751539865629 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.2777777777777778, |
|
"acc_stderr": 0.030546745264953195, |
|
"acc_norm": 0.2777777777777778, |
|
"acc_norm_stderr": 0.030546745264953195 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2435754189944134, |
|
"acc_stderr": 0.014355911964767857, |
|
"acc_norm": 0.2435754189944134, |
|
"acc_norm_stderr": 0.014355911964767857 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252605, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252605 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.43, |
|
"acc_stderr": 0.049756985195624284, |
|
"acc_norm": 0.43, |
|
"acc_norm_stderr": 0.049756985195624284 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.33088235294117646, |
|
"acc_stderr": 0.02858270975389843, |
|
"acc_norm": 0.33088235294117646, |
|
"acc_norm_stderr": 0.02858270975389843 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.44081632653061226, |
|
"acc_stderr": 0.03178419114175364, |
|
"acc_norm": 0.44081632653061226, |
|
"acc_norm_stderr": 0.03178419114175364 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.5569620253164557, |
|
"acc_stderr": 0.03233532777533484, |
|
"acc_norm": 0.5569620253164557, |
|
"acc_norm_stderr": 0.03233532777533484 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.32529335071707954, |
|
"acc_stderr": 0.011965311536571528, |
|
"acc_norm": 0.32529335071707954, |
|
"acc_norm_stderr": 0.011965311536571528 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.4362745098039216, |
|
"acc_stderr": 0.03480693138457039, |
|
"acc_norm": 0.4362745098039216, |
|
"acc_norm_stderr": 0.03480693138457039 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.5333333333333333, |
|
"acc_stderr": 0.03895658065271846, |
|
"acc_norm": 0.5333333333333333, |
|
"acc_norm_stderr": 0.03895658065271846 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.3084455324357405, |
|
"mc1_stderr": 0.01616803938315687, |
|
"mc2": 0.4532384559135145, |
|
"mc2_stderr": 0.015485047009493541 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.4793388429752066, |
|
"acc_stderr": 0.017175671279836446, |
|
"acc_norm": 0.5301062573789846, |
|
"acc_norm_stderr": 0.017159163590170223 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "42MARU/GenAI-llama2-ko-en-instruct-v7-13B", |
|
"model_sha": "40b2f1775ec5f92bfa8191fda6bb5f7c78564b3c", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |