|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3532423208191126, |
|
"acc_stderr": 0.013967822714840053, |
|
"acc_norm": 0.3984641638225256, |
|
"acc_norm_stderr": 0.014306946052735565 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3676558454491137, |
|
"acc_stderr": 0.004811815959388833, |
|
"acc_norm": 0.47410874327823144, |
|
"acc_norm_stderr": 0.004983087049281746 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.4502923976608187, |
|
"acc_stderr": 0.038158273659132366, |
|
"acc_norm": 0.4502923976608187, |
|
"acc_norm_stderr": 0.038158273659132366 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5922330097087378, |
|
"acc_stderr": 0.048657775704107696, |
|
"acc_norm": 0.5922330097087378, |
|
"acc_norm_stderr": 0.048657775704107696 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5108556832694764, |
|
"acc_stderr": 0.017875748840242407, |
|
"acc_norm": 0.5108556832694764, |
|
"acc_norm_stderr": 0.017875748840242407 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.3925925925925926, |
|
"acc_stderr": 0.04218506215368879, |
|
"acc_norm": 0.3925925925925926, |
|
"acc_norm_stderr": 0.04218506215368879 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621503, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621503 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.3659574468085106, |
|
"acc_stderr": 0.0314895582974553, |
|
"acc_norm": 0.3659574468085106, |
|
"acc_norm_stderr": 0.0314895582974553 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.35542168674698793, |
|
"acc_stderr": 0.03726214354322415, |
|
"acc_norm": 0.35542168674698793, |
|
"acc_norm_stderr": 0.03726214354322415 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.4662379421221865, |
|
"acc_stderr": 0.02833327710956279, |
|
"acc_norm": 0.4662379421221865, |
|
"acc_norm_stderr": 0.02833327710956279 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.45739910313901344, |
|
"acc_stderr": 0.033435777055830646, |
|
"acc_norm": 0.45739910313901344, |
|
"acc_norm_stderr": 0.033435777055830646 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.48091603053435117, |
|
"acc_stderr": 0.04382094705550989, |
|
"acc_norm": 0.48091603053435117, |
|
"acc_norm_stderr": 0.04382094705550989 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5202020202020202, |
|
"acc_stderr": 0.03559443565563919, |
|
"acc_norm": 0.5202020202020202, |
|
"acc_norm_stderr": 0.03559443565563919 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.41379310344827586, |
|
"acc_stderr": 0.04104269211806232, |
|
"acc_norm": 0.41379310344827586, |
|
"acc_norm_stderr": 0.04104269211806232 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.18627450980392157, |
|
"acc_stderr": 0.03873958714149352, |
|
"acc_norm": 0.18627450980392157, |
|
"acc_norm_stderr": 0.03873958714149352 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.44537815126050423, |
|
"acc_stderr": 0.0322841062671639, |
|
"acc_norm": 0.44537815126050423, |
|
"acc_norm_stderr": 0.0322841062671639 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4358974358974359, |
|
"acc_stderr": 0.025141801511177495, |
|
"acc_norm": 0.4358974358974359, |
|
"acc_norm_stderr": 0.025141801511177495 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.43, |
|
"acc_stderr": 0.04975698519562428, |
|
"acc_norm": 0.43, |
|
"acc_norm_stderr": 0.04975698519562428 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.049236596391733084, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.049236596391733084 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5277777777777778, |
|
"acc_stderr": 0.04826217294139894, |
|
"acc_norm": 0.5277777777777778, |
|
"acc_norm_stderr": 0.04826217294139894 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.33497536945812806, |
|
"acc_stderr": 0.033208527423483104, |
|
"acc_norm": 0.33497536945812806, |
|
"acc_norm_stderr": 0.033208527423483104 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.4645161290322581, |
|
"acc_stderr": 0.02837228779796295, |
|
"acc_norm": 0.4645161290322581, |
|
"acc_norm_stderr": 0.02837228779796295 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6495726495726496, |
|
"acc_stderr": 0.0312561082442188, |
|
"acc_norm": 0.6495726495726496, |
|
"acc_norm_stderr": 0.0312561082442188 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.44150943396226416, |
|
"acc_stderr": 0.030561590426731833, |
|
"acc_norm": 0.44150943396226416, |
|
"acc_norm_stderr": 0.030561590426731833 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.4727272727272727, |
|
"acc_stderr": 0.04782001791380063, |
|
"acc_norm": 0.4727272727272727, |
|
"acc_norm_stderr": 0.04782001791380063 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.2962962962962963, |
|
"acc_stderr": 0.027840811495871913, |
|
"acc_norm": 0.2962962962962963, |
|
"acc_norm_stderr": 0.027840811495871913 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2781456953642384, |
|
"acc_stderr": 0.03658603262763743, |
|
"acc_norm": 0.2781456953642384, |
|
"acc_norm_stderr": 0.03658603262763743 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.572139303482587, |
|
"acc_stderr": 0.03498541988407795, |
|
"acc_norm": 0.572139303482587, |
|
"acc_norm_stderr": 0.03498541988407795 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.3930635838150289, |
|
"acc_stderr": 0.03724249595817731, |
|
"acc_norm": 0.3930635838150289, |
|
"acc_norm_stderr": 0.03724249595817731 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.2857142857142857, |
|
"acc_stderr": 0.02326651221373057, |
|
"acc_norm": 0.2857142857142857, |
|
"acc_norm_stderr": 0.02326651221373057 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3541666666666667, |
|
"acc_stderr": 0.039994111357535424, |
|
"acc_norm": 0.3541666666666667, |
|
"acc_norm_stderr": 0.039994111357535424 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.59, |
|
"acc_stderr": 0.04943110704237102, |
|
"acc_norm": 0.59, |
|
"acc_norm_stderr": 0.04943110704237102 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5202312138728323, |
|
"acc_stderr": 0.02689704999638286, |
|
"acc_norm": 0.5202312138728323, |
|
"acc_norm_stderr": 0.02689704999638286 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.4049079754601227, |
|
"acc_stderr": 0.03856672163548913, |
|
"acc_norm": 0.4049079754601227, |
|
"acc_norm_stderr": 0.03856672163548913 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.4444444444444444, |
|
"acc_stderr": 0.027648477877413324, |
|
"acc_norm": 0.4444444444444444, |
|
"acc_norm_stderr": 0.027648477877413324 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.04824181513244218, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.04824181513244218 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.5129533678756477, |
|
"acc_stderr": 0.03607228061047749, |
|
"acc_norm": 0.5129533678756477, |
|
"acc_norm_stderr": 0.03607228061047749 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.23684210526315788, |
|
"acc_stderr": 0.03999423879281338, |
|
"acc_norm": 0.23684210526315788, |
|
"acc_norm_stderr": 0.03999423879281338 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.47706422018348627, |
|
"acc_stderr": 0.0214147570581755, |
|
"acc_norm": 0.47706422018348627, |
|
"acc_norm_stderr": 0.0214147570581755 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.36507936507936506, |
|
"acc_stderr": 0.04306241259127153, |
|
"acc_norm": 0.36507936507936506, |
|
"acc_norm_stderr": 0.04306241259127153 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.434640522875817, |
|
"acc_stderr": 0.028384256704883037, |
|
"acc_norm": 0.434640522875817, |
|
"acc_norm_stderr": 0.028384256704883037 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6611570247933884, |
|
"acc_stderr": 0.043207678075366705, |
|
"acc_norm": 0.6611570247933884, |
|
"acc_norm_stderr": 0.043207678075366705 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.4144736842105263, |
|
"acc_stderr": 0.04008973785779206, |
|
"acc_norm": 0.4144736842105263, |
|
"acc_norm_stderr": 0.04008973785779206 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.3562091503267974, |
|
"acc_stderr": 0.0193733324207245, |
|
"acc_norm": 0.3562091503267974, |
|
"acc_norm_stderr": 0.0193733324207245 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.02812163604063989, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.02812163604063989 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.29464285714285715, |
|
"acc_stderr": 0.043270409325787296, |
|
"acc_norm": 0.29464285714285715, |
|
"acc_norm_stderr": 0.043270409325787296 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.32407407407407407, |
|
"acc_stderr": 0.03191923445686186, |
|
"acc_norm": 0.32407407407407407, |
|
"acc_norm_stderr": 0.03191923445686186 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2581005586592179, |
|
"acc_stderr": 0.01463518561652782, |
|
"acc_norm": 0.2581005586592179, |
|
"acc_norm_stderr": 0.01463518561652782 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.04878317312145632, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.04878317312145632 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.45, |
|
"acc_stderr": 0.05, |
|
"acc_norm": 0.45, |
|
"acc_norm_stderr": 0.05 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.34558823529411764, |
|
"acc_stderr": 0.028888193103988644, |
|
"acc_norm": 0.34558823529411764, |
|
"acc_norm_stderr": 0.028888193103988644 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.5387755102040817, |
|
"acc_stderr": 0.03191282052669278, |
|
"acc_norm": 0.5387755102040817, |
|
"acc_norm_stderr": 0.03191282052669278 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.5569620253164557, |
|
"acc_stderr": 0.03233532777533484, |
|
"acc_norm": 0.5569620253164557, |
|
"acc_norm_stderr": 0.03233532777533484 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3494132985658409, |
|
"acc_stderr": 0.012177306252786702, |
|
"acc_norm": 0.3494132985658409, |
|
"acc_norm_stderr": 0.012177306252786702 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.4852941176470588, |
|
"acc_stderr": 0.03507793834791324, |
|
"acc_norm": 0.4852941176470588, |
|
"acc_norm_stderr": 0.03507793834791324 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.503030303030303, |
|
"acc_stderr": 0.03904272341431856, |
|
"acc_norm": 0.503030303030303, |
|
"acc_norm_stderr": 0.03904272341431856 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2962056303549572, |
|
"mc1_stderr": 0.015983595101811392, |
|
"mc2": 0.4656302357866705, |
|
"mc2_stderr": 0.01570298909368207 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.43683589138134593, |
|
"acc_stderr": 0.017052633559856076, |
|
"acc_norm": 0.45808736717827625, |
|
"acc_norm_stderr": 0.017129852117911144 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "HwiyeolJo/TeamJaeCorpo", |
|
"model_sha": "32816adb88f45d573f744311867caa7ec94b8e14", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |