|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.4564846416382253, |
|
"acc_stderr": 0.014555949760496439, |
|
"acc_norm": 0.5068259385665529, |
|
"acc_norm_stderr": 0.014610029151379813 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.46126269667396935, |
|
"acc_stderr": 0.004974783753309692, |
|
"acc_norm": 0.6171081457876917, |
|
"acc_norm_stderr": 0.0048509882151675345 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.6257309941520468, |
|
"acc_stderr": 0.03711601185389481, |
|
"acc_norm": 0.6257309941520468, |
|
"acc_norm_stderr": 0.03711601185389481 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6019417475728155, |
|
"acc_stderr": 0.04846748253977238, |
|
"acc_norm": 0.6019417475728155, |
|
"acc_norm_stderr": 0.04846748253977238 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.6462324393358876, |
|
"acc_stderr": 0.017098184708161903, |
|
"acc_norm": 0.6462324393358876, |
|
"acc_norm_stderr": 0.017098184708161903 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.5259259259259259, |
|
"acc_stderr": 0.043135316967505756, |
|
"acc_norm": 0.5259259259259259, |
|
"acc_norm_stderr": 0.043135316967505756 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.0446196043338474, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.0446196043338474 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.48936170212765956, |
|
"acc_stderr": 0.03267862331014063, |
|
"acc_norm": 0.48936170212765956, |
|
"acc_norm_stderr": 0.03267862331014063 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.463855421686747, |
|
"acc_stderr": 0.03882310850890594, |
|
"acc_norm": 0.463855421686747, |
|
"acc_norm_stderr": 0.03882310850890594 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.617363344051447, |
|
"acc_stderr": 0.027604689028581993, |
|
"acc_norm": 0.617363344051447, |
|
"acc_norm_stderr": 0.027604689028581993 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5695067264573991, |
|
"acc_stderr": 0.033231973029429394, |
|
"acc_norm": 0.5695067264573991, |
|
"acc_norm_stderr": 0.033231973029429394 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.648854961832061, |
|
"acc_stderr": 0.0418644516301375, |
|
"acc_norm": 0.648854961832061, |
|
"acc_norm_stderr": 0.0418644516301375 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.48, |
|
"acc_stderr": 0.05021167315686779, |
|
"acc_norm": 0.48, |
|
"acc_norm_stderr": 0.05021167315686779 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.7323232323232324, |
|
"acc_stderr": 0.03154449888270286, |
|
"acc_norm": 0.7323232323232324, |
|
"acc_norm_stderr": 0.03154449888270286 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.5310344827586206, |
|
"acc_stderr": 0.04158632762097828, |
|
"acc_norm": 0.5310344827586206, |
|
"acc_norm_stderr": 0.04158632762097828 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.22549019607843138, |
|
"acc_stderr": 0.04158307533083286, |
|
"acc_norm": 0.22549019607843138, |
|
"acc_norm_stderr": 0.04158307533083286 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.6176470588235294, |
|
"acc_stderr": 0.031566630992154156, |
|
"acc_norm": 0.6176470588235294, |
|
"acc_norm_stderr": 0.031566630992154156 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.5282051282051282, |
|
"acc_stderr": 0.025310639254933855, |
|
"acc_norm": 0.5282051282051282, |
|
"acc_norm_stderr": 0.025310639254933855 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.58, |
|
"acc_stderr": 0.04960449637488583, |
|
"acc_norm": 0.58, |
|
"acc_norm_stderr": 0.04960449637488583 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.048241815132442176, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.048241815132442176 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6574074074074074, |
|
"acc_stderr": 0.045879047413018105, |
|
"acc_norm": 0.6574074074074074, |
|
"acc_norm_stderr": 0.045879047413018105 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.42857142857142855, |
|
"acc_stderr": 0.034819048444388045, |
|
"acc_norm": 0.42857142857142855, |
|
"acc_norm_stderr": 0.034819048444388045 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.5774193548387097, |
|
"acc_stderr": 0.02810096472427264, |
|
"acc_norm": 0.5774193548387097, |
|
"acc_norm_stderr": 0.02810096472427264 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.811965811965812, |
|
"acc_stderr": 0.025598193686652247, |
|
"acc_norm": 0.811965811965812, |
|
"acc_norm_stderr": 0.025598193686652247 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.5207547169811321, |
|
"acc_stderr": 0.030746349975723463, |
|
"acc_norm": 0.5207547169811321, |
|
"acc_norm_stderr": 0.030746349975723463 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5909090909090909, |
|
"acc_stderr": 0.04709306978661895, |
|
"acc_norm": 0.5909090909090909, |
|
"acc_norm_stderr": 0.04709306978661895 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.362962962962963, |
|
"acc_stderr": 0.029318203645206865, |
|
"acc_norm": 0.362962962962963, |
|
"acc_norm_stderr": 0.029318203645206865 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3509933774834437, |
|
"acc_stderr": 0.03896981964257375, |
|
"acc_norm": 0.3509933774834437, |
|
"acc_norm_stderr": 0.03896981964257375 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.746268656716418, |
|
"acc_stderr": 0.030769444967296007, |
|
"acc_norm": 0.746268656716418, |
|
"acc_norm_stderr": 0.030769444967296007 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.49710982658959535, |
|
"acc_stderr": 0.038124005659748335, |
|
"acc_norm": 0.49710982658959535, |
|
"acc_norm_stderr": 0.038124005659748335 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.373015873015873, |
|
"acc_stderr": 0.02490699045899257, |
|
"acc_norm": 0.373015873015873, |
|
"acc_norm_stderr": 0.02490699045899257 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4791666666666667, |
|
"acc_stderr": 0.041775789507399935, |
|
"acc_norm": 0.4791666666666667, |
|
"acc_norm_stderr": 0.041775789507399935 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621505, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621505 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.75, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.75, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.6098265895953757, |
|
"acc_stderr": 0.026261677607806642, |
|
"acc_norm": 0.6098265895953757, |
|
"acc_norm_stderr": 0.026261677607806642 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.558282208588957, |
|
"acc_stderr": 0.03901591825836184, |
|
"acc_norm": 0.558282208588957, |
|
"acc_norm_stderr": 0.03901591825836184 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.6049382716049383, |
|
"acc_stderr": 0.027201117666925654, |
|
"acc_norm": 0.6049382716049383, |
|
"acc_norm_stderr": 0.027201117666925654 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.43, |
|
"acc_stderr": 0.049756985195624284, |
|
"acc_norm": 0.43, |
|
"acc_norm_stderr": 0.049756985195624284 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.7461139896373057, |
|
"acc_stderr": 0.03141024780565319, |
|
"acc_norm": 0.7461139896373057, |
|
"acc_norm_stderr": 0.03141024780565319 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.45614035087719296, |
|
"acc_stderr": 0.046854730419077895, |
|
"acc_norm": 0.45614035087719296, |
|
"acc_norm_stderr": 0.046854730419077895 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.6697247706422018, |
|
"acc_stderr": 0.020164466336342977, |
|
"acc_norm": 0.6697247706422018, |
|
"acc_norm_stderr": 0.020164466336342977 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.3888888888888889, |
|
"acc_stderr": 0.0436031486007746, |
|
"acc_norm": 0.3888888888888889, |
|
"acc_norm_stderr": 0.0436031486007746 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5947712418300654, |
|
"acc_stderr": 0.028110928492809075, |
|
"acc_norm": 0.5947712418300654, |
|
"acc_norm_stderr": 0.028110928492809075 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.61, |
|
"acc_stderr": 0.049020713000019756, |
|
"acc_norm": 0.61, |
|
"acc_norm_stderr": 0.049020713000019756 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.7272727272727273, |
|
"acc_stderr": 0.04065578140908705, |
|
"acc_norm": 0.7272727272727273, |
|
"acc_norm_stderr": 0.04065578140908705 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5855263157894737, |
|
"acc_stderr": 0.04008973785779205, |
|
"acc_norm": 0.5855263157894737, |
|
"acc_norm_stderr": 0.04008973785779205 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.5147058823529411, |
|
"acc_stderr": 0.020219083895133917, |
|
"acc_norm": 0.5147058823529411, |
|
"acc_norm_stderr": 0.020219083895133917 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.37943262411347517, |
|
"acc_stderr": 0.028947338851614105, |
|
"acc_norm": 0.37943262411347517, |
|
"acc_norm_stderr": 0.028947338851614105 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.375, |
|
"acc_stderr": 0.04595091388086298, |
|
"acc_norm": 0.375, |
|
"acc_norm_stderr": 0.04595091388086298 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.5092592592592593, |
|
"acc_stderr": 0.034093869469927006, |
|
"acc_norm": 0.5092592592592593, |
|
"acc_norm_stderr": 0.034093869469927006 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.22793296089385476, |
|
"acc_stderr": 0.014030149950805097, |
|
"acc_norm": 0.22793296089385476, |
|
"acc_norm_stderr": 0.014030149950805097 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.46, |
|
"acc_stderr": 0.05009082659620332, |
|
"acc_norm": 0.46, |
|
"acc_norm_stderr": 0.05009082659620332 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.64, |
|
"acc_stderr": 0.048241815132442176, |
|
"acc_norm": 0.64, |
|
"acc_norm_stderr": 0.048241815132442176 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.030372836961539352, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.030372836961539352 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.5755102040816327, |
|
"acc_stderr": 0.03164209487942942, |
|
"acc_norm": 0.5755102040816327, |
|
"acc_norm_stderr": 0.03164209487942942 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.7341772151898734, |
|
"acc_stderr": 0.02875679962965834, |
|
"acc_norm": 0.7341772151898734, |
|
"acc_norm_stderr": 0.02875679962965834 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.4041720990873533, |
|
"acc_stderr": 0.012533504046491367, |
|
"acc_norm": 0.4041720990873533, |
|
"acc_norm_stderr": 0.012533504046491367 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.6764705882352942, |
|
"acc_stderr": 0.0328347205610856, |
|
"acc_norm": 0.6764705882352942, |
|
"acc_norm_stderr": 0.0328347205610856 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.6848484848484848, |
|
"acc_stderr": 0.0362773057502241, |
|
"acc_norm": 0.6848484848484848, |
|
"acc_norm_stderr": 0.0362773057502241 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.31211750305997554, |
|
"mc1_stderr": 0.01622075676952091, |
|
"mc2": 0.47693449176935354, |
|
"mc2_stderr": 0.015381480297496484 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5041322314049587, |
|
"acc_stderr": 0.017189767032130817, |
|
"acc_norm": 0.5454545454545454, |
|
"acc_norm_stderr": 0.017119172208061504 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "JaeyeonKang/CCK-v1.1.0", |
|
"model_sha": "d0485d83b3f6d4738e239f9d15eb3edab9583770", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |