|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3575085324232082, |
|
"acc_stderr": 0.014005494275916576, |
|
"acc_norm": 0.42150170648464164, |
|
"acc_norm_stderr": 0.014430197069326028 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.38279227245568614, |
|
"acc_stderr": 0.004850748687859933, |
|
"acc_norm": 0.4874526986656045, |
|
"acc_norm_stderr": 0.004988210033832016 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.5029239766081871, |
|
"acc_stderr": 0.03834759370936839, |
|
"acc_norm": 0.5029239766081871, |
|
"acc_norm_stderr": 0.03834759370936839 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5728155339805825, |
|
"acc_stderr": 0.04897957737781168, |
|
"acc_norm": 0.5728155339805825, |
|
"acc_norm_stderr": 0.04897957737781168 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.4623243933588761, |
|
"acc_stderr": 0.017829131764287198, |
|
"acc_norm": 0.4623243933588761, |
|
"acc_norm_stderr": 0.017829131764287198 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.34814814814814815, |
|
"acc_stderr": 0.041153246103369526, |
|
"acc_norm": 0.34814814814814815, |
|
"acc_norm_stderr": 0.041153246103369526 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.04461960433384741, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.04461960433384741 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.032025630761017346, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.032025630761017346 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.41566265060240964, |
|
"acc_stderr": 0.038367221765980515, |
|
"acc_norm": 0.41566265060240964, |
|
"acc_norm_stderr": 0.038367221765980515 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.47266881028938906, |
|
"acc_stderr": 0.02835563356832818, |
|
"acc_norm": 0.47266881028938906, |
|
"acc_norm_stderr": 0.02835563356832818 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.4618834080717489, |
|
"acc_stderr": 0.033460150119732274, |
|
"acc_norm": 0.4618834080717489, |
|
"acc_norm_stderr": 0.033460150119732274 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.44274809160305345, |
|
"acc_stderr": 0.04356447202665069, |
|
"acc_norm": 0.44274809160305345, |
|
"acc_norm_stderr": 0.04356447202665069 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.049236596391733084, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.049236596391733084 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5656565656565656, |
|
"acc_stderr": 0.03531505879359182, |
|
"acc_norm": 0.5656565656565656, |
|
"acc_norm_stderr": 0.03531505879359182 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.45517241379310347, |
|
"acc_stderr": 0.04149886942192117, |
|
"acc_norm": 0.45517241379310347, |
|
"acc_norm_stderr": 0.04149886942192117 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.17647058823529413, |
|
"acc_stderr": 0.03793281185307807, |
|
"acc_norm": 0.17647058823529413, |
|
"acc_norm_stderr": 0.03793281185307807 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.46638655462184875, |
|
"acc_stderr": 0.03240501447690071, |
|
"acc_norm": 0.46638655462184875, |
|
"acc_norm_stderr": 0.03240501447690071 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4282051282051282, |
|
"acc_stderr": 0.02508830145469484, |
|
"acc_norm": 0.4282051282051282, |
|
"acc_norm_stderr": 0.02508830145469484 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.7, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.7, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.04512608598542128, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.04512608598542128 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.49074074074074076, |
|
"acc_stderr": 0.04832853553437055, |
|
"acc_norm": 0.49074074074074076, |
|
"acc_norm_stderr": 0.04832853553437055 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.3793103448275862, |
|
"acc_stderr": 0.034139638059062345, |
|
"acc_norm": 0.3793103448275862, |
|
"acc_norm_stderr": 0.034139638059062345 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.432258064516129, |
|
"acc_stderr": 0.028181739720019413, |
|
"acc_norm": 0.432258064516129, |
|
"acc_norm_stderr": 0.028181739720019413 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6923076923076923, |
|
"acc_stderr": 0.030236389942173095, |
|
"acc_norm": 0.6923076923076923, |
|
"acc_norm_stderr": 0.030236389942173095 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.4679245283018868, |
|
"acc_stderr": 0.030709486992556538, |
|
"acc_norm": 0.4679245283018868, |
|
"acc_norm_stderr": 0.030709486992556538 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5363636363636364, |
|
"acc_stderr": 0.04776449162396197, |
|
"acc_norm": 0.5363636363636364, |
|
"acc_norm_stderr": 0.04776449162396197 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.2814814814814815, |
|
"acc_stderr": 0.027420019350945277, |
|
"acc_norm": 0.2814814814814815, |
|
"acc_norm_stderr": 0.027420019350945277 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2781456953642384, |
|
"acc_stderr": 0.03658603262763744, |
|
"acc_norm": 0.2781456953642384, |
|
"acc_norm_stderr": 0.03658603262763744 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6019900497512438, |
|
"acc_stderr": 0.03461199429040013, |
|
"acc_norm": 0.6019900497512438, |
|
"acc_norm_stderr": 0.03461199429040013 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.3815028901734104, |
|
"acc_stderr": 0.03703851193099521, |
|
"acc_norm": 0.3815028901734104, |
|
"acc_norm_stderr": 0.03703851193099521 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.0242785680243077, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.0242785680243077 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3263888888888889, |
|
"acc_stderr": 0.03921067198982266, |
|
"acc_norm": 0.3263888888888889, |
|
"acc_norm_stderr": 0.03921067198982266 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.0479372485441102, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.0479372485441102 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.6, |
|
"acc_stderr": 0.049236596391733084, |
|
"acc_norm": 0.6, |
|
"acc_norm_stderr": 0.049236596391733084 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.4913294797687861, |
|
"acc_stderr": 0.026915047355369818, |
|
"acc_norm": 0.4913294797687861, |
|
"acc_norm_stderr": 0.026915047355369818 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.4601226993865031, |
|
"acc_stderr": 0.03915857291436971, |
|
"acc_norm": 0.4601226993865031, |
|
"acc_norm_stderr": 0.03915857291436971 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.43209876543209874, |
|
"acc_stderr": 0.02756301097160668, |
|
"acc_norm": 0.43209876543209874, |
|
"acc_norm_stderr": 0.02756301097160668 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.5181347150259067, |
|
"acc_stderr": 0.036060650018329185, |
|
"acc_norm": 0.5181347150259067, |
|
"acc_norm_stderr": 0.036060650018329185 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.3157894736842105, |
|
"acc_stderr": 0.043727482902780085, |
|
"acc_norm": 0.3157894736842105, |
|
"acc_norm_stderr": 0.043727482902780085 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.47522935779816516, |
|
"acc_stderr": 0.02141099975363592, |
|
"acc_norm": 0.47522935779816516, |
|
"acc_norm_stderr": 0.02141099975363592 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.35714285714285715, |
|
"acc_stderr": 0.04285714285714281, |
|
"acc_norm": 0.35714285714285715, |
|
"acc_norm_stderr": 0.04285714285714281 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.45098039215686275, |
|
"acc_stderr": 0.028491993586171573, |
|
"acc_norm": 0.45098039215686275, |
|
"acc_norm_stderr": 0.028491993586171573 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.050251890762960605, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.050251890762960605 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6528925619834711, |
|
"acc_stderr": 0.04345724570292534, |
|
"acc_norm": 0.6528925619834711, |
|
"acc_norm_stderr": 0.04345724570292534 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.39473684210526316, |
|
"acc_stderr": 0.03977749934622074, |
|
"acc_norm": 0.39473684210526316, |
|
"acc_norm_stderr": 0.03977749934622074 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.3872549019607843, |
|
"acc_stderr": 0.019706875804085627, |
|
"acc_norm": 0.3872549019607843, |
|
"acc_norm_stderr": 0.019706875804085627 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.36524822695035464, |
|
"acc_stderr": 0.028723863853281278, |
|
"acc_norm": 0.36524822695035464, |
|
"acc_norm_stderr": 0.028723863853281278 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.39285714285714285, |
|
"acc_stderr": 0.04635550135609976, |
|
"acc_norm": 0.39285714285714285, |
|
"acc_norm_stderr": 0.04635550135609976 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.39351851851851855, |
|
"acc_stderr": 0.03331747876370312, |
|
"acc_norm": 0.39351851851851855, |
|
"acc_norm_stderr": 0.03331747876370312 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.21564245810055865, |
|
"acc_stderr": 0.013754835975482355, |
|
"acc_norm": 0.21564245810055865, |
|
"acc_norm_stderr": 0.013754835975482355 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.58, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.58, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.36764705882352944, |
|
"acc_stderr": 0.029289413409403192, |
|
"acc_norm": 0.36764705882352944, |
|
"acc_norm_stderr": 0.029289413409403192 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.43673469387755104, |
|
"acc_stderr": 0.03175195237583322, |
|
"acc_norm": 0.43673469387755104, |
|
"acc_norm_stderr": 0.03175195237583322 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.5907172995780591, |
|
"acc_stderr": 0.03200704183359591, |
|
"acc_norm": 0.5907172995780591, |
|
"acc_norm_stderr": 0.03200704183359591 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.31747066492829207, |
|
"acc_stderr": 0.01188889206880931, |
|
"acc_norm": 0.31747066492829207, |
|
"acc_norm_stderr": 0.01188889206880931 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.5098039215686274, |
|
"acc_stderr": 0.03508637358630572, |
|
"acc_norm": 0.5098039215686274, |
|
"acc_norm_stderr": 0.03508637358630572 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.4666666666666667, |
|
"acc_stderr": 0.03895658065271846, |
|
"acc_norm": 0.4666666666666667, |
|
"acc_norm_stderr": 0.03895658065271846 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2876376988984088, |
|
"mc1_stderr": 0.015846315101394823, |
|
"mc2": 0.47510378175366297, |
|
"mc2_stderr": 0.015686785961170725 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.4297520661157025, |
|
"acc_stderr": 0.017019847535972205, |
|
"acc_norm": 0.48642266824085006, |
|
"acc_norm_stderr": 0.017184015060401448 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "MNCJ1hun/MIstral-11B-Omni-OP-u1k-ver0.1", |
|
"model_sha": "3cf7eb4c014f181bec2a9b36897771b2710422d1", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |