|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3993174061433447, |
|
"acc_stderr": 0.014312094557946704, |
|
"acc_norm": 0.46501706484641636, |
|
"acc_norm_stderr": 0.01457558392201967 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.4229237203744274, |
|
"acc_stderr": 0.004930138842768219, |
|
"acc_norm": 0.5690101573391755, |
|
"acc_norm_stderr": 0.004942026200279584 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.4678362573099415, |
|
"acc_stderr": 0.03826882417660369, |
|
"acc_norm": 0.4678362573099415, |
|
"acc_norm_stderr": 0.03826882417660369 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5339805825242718, |
|
"acc_stderr": 0.0493929144727348, |
|
"acc_norm": 0.5339805825242718, |
|
"acc_norm_stderr": 0.0493929144727348 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5606641123882503, |
|
"acc_stderr": 0.017747874245683602, |
|
"acc_norm": 0.5606641123882503, |
|
"acc_norm_stderr": 0.017747874245683602 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4666666666666667, |
|
"acc_stderr": 0.04309732901036354, |
|
"acc_norm": 0.4666666666666667, |
|
"acc_norm_stderr": 0.04309732901036354 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.045604802157206824, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.045604802157206824 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.4297872340425532, |
|
"acc_stderr": 0.03236214467715564, |
|
"acc_norm": 0.4297872340425532, |
|
"acc_norm_stderr": 0.03236214467715564 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.42168674698795183, |
|
"acc_stderr": 0.038444531817709175, |
|
"acc_norm": 0.42168674698795183, |
|
"acc_norm_stderr": 0.038444531817709175 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.4855305466237942, |
|
"acc_stderr": 0.02838619808417768, |
|
"acc_norm": 0.4855305466237942, |
|
"acc_norm_stderr": 0.02838619808417768 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5381165919282511, |
|
"acc_stderr": 0.033460150119732274, |
|
"acc_norm": 0.5381165919282511, |
|
"acc_norm_stderr": 0.033460150119732274 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.4961832061068702, |
|
"acc_stderr": 0.043851623256015534, |
|
"acc_norm": 0.4961832061068702, |
|
"acc_norm_stderr": 0.043851623256015534 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.04960449637488583, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.04960449637488583 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5757575757575758, |
|
"acc_stderr": 0.03521224908841586, |
|
"acc_norm": 0.5757575757575758, |
|
"acc_norm_stderr": 0.03521224908841586 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.42758620689655175, |
|
"acc_stderr": 0.041227371113703316, |
|
"acc_norm": 0.42758620689655175, |
|
"acc_norm_stderr": 0.041227371113703316 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.2647058823529412, |
|
"acc_stderr": 0.04389869956808778, |
|
"acc_norm": 0.2647058823529412, |
|
"acc_norm_stderr": 0.04389869956808778 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.47478991596638653, |
|
"acc_stderr": 0.03243718055137411, |
|
"acc_norm": 0.47478991596638653, |
|
"acc_norm_stderr": 0.03243718055137411 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4564102564102564, |
|
"acc_stderr": 0.02525448542479961, |
|
"acc_norm": 0.4564102564102564, |
|
"acc_norm_stderr": 0.02525448542479961 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.050251890762960605, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.050251890762960605 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.048523658709391, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.048523658709391 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5277777777777778, |
|
"acc_stderr": 0.04826217294139894, |
|
"acc_norm": 0.5277777777777778, |
|
"acc_norm_stderr": 0.04826217294139894 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.43842364532019706, |
|
"acc_stderr": 0.03491207857486518, |
|
"acc_norm": 0.43842364532019706, |
|
"acc_norm_stderr": 0.03491207857486518 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.4838709677419355, |
|
"acc_stderr": 0.028429203176724555, |
|
"acc_norm": 0.4838709677419355, |
|
"acc_norm_stderr": 0.028429203176724555 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6282051282051282, |
|
"acc_stderr": 0.031660988918880785, |
|
"acc_norm": 0.6282051282051282, |
|
"acc_norm_stderr": 0.031660988918880785 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.4528301886792453, |
|
"acc_stderr": 0.03063562795796182, |
|
"acc_norm": 0.4528301886792453, |
|
"acc_norm_stderr": 0.03063562795796182 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5181818181818182, |
|
"acc_stderr": 0.04785964010794916, |
|
"acc_norm": 0.5181818181818182, |
|
"acc_norm_stderr": 0.04785964010794916 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.25925925925925924, |
|
"acc_stderr": 0.026719240783712166, |
|
"acc_norm": 0.25925925925925924, |
|
"acc_norm_stderr": 0.026719240783712166 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2847682119205298, |
|
"acc_stderr": 0.03684881521389024, |
|
"acc_norm": 0.2847682119205298, |
|
"acc_norm_stderr": 0.03684881521389024 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.5920398009950248, |
|
"acc_stderr": 0.03475116365194092, |
|
"acc_norm": 0.5920398009950248, |
|
"acc_norm_stderr": 0.03475116365194092 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.4161849710982659, |
|
"acc_stderr": 0.037585177754049466, |
|
"acc_norm": 0.4161849710982659, |
|
"acc_norm_stderr": 0.037585177754049466 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.2724867724867725, |
|
"acc_stderr": 0.02293097307163335, |
|
"acc_norm": 0.2724867724867725, |
|
"acc_norm_stderr": 0.02293097307163335 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3819444444444444, |
|
"acc_stderr": 0.040629907841466674, |
|
"acc_norm": 0.3819444444444444, |
|
"acc_norm_stderr": 0.040629907841466674 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.61, |
|
"acc_stderr": 0.04902071300001975, |
|
"acc_norm": 0.61, |
|
"acc_norm_stderr": 0.04902071300001975 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5028901734104047, |
|
"acc_stderr": 0.02691864538323901, |
|
"acc_norm": 0.5028901734104047, |
|
"acc_norm_stderr": 0.02691864538323901 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.5276073619631901, |
|
"acc_stderr": 0.0392237829061099, |
|
"acc_norm": 0.5276073619631901, |
|
"acc_norm_stderr": 0.0392237829061099 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.4783950617283951, |
|
"acc_stderr": 0.027794760105008736, |
|
"acc_norm": 0.4783950617283951, |
|
"acc_norm_stderr": 0.027794760105008736 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.5492227979274611, |
|
"acc_stderr": 0.035909109522355244, |
|
"acc_norm": 0.5492227979274611, |
|
"acc_norm_stderr": 0.035909109522355244 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2894736842105263, |
|
"acc_stderr": 0.04266339443159395, |
|
"acc_norm": 0.2894736842105263, |
|
"acc_norm_stderr": 0.04266339443159395 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.5743119266055046, |
|
"acc_stderr": 0.021199235972470795, |
|
"acc_norm": 0.5743119266055046, |
|
"acc_norm_stderr": 0.021199235972470795 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.30952380952380953, |
|
"acc_stderr": 0.04134913018303316, |
|
"acc_norm": 0.30952380952380953, |
|
"acc_norm_stderr": 0.04134913018303316 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.40522875816993464, |
|
"acc_stderr": 0.02811092849280908, |
|
"acc_norm": 0.40522875816993464, |
|
"acc_norm_stderr": 0.02811092849280908 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.049236596391733084, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.049236596391733084 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6363636363636364, |
|
"acc_stderr": 0.043913262867240704, |
|
"acc_norm": 0.6363636363636364, |
|
"acc_norm_stderr": 0.043913262867240704 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.40131578947368424, |
|
"acc_stderr": 0.03988903703336284, |
|
"acc_norm": 0.40131578947368424, |
|
"acc_norm_stderr": 0.03988903703336284 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.38562091503267976, |
|
"acc_stderr": 0.01969145905235416, |
|
"acc_norm": 0.38562091503267976, |
|
"acc_norm_stderr": 0.01969145905235416 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.32269503546099293, |
|
"acc_stderr": 0.027889139300534792, |
|
"acc_norm": 0.32269503546099293, |
|
"acc_norm_stderr": 0.027889139300534792 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.1875, |
|
"acc_stderr": 0.0370468111477387, |
|
"acc_norm": 0.1875, |
|
"acc_norm_stderr": 0.0370468111477387 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.3425925925925926, |
|
"acc_stderr": 0.03236585252602157, |
|
"acc_norm": 0.3425925925925926, |
|
"acc_norm_stderr": 0.03236585252602157 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2424581005586592, |
|
"acc_stderr": 0.01433352205921789, |
|
"acc_norm": 0.2424581005586592, |
|
"acc_norm_stderr": 0.01433352205921789 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.34, |
|
"acc_stderr": 0.04760952285695235, |
|
"acc_norm": 0.34, |
|
"acc_norm_stderr": 0.04760952285695235 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.44485294117647056, |
|
"acc_stderr": 0.03018753206032938, |
|
"acc_norm": 0.44485294117647056, |
|
"acc_norm_stderr": 0.03018753206032938 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.5142857142857142, |
|
"acc_stderr": 0.03199615232806286, |
|
"acc_norm": 0.5142857142857142, |
|
"acc_norm_stderr": 0.03199615232806286 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.6286919831223629, |
|
"acc_stderr": 0.03145068600744858, |
|
"acc_norm": 0.6286919831223629, |
|
"acc_norm_stderr": 0.03145068600744858 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3474576271186441, |
|
"acc_stderr": 0.012161417729749806, |
|
"acc_norm": 0.3474576271186441, |
|
"acc_norm_stderr": 0.012161417729749806 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.49019607843137253, |
|
"acc_stderr": 0.03508637358630573, |
|
"acc_norm": 0.49019607843137253, |
|
"acc_norm_stderr": 0.03508637358630573 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.5393939393939394, |
|
"acc_stderr": 0.03892207016552012, |
|
"acc_norm": 0.5393939393939394, |
|
"acc_norm_stderr": 0.03892207016552012 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2558139534883721, |
|
"mc1_stderr": 0.015274176219283344, |
|
"mc2": 0.4199929776899167, |
|
"mc2_stderr": 0.014679195459056854 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.4923258559622196, |
|
"acc_stderr": 0.017188329219654276, |
|
"acc_norm": 0.5844155844155844, |
|
"acc_norm_stderr": 0.016943586313076565 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "daekeun-ml/Llama-2-ko-instruct-13B", |
|
"model_sha": "a29fb540227b3cbc88a308cc5ed62c26b28d84f1", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |