open-ko-llm-bot's picture
Add results for 2023-11-29 11:13:19
d9ff906
raw
history blame
No virus
17.9 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.371160409556314,
"acc_stderr": 0.014117971901142818,
"acc_norm": 0.4496587030716723,
"acc_norm_stderr": 0.014537144444284738
},
"harness|ko_hellaswag|10": {
"acc": 0.41884086835291773,
"acc_stderr": 0.004923609207861539,
"acc_norm": 0.5671181039633539,
"acc_norm_stderr": 0.004944620712318274
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.47953216374269003,
"acc_stderr": 0.038316105328219316,
"acc_norm": 0.47953216374269003,
"acc_norm_stderr": 0.038316105328219316
},
"harness|ko_mmlu_management|5": {
"acc": 0.4077669902912621,
"acc_stderr": 0.048657775704107696,
"acc_norm": 0.4077669902912621,
"acc_norm_stderr": 0.048657775704107696
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.5095785440613027,
"acc_stderr": 0.01787668227534085,
"acc_norm": 0.5095785440613027,
"acc_norm_stderr": 0.01787668227534085
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.4666666666666667,
"acc_stderr": 0.043097329010363554,
"acc_norm": 0.4666666666666667,
"acc_norm_stderr": 0.043097329010363554
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.37446808510638296,
"acc_stderr": 0.031639106653672915,
"acc_norm": 0.37446808510638296,
"acc_norm_stderr": 0.031639106653672915
},
"harness|ko_mmlu_virology|5": {
"acc": 0.39759036144578314,
"acc_stderr": 0.038099730845402184,
"acc_norm": 0.39759036144578314,
"acc_norm_stderr": 0.038099730845402184
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.4565916398713826,
"acc_stderr": 0.028290869054197604,
"acc_norm": 0.4565916398713826,
"acc_norm_stderr": 0.028290869054197604
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.5112107623318386,
"acc_stderr": 0.033549366530984746,
"acc_norm": 0.5112107623318386,
"acc_norm_stderr": 0.033549366530984746
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.4198473282442748,
"acc_stderr": 0.04328577215262971,
"acc_norm": 0.4198473282442748,
"acc_norm_stderr": 0.04328577215262971
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252605,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252605
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.4898989898989899,
"acc_stderr": 0.035616254886737454,
"acc_norm": 0.4898989898989899,
"acc_norm_stderr": 0.035616254886737454
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.4,
"acc_stderr": 0.04082482904638628,
"acc_norm": 0.4,
"acc_norm_stderr": 0.04082482904638628
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.18627450980392157,
"acc_stderr": 0.038739587141493524,
"acc_norm": 0.18627450980392157,
"acc_norm_stderr": 0.038739587141493524
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.3907563025210084,
"acc_stderr": 0.031693802357129965,
"acc_norm": 0.3907563025210084,
"acc_norm_stderr": 0.031693802357129965
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.4205128205128205,
"acc_stderr": 0.025028610276710855,
"acc_norm": 0.4205128205128205,
"acc_norm_stderr": 0.025028610276710855
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.52,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.52,
"acc_norm_stderr": 0.050211673156867795
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.4722222222222222,
"acc_stderr": 0.048262172941398944,
"acc_norm": 0.4722222222222222,
"acc_norm_stderr": 0.048262172941398944
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.4088669950738916,
"acc_stderr": 0.034590588158832314,
"acc_norm": 0.4088669950738916,
"acc_norm_stderr": 0.034590588158832314
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.4129032258064516,
"acc_stderr": 0.02800913812540039,
"acc_norm": 0.4129032258064516,
"acc_norm_stderr": 0.02800913812540039
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.6068376068376068,
"acc_stderr": 0.03199957924651048,
"acc_norm": 0.6068376068376068,
"acc_norm_stderr": 0.03199957924651048
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4188679245283019,
"acc_stderr": 0.030365050829115205,
"acc_norm": 0.4188679245283019,
"acc_norm_stderr": 0.030365050829115205
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.44545454545454544,
"acc_stderr": 0.047605488214603246,
"acc_norm": 0.44545454545454544,
"acc_norm_stderr": 0.047605488214603246
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.26666666666666666,
"acc_stderr": 0.02696242432507382,
"acc_norm": 0.26666666666666666,
"acc_norm_stderr": 0.02696242432507382
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.271523178807947,
"acc_stderr": 0.03631329803969653,
"acc_norm": 0.271523178807947,
"acc_norm_stderr": 0.03631329803969653
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.5323383084577115,
"acc_stderr": 0.035281314729336065,
"acc_norm": 0.5323383084577115,
"acc_norm_stderr": 0.035281314729336065
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.3583815028901734,
"acc_stderr": 0.0365634365335316,
"acc_norm": 0.3583815028901734,
"acc_norm_stderr": 0.0365634365335316
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.2751322751322751,
"acc_stderr": 0.02300008685906864,
"acc_norm": 0.2751322751322751,
"acc_norm_stderr": 0.02300008685906864
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.3680555555555556,
"acc_stderr": 0.04032999053960719,
"acc_norm": 0.3680555555555556,
"acc_norm_stderr": 0.04032999053960719
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695236,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695236
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.63,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.63,
"acc_norm_stderr": 0.048523658709391
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.45375722543352603,
"acc_stderr": 0.026803720583206174,
"acc_norm": 0.45375722543352603,
"acc_norm_stderr": 0.026803720583206174
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.4294478527607362,
"acc_stderr": 0.038890666191127216,
"acc_norm": 0.4294478527607362,
"acc_norm_stderr": 0.038890666191127216
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.42901234567901236,
"acc_stderr": 0.027538925613470863,
"acc_norm": 0.42901234567901236,
"acc_norm_stderr": 0.027538925613470863
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252606,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252606
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.45595854922279794,
"acc_stderr": 0.03594413711272438,
"acc_norm": 0.45595854922279794,
"acc_norm_stderr": 0.03594413711272438
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.23684210526315788,
"acc_stderr": 0.03999423879281337,
"acc_norm": 0.23684210526315788,
"acc_norm_stderr": 0.03999423879281337
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.4972477064220184,
"acc_stderr": 0.021436998359765317,
"acc_norm": 0.4972477064220184,
"acc_norm_stderr": 0.021436998359765317
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.24603174603174602,
"acc_stderr": 0.03852273364924314,
"acc_norm": 0.24603174603174602,
"acc_norm_stderr": 0.03852273364924314
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.4019607843137255,
"acc_stderr": 0.028074158947600666,
"acc_norm": 0.4019607843137255,
"acc_norm_stderr": 0.028074158947600666
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.43,
"acc_stderr": 0.04975698519562428,
"acc_norm": 0.43,
"acc_norm_stderr": 0.04975698519562428
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6115702479338843,
"acc_stderr": 0.04449270350068382,
"acc_norm": 0.6115702479338843,
"acc_norm_stderr": 0.04449270350068382
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.39473684210526316,
"acc_stderr": 0.039777499346220734,
"acc_norm": 0.39473684210526316,
"acc_norm_stderr": 0.039777499346220734
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.3480392156862745,
"acc_stderr": 0.019270998708223974,
"acc_norm": 0.3480392156862745,
"acc_norm_stderr": 0.019270998708223974
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.30141843971631205,
"acc_stderr": 0.02737412888263115,
"acc_norm": 0.30141843971631205,
"acc_norm_stderr": 0.02737412888263115
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.22321428571428573,
"acc_stderr": 0.039523019677025116,
"acc_norm": 0.22321428571428573,
"acc_norm_stderr": 0.039523019677025116
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.27314814814814814,
"acc_stderr": 0.03038805130167812,
"acc_norm": 0.27314814814814814,
"acc_norm_stderr": 0.03038805130167812
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2424581005586592,
"acc_stderr": 0.01433352205921789,
"acc_norm": 0.2424581005586592,
"acc_norm_stderr": 0.01433352205921789
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.41,
"acc_stderr": 0.04943110704237102,
"acc_norm": 0.41,
"acc_norm_stderr": 0.04943110704237102
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.34191176470588236,
"acc_stderr": 0.02881472242225417,
"acc_norm": 0.34191176470588236,
"acc_norm_stderr": 0.02881472242225417
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.47346938775510206,
"acc_stderr": 0.03196412734523272,
"acc_norm": 0.47346938775510206,
"acc_norm_stderr": 0.03196412734523272
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.5443037974683544,
"acc_stderr": 0.032419206846933335,
"acc_norm": 0.5443037974683544,
"acc_norm_stderr": 0.032419206846933335
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3070404172099087,
"acc_stderr": 0.01178095911451378,
"acc_norm": 0.3070404172099087,
"acc_norm_stderr": 0.01178095911451378
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.4264705882352941,
"acc_stderr": 0.034711579079534254,
"acc_norm": 0.4264705882352941,
"acc_norm_stderr": 0.034711579079534254
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.509090909090909,
"acc_stderr": 0.03903698647748441,
"acc_norm": 0.509090909090909,
"acc_norm_stderr": 0.03903698647748441
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.2533659730722154,
"mc1_stderr": 0.01522589934082683,
"mc2": 0.39177380761625485,
"mc2_stderr": 0.014625221380747738
},
"harness|ko_commongen_v2|2": {
"acc": 0.4805194805194805,
"acc_stderr": 0.017177301992342558,
"acc_norm": 0.5584415584415584,
"acc_norm_stderr": 0.017072525875563103
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "Puluming/AISquare-Instruct-llama2-koen-13b-v0.9.2",
"model_sha": "e61e6122ceca6995569c008901a3e1d4a0a58972",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}