results / kyujinpy /KoT-platypus2-13B /result_2023-10-07 18:04:13.json
open-ko-llm-bot's picture
Add results for 2023-10-07 18:04:13
c284534
raw
history blame
17.8 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.371160409556314,
"acc_stderr": 0.014117971901142818,
"acc_norm": 0.43686006825938567,
"acc_norm_stderr": 0.014494421584256515
},
"harness|ko_hellaswag|10": {
"acc": 0.40221071499701255,
"acc_stderr": 0.004893418929918276,
"acc_norm": 0.5304720175263892,
"acc_norm_stderr": 0.004980506329407588
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.52046783625731,
"acc_stderr": 0.038316105328219316,
"acc_norm": 0.52046783625731,
"acc_norm_stderr": 0.038316105328219316
},
"harness|ko_mmlu_management|5": {
"acc": 0.5145631067961165,
"acc_stderr": 0.04948637324026637,
"acc_norm": 0.5145631067961165,
"acc_norm_stderr": 0.04948637324026637
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.5172413793103449,
"acc_stderr": 0.017869330154003698,
"acc_norm": 0.5172413793103449,
"acc_norm_stderr": 0.017869330154003698
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.45925925925925926,
"acc_stderr": 0.04304979692464243,
"acc_norm": 0.45925925925925926,
"acc_norm_stderr": 0.04304979692464243
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.27,
"acc_stderr": 0.0446196043338474,
"acc_norm": 0.27,
"acc_norm_stderr": 0.0446196043338474
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.35319148936170214,
"acc_stderr": 0.03124532520276193,
"acc_norm": 0.35319148936170214,
"acc_norm_stderr": 0.03124532520276193
},
"harness|ko_mmlu_virology|5": {
"acc": 0.3674698795180723,
"acc_stderr": 0.03753267402120574,
"acc_norm": 0.3674698795180723,
"acc_norm_stderr": 0.03753267402120574
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.45016077170418006,
"acc_stderr": 0.02825666072336018,
"acc_norm": 0.45016077170418006,
"acc_norm_stderr": 0.02825666072336018
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.4080717488789238,
"acc_stderr": 0.03298574607842821,
"acc_norm": 0.4080717488789238,
"acc_norm_stderr": 0.03298574607842821
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.44274809160305345,
"acc_stderr": 0.043564472026650695,
"acc_norm": 0.44274809160305345,
"acc_norm_stderr": 0.043564472026650695
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.38,
"acc_stderr": 0.048783173121456316,
"acc_norm": 0.38,
"acc_norm_stderr": 0.048783173121456316
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.5202020202020202,
"acc_stderr": 0.03559443565563918,
"acc_norm": 0.5202020202020202,
"acc_norm_stderr": 0.03559443565563918
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.4206896551724138,
"acc_stderr": 0.0411391498118926,
"acc_norm": 0.4206896551724138,
"acc_norm_stderr": 0.0411391498118926
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.2647058823529412,
"acc_stderr": 0.04389869956808777,
"acc_norm": 0.2647058823529412,
"acc_norm_stderr": 0.04389869956808777
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.4117647058823529,
"acc_stderr": 0.03196876989195779,
"acc_norm": 0.4117647058823529,
"acc_norm_stderr": 0.03196876989195779
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.4025641025641026,
"acc_stderr": 0.02486499515976777,
"acc_norm": 0.4025641025641026,
"acc_norm_stderr": 0.02486499515976777
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.49,
"acc_stderr": 0.05024183937956911,
"acc_norm": 0.49,
"acc_norm_stderr": 0.05024183937956911
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.5462962962962963,
"acc_stderr": 0.04812917324536823,
"acc_norm": 0.5462962962962963,
"acc_norm_stderr": 0.04812917324536823
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.35467980295566504,
"acc_stderr": 0.03366124489051448,
"acc_norm": 0.35467980295566504,
"acc_norm_stderr": 0.03366124489051448
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.4483870967741935,
"acc_stderr": 0.028292056830112735,
"acc_norm": 0.4483870967741935,
"acc_norm_stderr": 0.028292056830112735
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.6068376068376068,
"acc_stderr": 0.03199957924651048,
"acc_norm": 0.6068376068376068,
"acc_norm_stderr": 0.03199957924651048
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4377358490566038,
"acc_stderr": 0.03053333843046751,
"acc_norm": 0.4377358490566038,
"acc_norm_stderr": 0.03053333843046751
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.5545454545454546,
"acc_stderr": 0.047605488214603246,
"acc_norm": 0.5545454545454546,
"acc_norm_stderr": 0.047605488214603246
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.29259259259259257,
"acc_stderr": 0.02773896963217609,
"acc_norm": 0.29259259259259257,
"acc_norm_stderr": 0.02773896963217609
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.33774834437086093,
"acc_stderr": 0.038615575462551684,
"acc_norm": 0.33774834437086093,
"acc_norm_stderr": 0.038615575462551684
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.5920398009950248,
"acc_stderr": 0.03475116365194092,
"acc_norm": 0.5920398009950248,
"acc_norm_stderr": 0.03475116365194092
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.4046242774566474,
"acc_stderr": 0.0374246119388725,
"acc_norm": 0.4046242774566474,
"acc_norm_stderr": 0.0374246119388725
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.2804232804232804,
"acc_stderr": 0.02313528797432564,
"acc_norm": 0.2804232804232804,
"acc_norm_stderr": 0.02313528797432564
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.375,
"acc_stderr": 0.04048439222695598,
"acc_norm": 0.375,
"acc_norm_stderr": 0.04048439222695598
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.29,
"acc_stderr": 0.045604802157206845,
"acc_norm": 0.29,
"acc_norm_stderr": 0.045604802157206845
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.65,
"acc_stderr": 0.04793724854411018,
"acc_norm": 0.65,
"acc_norm_stderr": 0.04793724854411018
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.47398843930635837,
"acc_stderr": 0.02688264343402289,
"acc_norm": 0.47398843930635837,
"acc_norm_stderr": 0.02688264343402289
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.4171779141104294,
"acc_stderr": 0.038741028598180814,
"acc_norm": 0.4171779141104294,
"acc_norm_stderr": 0.038741028598180814
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.4567901234567901,
"acc_stderr": 0.027716661650194048,
"acc_norm": 0.4567901234567901,
"acc_norm_stderr": 0.027716661650194048
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.32,
"acc_stderr": 0.04688261722621504,
"acc_norm": 0.32,
"acc_norm_stderr": 0.04688261722621504
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.5233160621761658,
"acc_stderr": 0.03604513672442202,
"acc_norm": 0.5233160621761658,
"acc_norm_stderr": 0.03604513672442202
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.22807017543859648,
"acc_stderr": 0.03947152782669415,
"acc_norm": 0.22807017543859648,
"acc_norm_stderr": 0.03947152782669415
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.5541284403669725,
"acc_stderr": 0.02131133500970857,
"acc_norm": 0.5541284403669725,
"acc_norm_stderr": 0.02131133500970857
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.31746031746031744,
"acc_stderr": 0.0416345303130286,
"acc_norm": 0.31746031746031744,
"acc_norm_stderr": 0.0416345303130286
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.434640522875817,
"acc_stderr": 0.028384256704883037,
"acc_norm": 0.434640522875817,
"acc_norm_stderr": 0.028384256704883037
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.49,
"acc_stderr": 0.05024183937956912,
"acc_norm": 0.49,
"acc_norm_stderr": 0.05024183937956912
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6363636363636364,
"acc_stderr": 0.043913262867240704,
"acc_norm": 0.6363636363636364,
"acc_norm_stderr": 0.043913262867240704
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.4144736842105263,
"acc_stderr": 0.04008973785779206,
"acc_norm": 0.4144736842105263,
"acc_norm_stderr": 0.04008973785779206
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.3562091503267974,
"acc_stderr": 0.0193733324207245,
"acc_norm": 0.3562091503267974,
"acc_norm_stderr": 0.0193733324207245
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.34397163120567376,
"acc_stderr": 0.028338017428611317,
"acc_norm": 0.34397163120567376,
"acc_norm_stderr": 0.028338017428611317
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.22321428571428573,
"acc_stderr": 0.039523019677025116,
"acc_norm": 0.22321428571428573,
"acc_norm_stderr": 0.039523019677025116
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.3611111111111111,
"acc_stderr": 0.03275773486100999,
"acc_norm": 0.3611111111111111,
"acc_norm_stderr": 0.03275773486100999
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2558659217877095,
"acc_stderr": 0.014593620923210756,
"acc_norm": 0.2558659217877095,
"acc_norm_stderr": 0.014593620923210756
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.42,
"acc_stderr": 0.049604496374885836,
"acc_norm": 0.42,
"acc_norm_stderr": 0.049604496374885836
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.45,
"acc_stderr": 0.05,
"acc_norm": 0.45,
"acc_norm_stderr": 0.05
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.40441176470588236,
"acc_stderr": 0.029812630701569743,
"acc_norm": 0.40441176470588236,
"acc_norm_stderr": 0.029812630701569743
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.4775510204081633,
"acc_stderr": 0.031976941187136725,
"acc_norm": 0.4775510204081633,
"acc_norm_stderr": 0.031976941187136725
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.5864978902953587,
"acc_stderr": 0.03205649904851858,
"acc_norm": 0.5864978902953587,
"acc_norm_stderr": 0.03205649904851858
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3683181225554107,
"acc_stderr": 0.01231940336956464,
"acc_norm": 0.3683181225554107,
"acc_norm_stderr": 0.01231940336956464
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.4411764705882353,
"acc_stderr": 0.034849415144292316,
"acc_norm": 0.4411764705882353,
"acc_norm_stderr": 0.034849415144292316
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.47878787878787876,
"acc_stderr": 0.03900828913737302,
"acc_norm": 0.47878787878787876,
"acc_norm_stderr": 0.03900828913737302
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.2668298653610771,
"mc1_stderr": 0.015483691939237272,
"mc2": 0.4334291763920242,
"mc2_stderr": 0.014968924711902113
},
"harness|ko_commongen_v2|2": {
"acc": 0.5586854460093896,
"acc_stderr": 0.017021311671847474,
"acc_norm": 0.653755868544601,
"acc_norm_stderr": 0.016309259043203083
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "kyujinpy/KoT-platypus2-13B",
"model_sha": "a5d295abdb2fca50971ee29b4db84f7565c67ab3",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}