results / LDCC /LDCC-Instruct-Llama-2-ko-13B-v1.2 /result_2023-10-31 01:25:10.json
open-ko-llm-bot's picture
Add results for 2023-10-31 01:25:10
e1d26ee
raw history blame
No virus
17.9 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.40017064846416384,
"acc_stderr": 0.014317197787809169,
"acc_norm": 0.45307167235494883,
"acc_norm_stderr": 0.014546892052005628
},
"harness|ko_hellaswag|10": {
"acc": 0.4222266480780721,
"acc_stderr": 0.0049290484827604515,
"acc_norm": 0.566620195180243,
"acc_norm_stderr": 0.004945291270072434
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.49122807017543857,
"acc_stderr": 0.038342347441649924,
"acc_norm": 0.49122807017543857,
"acc_norm_stderr": 0.038342347441649924
},
"harness|ko_mmlu_management|5": {
"acc": 0.44660194174757284,
"acc_stderr": 0.04922424153458933,
"acc_norm": 0.44660194174757284,
"acc_norm_stderr": 0.04922424153458933
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.5389527458492975,
"acc_stderr": 0.01782562179323901,
"acc_norm": 0.5389527458492975,
"acc_norm_stderr": 0.01782562179323901
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.45185185185185184,
"acc_stderr": 0.042992689054808624,
"acc_norm": 0.45185185185185184,
"acc_norm_stderr": 0.042992689054808624
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.21,
"acc_stderr": 0.040936018074033256,
"acc_norm": 0.21,
"acc_norm_stderr": 0.040936018074033256
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.3276595744680851,
"acc_stderr": 0.03068302084323101,
"acc_norm": 0.3276595744680851,
"acc_norm_stderr": 0.03068302084323101
},
"harness|ko_mmlu_virology|5": {
"acc": 0.42168674698795183,
"acc_stderr": 0.03844453181770917,
"acc_norm": 0.42168674698795183,
"acc_norm_stderr": 0.03844453181770917
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.4887459807073955,
"acc_stderr": 0.028390897396863526,
"acc_norm": 0.4887459807073955,
"acc_norm_stderr": 0.028390897396863526
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.4260089686098655,
"acc_stderr": 0.033188332862172806,
"acc_norm": 0.4260089686098655,
"acc_norm_stderr": 0.033188332862172806
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.5114503816793893,
"acc_stderr": 0.043841400240780176,
"acc_norm": 0.5114503816793893,
"acc_norm_stderr": 0.043841400240780176
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.38,
"acc_stderr": 0.048783173121456316,
"acc_norm": 0.38,
"acc_norm_stderr": 0.048783173121456316
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.5202020202020202,
"acc_stderr": 0.035594435655639196,
"acc_norm": 0.5202020202020202,
"acc_norm_stderr": 0.035594435655639196
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.4413793103448276,
"acc_stderr": 0.04137931034482758,
"acc_norm": 0.4413793103448276,
"acc_norm_stderr": 0.04137931034482758
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.2549019607843137,
"acc_stderr": 0.043364327079931785,
"acc_norm": 0.2549019607843137,
"acc_norm_stderr": 0.043364327079931785
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.453781512605042,
"acc_stderr": 0.03233943468182088,
"acc_norm": 0.453781512605042,
"acc_norm_stderr": 0.03233943468182088
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.4282051282051282,
"acc_stderr": 0.025088301454694834,
"acc_norm": 0.4282051282051282,
"acc_norm_stderr": 0.025088301454694834
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.53,
"acc_stderr": 0.05016135580465919,
"acc_norm": 0.53,
"acc_norm_stderr": 0.05016135580465919
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.37,
"acc_stderr": 0.04852365870939099,
"acc_norm": 0.37,
"acc_norm_stderr": 0.04852365870939099
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.5,
"acc_stderr": 0.04833682445228318,
"acc_norm": 0.5,
"acc_norm_stderr": 0.04833682445228318
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.39901477832512317,
"acc_stderr": 0.03445487686264715,
"acc_norm": 0.39901477832512317,
"acc_norm_stderr": 0.03445487686264715
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.44516129032258067,
"acc_stderr": 0.028272410186214906,
"acc_norm": 0.44516129032258067,
"acc_norm_stderr": 0.028272410186214906
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.5854700854700855,
"acc_stderr": 0.03227396567623779,
"acc_norm": 0.5854700854700855,
"acc_norm_stderr": 0.03227396567623779
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4377358490566038,
"acc_stderr": 0.030533338430467512,
"acc_norm": 0.4377358490566038,
"acc_norm_stderr": 0.030533338430467512
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.509090909090909,
"acc_stderr": 0.04788339768702861,
"acc_norm": 0.509090909090909,
"acc_norm_stderr": 0.04788339768702861
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.26666666666666666,
"acc_stderr": 0.026962424325073838,
"acc_norm": 0.26666666666666666,
"acc_norm_stderr": 0.026962424325073838
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.3576158940397351,
"acc_stderr": 0.03913453431177258,
"acc_norm": 0.3576158940397351,
"acc_norm_stderr": 0.03913453431177258
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.5870646766169154,
"acc_stderr": 0.03481520803367348,
"acc_norm": 0.5870646766169154,
"acc_norm_stderr": 0.03481520803367348
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.3930635838150289,
"acc_stderr": 0.0372424959581773,
"acc_norm": 0.3930635838150289,
"acc_norm_stderr": 0.0372424959581773
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.328042328042328,
"acc_stderr": 0.024180497164376896,
"acc_norm": 0.328042328042328,
"acc_norm_stderr": 0.024180497164376896
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.4097222222222222,
"acc_stderr": 0.04112490974670788,
"acc_norm": 0.4097222222222222,
"acc_norm_stderr": 0.04112490974670788
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695235,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695235
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.66,
"acc_stderr": 0.04760952285695237,
"acc_norm": 0.66,
"acc_norm_stderr": 0.04760952285695237
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.4884393063583815,
"acc_stderr": 0.026911898686377906,
"acc_norm": 0.4884393063583815,
"acc_norm_stderr": 0.026911898686377906
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.5276073619631901,
"acc_stderr": 0.0392237829061099,
"acc_norm": 0.5276073619631901,
"acc_norm_stderr": 0.0392237829061099
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.4722222222222222,
"acc_stderr": 0.027777777777777804,
"acc_norm": 0.4722222222222222,
"acc_norm_stderr": 0.027777777777777804
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.35,
"acc_stderr": 0.0479372485441102,
"acc_norm": 0.35,
"acc_norm_stderr": 0.0479372485441102
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.49222797927461137,
"acc_stderr": 0.036080032255696545,
"acc_norm": 0.49222797927461137,
"acc_norm_stderr": 0.036080032255696545
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.2719298245614035,
"acc_stderr": 0.041857744240220575,
"acc_norm": 0.2719298245614035,
"acc_norm_stderr": 0.041857744240220575
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.5798165137614679,
"acc_stderr": 0.0211624200482735,
"acc_norm": 0.5798165137614679,
"acc_norm_stderr": 0.0211624200482735
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.04216370213557835,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.04216370213557835
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.41830065359477125,
"acc_stderr": 0.02824513402438729,
"acc_norm": 0.41830065359477125,
"acc_norm_stderr": 0.02824513402438729
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.42,
"acc_stderr": 0.049604496374885836,
"acc_norm": 0.42,
"acc_norm_stderr": 0.049604496374885836
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.628099173553719,
"acc_stderr": 0.044120158066245044,
"acc_norm": 0.628099173553719,
"acc_norm_stderr": 0.044120158066245044
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.4144736842105263,
"acc_stderr": 0.04008973785779206,
"acc_norm": 0.4144736842105263,
"acc_norm_stderr": 0.04008973785779206
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.369281045751634,
"acc_stderr": 0.01952431674486635,
"acc_norm": 0.369281045751634,
"acc_norm_stderr": 0.01952431674486635
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.36524822695035464,
"acc_stderr": 0.02872386385328128,
"acc_norm": 0.36524822695035464,
"acc_norm_stderr": 0.02872386385328128
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.22321428571428573,
"acc_stderr": 0.03952301967702511,
"acc_norm": 0.22321428571428573,
"acc_norm_stderr": 0.03952301967702511
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.37962962962962965,
"acc_stderr": 0.03309682581119035,
"acc_norm": 0.37962962962962965,
"acc_norm_stderr": 0.03309682581119035
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2424581005586592,
"acc_stderr": 0.01433352205921789,
"acc_norm": 0.2424581005586592,
"acc_norm_stderr": 0.01433352205921789
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.36,
"acc_stderr": 0.04824181513244218,
"acc_norm": 0.36,
"acc_norm_stderr": 0.04824181513244218
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.44,
"acc_stderr": 0.049888765156985884,
"acc_norm": 0.44,
"acc_norm_stderr": 0.049888765156985884
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.45955882352941174,
"acc_stderr": 0.03027332507734575,
"acc_norm": 0.45955882352941174,
"acc_norm_stderr": 0.03027332507734575
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.4489795918367347,
"acc_stderr": 0.03184213866687579,
"acc_norm": 0.4489795918367347,
"acc_norm_stderr": 0.03184213866687579
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.5991561181434599,
"acc_stderr": 0.03190080389473236,
"acc_norm": 0.5991561181434599,
"acc_norm_stderr": 0.03190080389473236
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3670143415906128,
"acc_stderr": 0.01231026424484213,
"acc_norm": 0.3670143415906128,
"acc_norm_stderr": 0.01231026424484213
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.4950980392156863,
"acc_stderr": 0.03509143375606785,
"acc_norm": 0.4950980392156863,
"acc_norm_stderr": 0.03509143375606785
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.5151515151515151,
"acc_stderr": 0.03902551007374448,
"acc_norm": 0.5151515151515151,
"acc_norm_stderr": 0.03902551007374448
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.2631578947368421,
"mc1_stderr": 0.01541524174023703,
"mc2": 0.41898060116595187,
"mc2_stderr": 0.014731537822096375
},
"harness|ko_commongen_v2|2": {
"acc": 0.4935064935064935,
"acc_stderr": 0.01718890435907731,
"acc_norm": 0.6198347107438017,
"acc_norm_stderr": 0.016689333596980094
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "LDCC/LDCC-Instruct-Llama-2-ko-13B-v1.2",
"model_sha": "01fd24c5633c041fc150a92d285b67a58aa42d1b",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}