results / DopeorNope /COKAL_pre_DPO_Test_v1-13b /result_2023-11-10 04:48:39.json
open-ko-llm-bot's picture
Add results for 2023-11-10 04:48:39
b90090e
raw history blame
No virus
17.9 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.4206484641638225,
"acc_stderr": 0.014426211252508406,
"acc_norm": 0.4786689419795222,
"acc_norm_stderr": 0.014598087973127104
},
"harness|ko_hellaswag|10": {
"acc": 0.4311890061740689,
"acc_stderr": 0.004942302768002103,
"acc_norm": 0.5746863174666401,
"acc_norm_stderr": 0.004933800927560538
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.5087719298245614,
"acc_stderr": 0.038342347441649924,
"acc_norm": 0.5087719298245614,
"acc_norm_stderr": 0.038342347441649924
},
"harness|ko_mmlu_management|5": {
"acc": 0.5728155339805825,
"acc_stderr": 0.04897957737781168,
"acc_norm": 0.5728155339805825,
"acc_norm_stderr": 0.04897957737781168
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.558109833971903,
"acc_stderr": 0.01775880053421441,
"acc_norm": 0.558109833971903,
"acc_norm_stderr": 0.01775880053421441
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.45185185185185184,
"acc_stderr": 0.042992689054808624,
"acc_norm": 0.45185185185185184,
"acc_norm_stderr": 0.042992689054808624
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.28,
"acc_stderr": 0.045126085985421255,
"acc_norm": 0.28,
"acc_norm_stderr": 0.045126085985421255
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.4127659574468085,
"acc_stderr": 0.03218471141400351,
"acc_norm": 0.4127659574468085,
"acc_norm_stderr": 0.03218471141400351
},
"harness|ko_mmlu_virology|5": {
"acc": 0.41566265060240964,
"acc_stderr": 0.03836722176598053,
"acc_norm": 0.41566265060240964,
"acc_norm_stderr": 0.03836722176598053
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.5176848874598071,
"acc_stderr": 0.02838032284907713,
"acc_norm": 0.5176848874598071,
"acc_norm_stderr": 0.02838032284907713
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.4977578475336323,
"acc_stderr": 0.033557465352232634,
"acc_norm": 0.4977578475336323,
"acc_norm_stderr": 0.033557465352232634
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.4961832061068702,
"acc_stderr": 0.043851623256015534,
"acc_norm": 0.4961832061068702,
"acc_norm_stderr": 0.043851623256015534
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.35,
"acc_stderr": 0.04793724854411022,
"acc_norm": 0.35,
"acc_norm_stderr": 0.04793724854411022
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.6060606060606061,
"acc_stderr": 0.03481285338232964,
"acc_norm": 0.6060606060606061,
"acc_norm_stderr": 0.03481285338232964
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.4206896551724138,
"acc_stderr": 0.0411391498118926,
"acc_norm": 0.4206896551724138,
"acc_norm_stderr": 0.0411391498118926
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.24509803921568626,
"acc_stderr": 0.042801058373643966,
"acc_norm": 0.24509803921568626,
"acc_norm_stderr": 0.042801058373643966
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.4789915966386555,
"acc_stderr": 0.03244980849990029,
"acc_norm": 0.4789915966386555,
"acc_norm_stderr": 0.03244980849990029
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.47692307692307695,
"acc_stderr": 0.025323990861736125,
"acc_norm": 0.47692307692307695,
"acc_norm_stderr": 0.025323990861736125
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.52,
"acc_stderr": 0.050211673156867795,
"acc_norm": 0.52,
"acc_norm_stderr": 0.050211673156867795
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.36,
"acc_stderr": 0.04824181513244218,
"acc_norm": 0.36,
"acc_norm_stderr": 0.04824181513244218
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.5,
"acc_stderr": 0.04833682445228318,
"acc_norm": 0.5,
"acc_norm_stderr": 0.04833682445228318
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.4088669950738916,
"acc_stderr": 0.034590588158832314,
"acc_norm": 0.4088669950738916,
"acc_norm_stderr": 0.034590588158832314
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.47419354838709676,
"acc_stderr": 0.028406095057653315,
"acc_norm": 0.47419354838709676,
"acc_norm_stderr": 0.028406095057653315
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.6666666666666666,
"acc_stderr": 0.03088273697413866,
"acc_norm": 0.6666666666666666,
"acc_norm_stderr": 0.03088273697413866
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.45660377358490567,
"acc_stderr": 0.03065674869673943,
"acc_norm": 0.45660377358490567,
"acc_norm_stderr": 0.03065674869673943
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.5454545454545454,
"acc_stderr": 0.04769300568972744,
"acc_norm": 0.5454545454545454,
"acc_norm_stderr": 0.04769300568972744
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.27037037037037037,
"acc_stderr": 0.027080372815145658,
"acc_norm": 0.27037037037037037,
"acc_norm_stderr": 0.027080372815145658
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.2913907284768212,
"acc_stderr": 0.037101857261199946,
"acc_norm": 0.2913907284768212,
"acc_norm_stderr": 0.037101857261199946
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.5870646766169154,
"acc_stderr": 0.03481520803367348,
"acc_norm": 0.5870646766169154,
"acc_norm_stderr": 0.03481520803367348
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.4277456647398844,
"acc_stderr": 0.037724468575180276,
"acc_norm": 0.4277456647398844,
"acc_norm_stderr": 0.037724468575180276
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.29365079365079366,
"acc_stderr": 0.023456037383982026,
"acc_norm": 0.29365079365079366,
"acc_norm_stderr": 0.023456037383982026
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.4097222222222222,
"acc_stderr": 0.04112490974670787,
"acc_norm": 0.4097222222222222,
"acc_norm_stderr": 0.04112490974670787
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.36,
"acc_stderr": 0.04824181513244218,
"acc_norm": 0.36,
"acc_norm_stderr": 0.04824181513244218
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.62,
"acc_stderr": 0.04878317312145634,
"acc_norm": 0.62,
"acc_norm_stderr": 0.04878317312145634
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.49710982658959535,
"acc_stderr": 0.02691864538323901,
"acc_norm": 0.49710982658959535,
"acc_norm_stderr": 0.02691864538323901
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.5214723926380368,
"acc_stderr": 0.03924746876751129,
"acc_norm": 0.5214723926380368,
"acc_norm_stderr": 0.03924746876751129
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.4691358024691358,
"acc_stderr": 0.027767689606833935,
"acc_norm": 0.4691358024691358,
"acc_norm_stderr": 0.027767689606833935
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.6113989637305699,
"acc_stderr": 0.03517739796373132,
"acc_norm": 0.6113989637305699,
"acc_norm_stderr": 0.03517739796373132
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.2807017543859649,
"acc_stderr": 0.042270544512321984,
"acc_norm": 0.2807017543859649,
"acc_norm_stderr": 0.042270544512321984
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.6311926605504588,
"acc_stderr": 0.020686227560729534,
"acc_norm": 0.6311926605504588,
"acc_norm_stderr": 0.020686227560729534
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.04216370213557835,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.04216370213557835
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.42810457516339867,
"acc_stderr": 0.028332397483664274,
"acc_norm": 0.42810457516339867,
"acc_norm_stderr": 0.028332397483664274
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.43,
"acc_stderr": 0.04975698519562428,
"acc_norm": 0.43,
"acc_norm_stderr": 0.04975698519562428
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.628099173553719,
"acc_stderr": 0.04412015806624504,
"acc_norm": 0.628099173553719,
"acc_norm_stderr": 0.04412015806624504
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.3618421052631579,
"acc_stderr": 0.039105257528497264,
"acc_norm": 0.3618421052631579,
"acc_norm_stderr": 0.039105257528497264
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.369281045751634,
"acc_stderr": 0.019524316744866342,
"acc_norm": 0.369281045751634,
"acc_norm_stderr": 0.019524316744866342
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.36524822695035464,
"acc_stderr": 0.028723863853281278,
"acc_norm": 0.36524822695035464,
"acc_norm_stderr": 0.028723863853281278
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.22321428571428573,
"acc_stderr": 0.039523019677025116,
"acc_norm": 0.22321428571428573,
"acc_norm_stderr": 0.039523019677025116
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.37962962962962965,
"acc_stderr": 0.03309682581119035,
"acc_norm": 0.37962962962962965,
"acc_norm_stderr": 0.03309682581119035
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2424581005586592,
"acc_stderr": 0.01433352205921789,
"acc_norm": 0.2424581005586592,
"acc_norm_stderr": 0.01433352205921789
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.36,
"acc_stderr": 0.04824181513244218,
"acc_norm": 0.36,
"acc_norm_stderr": 0.04824181513244218
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001975
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.4852941176470588,
"acc_stderr": 0.03035969707904612,
"acc_norm": 0.4852941176470588,
"acc_norm_stderr": 0.03035969707904612
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.49387755102040815,
"acc_stderr": 0.032006820201639086,
"acc_norm": 0.49387755102040815,
"acc_norm_stderr": 0.032006820201639086
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.6455696202531646,
"acc_stderr": 0.031137304297185805,
"acc_norm": 0.6455696202531646,
"acc_norm_stderr": 0.031137304297185805
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3533246414602347,
"acc_stderr": 0.012208408211082425,
"acc_norm": 0.3533246414602347,
"acc_norm_stderr": 0.012208408211082425
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.5147058823529411,
"acc_stderr": 0.03507793834791324,
"acc_norm": 0.5147058823529411,
"acc_norm_stderr": 0.03507793834791324
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.5333333333333333,
"acc_stderr": 0.03895658065271846,
"acc_norm": 0.5333333333333333,
"acc_norm_stderr": 0.03895658065271846
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.3047735618115055,
"mc1_stderr": 0.016114124156882462,
"mc2": 0.4661645299206862,
"mc2_stderr": 0.015362128030709755
},
"harness|ko_commongen_v2|2": {
"acc": 0.47107438016528924,
"acc_stderr": 0.017161563949916345,
"acc_norm": 0.5230224321133412,
"acc_norm_stderr": 0.017172121546727634
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "DopeorNope/COKAL_pre_DPO_Test_v1-13b",
"model_sha": "bb7af9fcf945355418b9457538f30f0fd41b5aac",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}