results / DopeorNope /mistralopithecus-v1-SFT /result_2023-11-26 08:04:14.json
open-ko-llm-bot's picture
Add results for 2023-11-26 08:04:14
8f8be45
raw history blame
No virus
17.9 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.4283276450511945,
"acc_stderr": 0.014460496367599008,
"acc_norm": 0.47696245733788395,
"acc_norm_stderr": 0.014595873205358269
},
"harness|ko_hellaswag|10": {
"acc": 0.40689105755825533,
"acc_stderr": 0.004902502514738604,
"acc_norm": 0.5343557060346544,
"acc_norm_stderr": 0.00497798845250264
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.5146198830409356,
"acc_stderr": 0.03833185275213025,
"acc_norm": 0.5146198830409356,
"acc_norm_stderr": 0.03833185275213025
},
"harness|ko_mmlu_management|5": {
"acc": 0.6019417475728155,
"acc_stderr": 0.048467482539772386,
"acc_norm": 0.6019417475728155,
"acc_norm_stderr": 0.048467482539772386
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.5504469987228607,
"acc_stderr": 0.017788725283507337,
"acc_norm": 0.5504469987228607,
"acc_norm_stderr": 0.017788725283507337
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.362962962962963,
"acc_stderr": 0.041539484047424,
"acc_norm": 0.362962962962963,
"acc_norm_stderr": 0.041539484047424
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.27,
"acc_stderr": 0.0446196043338474,
"acc_norm": 0.27,
"acc_norm_stderr": 0.0446196043338474
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.425531914893617,
"acc_stderr": 0.03232146916224468,
"acc_norm": 0.425531914893617,
"acc_norm_stderr": 0.03232146916224468
},
"harness|ko_mmlu_virology|5": {
"acc": 0.41566265060240964,
"acc_stderr": 0.038367221765980515,
"acc_norm": 0.41566265060240964,
"acc_norm_stderr": 0.038367221765980515
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.4533762057877814,
"acc_stderr": 0.02827435985489424,
"acc_norm": 0.4533762057877814,
"acc_norm_stderr": 0.02827435985489424
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.4618834080717489,
"acc_stderr": 0.033460150119732274,
"acc_norm": 0.4618834080717489,
"acc_norm_stderr": 0.033460150119732274
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.5038167938931297,
"acc_stderr": 0.043851623256015534,
"acc_norm": 0.5038167938931297,
"acc_norm_stderr": 0.043851623256015534
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.47,
"acc_stderr": 0.05016135580465919,
"acc_norm": 0.47,
"acc_norm_stderr": 0.05016135580465919
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.6111111111111112,
"acc_stderr": 0.0347327959083696,
"acc_norm": 0.6111111111111112,
"acc_norm_stderr": 0.0347327959083696
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.4689655172413793,
"acc_stderr": 0.04158632762097828,
"acc_norm": 0.4689655172413793,
"acc_norm_stderr": 0.04158632762097828
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.28431372549019607,
"acc_stderr": 0.04488482852329017,
"acc_norm": 0.28431372549019607,
"acc_norm_stderr": 0.04488482852329017
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.5294117647058824,
"acc_stderr": 0.03242225027115006,
"acc_norm": 0.5294117647058824,
"acc_norm_stderr": 0.03242225027115006
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.4897435897435897,
"acc_stderr": 0.025345672221942374,
"acc_norm": 0.4897435897435897,
"acc_norm_stderr": 0.025345672221942374
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.64,
"acc_stderr": 0.04824181513244218,
"acc_norm": 0.64,
"acc_norm_stderr": 0.04824181513244218
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.32,
"acc_stderr": 0.04688261722621504,
"acc_norm": 0.32,
"acc_norm_stderr": 0.04688261722621504
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.5092592592592593,
"acc_stderr": 0.04832853553437055,
"acc_norm": 0.5092592592592593,
"acc_norm_stderr": 0.04832853553437055
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.3891625615763547,
"acc_stderr": 0.034304624161038716,
"acc_norm": 0.3891625615763547,
"acc_norm_stderr": 0.034304624161038716
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.4870967741935484,
"acc_stderr": 0.028434533152681848,
"acc_norm": 0.4870967741935484,
"acc_norm_stderr": 0.028434533152681848
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.7222222222222222,
"acc_stderr": 0.02934311479809447,
"acc_norm": 0.7222222222222222,
"acc_norm_stderr": 0.02934311479809447
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4981132075471698,
"acc_stderr": 0.030772653642075664,
"acc_norm": 0.4981132075471698,
"acc_norm_stderr": 0.030772653642075664
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.509090909090909,
"acc_stderr": 0.04788339768702861,
"acc_norm": 0.509090909090909,
"acc_norm_stderr": 0.04788339768702861
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.3111111111111111,
"acc_stderr": 0.02822644674968352,
"acc_norm": 0.3111111111111111,
"acc_norm_stderr": 0.02822644674968352
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.3443708609271523,
"acc_stderr": 0.038796870240733264,
"acc_norm": 0.3443708609271523,
"acc_norm_stderr": 0.038796870240733264
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.6069651741293532,
"acc_stderr": 0.0345368246603156,
"acc_norm": 0.6069651741293532,
"acc_norm_stderr": 0.0345368246603156
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.41040462427745666,
"acc_stderr": 0.03750757044895538,
"acc_norm": 0.41040462427745666,
"acc_norm_stderr": 0.03750757044895538
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.36772486772486773,
"acc_stderr": 0.02483383982556242,
"acc_norm": 0.36772486772486773,
"acc_norm_stderr": 0.02483383982556242
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.3611111111111111,
"acc_stderr": 0.040166600304512336,
"acc_norm": 0.3611111111111111,
"acc_norm_stderr": 0.040166600304512336
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695235,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695235
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.64,
"acc_stderr": 0.048241815132442176,
"acc_norm": 0.64,
"acc_norm_stderr": 0.048241815132442176
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.546242774566474,
"acc_stderr": 0.026803720583206174,
"acc_norm": 0.546242774566474,
"acc_norm_stderr": 0.026803720583206174
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.50920245398773,
"acc_stderr": 0.03927705600787443,
"acc_norm": 0.50920245398773,
"acc_norm_stderr": 0.03927705600787443
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.4691358024691358,
"acc_stderr": 0.027767689606833942,
"acc_norm": 0.4691358024691358,
"acc_norm_stderr": 0.027767689606833942
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.33,
"acc_stderr": 0.04725815626252605,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252605
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.5595854922279793,
"acc_stderr": 0.035827245300360945,
"acc_norm": 0.5595854922279793,
"acc_norm_stderr": 0.035827245300360945
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.34210526315789475,
"acc_stderr": 0.04462917535336938,
"acc_norm": 0.34210526315789475,
"acc_norm_stderr": 0.04462917535336938
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.5541284403669725,
"acc_stderr": 0.021311335009708582,
"acc_norm": 0.5541284403669725,
"acc_norm_stderr": 0.021311335009708582
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.40476190476190477,
"acc_stderr": 0.043902592653775614,
"acc_norm": 0.40476190476190477,
"acc_norm_stderr": 0.043902592653775614
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.46078431372549017,
"acc_stderr": 0.028541722692618874,
"acc_norm": 0.46078431372549017,
"acc_norm_stderr": 0.028541722692618874
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.46,
"acc_stderr": 0.05009082659620332,
"acc_norm": 0.46,
"acc_norm_stderr": 0.05009082659620332
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6776859504132231,
"acc_stderr": 0.042664163633521664,
"acc_norm": 0.6776859504132231,
"acc_norm_stderr": 0.042664163633521664
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.4342105263157895,
"acc_stderr": 0.04033565667848319,
"acc_norm": 0.4342105263157895,
"acc_norm_stderr": 0.04033565667848319
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.39869281045751637,
"acc_stderr": 0.019808281317449848,
"acc_norm": 0.39869281045751637,
"acc_norm_stderr": 0.019808281317449848
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.3049645390070922,
"acc_stderr": 0.02746470844202214,
"acc_norm": 0.3049645390070922,
"acc_norm_stderr": 0.02746470844202214
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.4375,
"acc_stderr": 0.04708567521880525,
"acc_norm": 0.4375,
"acc_norm_stderr": 0.04708567521880525
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.3611111111111111,
"acc_stderr": 0.032757734861009996,
"acc_norm": 0.3611111111111111,
"acc_norm_stderr": 0.032757734861009996
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.18659217877094972,
"acc_stderr": 0.013029631416358349,
"acc_norm": 0.18659217877094972,
"acc_norm_stderr": 0.013029631416358349
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.44,
"acc_stderr": 0.0498887651569859,
"acc_norm": 0.44,
"acc_norm_stderr": 0.0498887651569859
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.58,
"acc_stderr": 0.049604496374885836,
"acc_norm": 0.58,
"acc_norm_stderr": 0.049604496374885836
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.3639705882352941,
"acc_stderr": 0.02922719246003202,
"acc_norm": 0.3639705882352941,
"acc_norm_stderr": 0.02922719246003202
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.5795918367346938,
"acc_stderr": 0.031601069934496004,
"acc_norm": 0.5795918367346938,
"acc_norm_stderr": 0.031601069934496004
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.6286919831223629,
"acc_stderr": 0.03145068600744859,
"acc_norm": 0.6286919831223629,
"acc_norm_stderr": 0.03145068600744859
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3318122555410691,
"acc_stderr": 0.012026088259897634,
"acc_norm": 0.3318122555410691,
"acc_norm_stderr": 0.012026088259897634
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.4803921568627451,
"acc_stderr": 0.03506612560524866,
"acc_norm": 0.4803921568627451,
"acc_norm_stderr": 0.03506612560524866
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.47878787878787876,
"acc_stderr": 0.03900828913737301,
"acc_norm": 0.47878787878787876,
"acc_norm_stderr": 0.03900828913737301
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.2839657282741738,
"mc1_stderr": 0.015785370858396708,
"mc2": 0.4537563569615343,
"mc2_stderr": 0.015481816857869497
},
"harness|ko_commongen_v2|2": {
"acc": 0.44391971664698937,
"acc_stderr": 0.017081884623542543,
"acc_norm": 0.47461629279811096,
"acc_norm_stderr": 0.017168187201429253
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "DopeorNope/mistralopithecus-v1-SFT",
"model_sha": "d287f71c14d2bfbcaa053dcaa6b9b22dd5bc0f1a",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}