results / Ja3ck /Mistral-instruct-DPO-Y24-v2 /result_2023-12-06 23:00:18.json
open-ko-llm-bot's picture
Add results for 2023-12-06 23:00:18
96e8cd4
raw history blame
No virus
17.9 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.3532423208191126,
"acc_stderr": 0.013967822714840053,
"acc_norm": 0.4197952218430034,
"acc_norm_stderr": 0.014422181226303028
},
"harness|ko_hellaswag|10": {
"acc": 0.38697470623381797,
"acc_stderr": 0.004860623733461132,
"acc_norm": 0.5070703047201752,
"acc_norm_stderr": 0.004989282516055394
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.5730994152046783,
"acc_stderr": 0.03793620616529916,
"acc_norm": 0.5730994152046783,
"acc_norm_stderr": 0.03793620616529916
},
"harness|ko_mmlu_management|5": {
"acc": 0.5631067961165048,
"acc_stderr": 0.049111471073657764,
"acc_norm": 0.5631067961165048,
"acc_norm_stderr": 0.049111471073657764
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.49169859514687103,
"acc_stderr": 0.017877498991072008,
"acc_norm": 0.49169859514687103,
"acc_norm_stderr": 0.017877498991072008
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.37777777777777777,
"acc_stderr": 0.04188307537595853,
"acc_norm": 0.37777777777777777,
"acc_norm_stderr": 0.04188307537595853
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.22,
"acc_stderr": 0.04163331998932269,
"acc_norm": 0.22,
"acc_norm_stderr": 0.04163331998932269
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.3872340425531915,
"acc_stderr": 0.03184389265339526,
"acc_norm": 0.3872340425531915,
"acc_norm_stderr": 0.03184389265339526
},
"harness|ko_mmlu_virology|5": {
"acc": 0.42771084337349397,
"acc_stderr": 0.03851597683718533,
"acc_norm": 0.42771084337349397,
"acc_norm_stderr": 0.03851597683718533
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.4983922829581994,
"acc_stderr": 0.028397944907806612,
"acc_norm": 0.4983922829581994,
"acc_norm_stderr": 0.028397944907806612
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.45739910313901344,
"acc_stderr": 0.033435777055830646,
"acc_norm": 0.45739910313901344,
"acc_norm_stderr": 0.033435777055830646
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.4122137404580153,
"acc_stderr": 0.04317171194870254,
"acc_norm": 0.4122137404580153,
"acc_norm_stderr": 0.04317171194870254
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.44,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.44,
"acc_norm_stderr": 0.04988876515698589
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.5757575757575758,
"acc_stderr": 0.03521224908841585,
"acc_norm": 0.5757575757575758,
"acc_norm_stderr": 0.03521224908841585
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.47586206896551725,
"acc_stderr": 0.041618085035015295,
"acc_norm": 0.47586206896551725,
"acc_norm_stderr": 0.041618085035015295
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.28431372549019607,
"acc_stderr": 0.04488482852329017,
"acc_norm": 0.28431372549019607,
"acc_norm_stderr": 0.04488482852329017
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.5042016806722689,
"acc_stderr": 0.03247734334448111,
"acc_norm": 0.5042016806722689,
"acc_norm_stderr": 0.03247734334448111
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.44871794871794873,
"acc_stderr": 0.025217315184846482,
"acc_norm": 0.44871794871794873,
"acc_norm_stderr": 0.025217315184846482
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.58,
"acc_stderr": 0.04960449637488583,
"acc_norm": 0.58,
"acc_norm_stderr": 0.04960449637488583
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.37,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.37,
"acc_norm_stderr": 0.048523658709391
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.5092592592592593,
"acc_stderr": 0.04832853553437055,
"acc_norm": 0.5092592592592593,
"acc_norm_stderr": 0.04832853553437055
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.46798029556650245,
"acc_stderr": 0.035107665979592154,
"acc_norm": 0.46798029556650245,
"acc_norm_stderr": 0.035107665979592154
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.4774193548387097,
"acc_stderr": 0.028414985019707868,
"acc_norm": 0.4774193548387097,
"acc_norm_stderr": 0.028414985019707868
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.7307692307692307,
"acc_stderr": 0.029058588303748845,
"acc_norm": 0.7307692307692307,
"acc_norm_stderr": 0.029058588303748845
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4716981132075472,
"acc_stderr": 0.030723535249006107,
"acc_norm": 0.4716981132075472,
"acc_norm_stderr": 0.030723535249006107
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.4727272727272727,
"acc_stderr": 0.04782001791380063,
"acc_norm": 0.4727272727272727,
"acc_norm_stderr": 0.04782001791380063
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.02874204090394849,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.02874204090394849
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.31125827814569534,
"acc_stderr": 0.03780445850526733,
"acc_norm": 0.31125827814569534,
"acc_norm_stderr": 0.03780445850526733
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.6218905472636815,
"acc_stderr": 0.03428867848778658,
"acc_norm": 0.6218905472636815,
"acc_norm_stderr": 0.03428867848778658
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.4161849710982659,
"acc_stderr": 0.03758517775404947,
"acc_norm": 0.4161849710982659,
"acc_norm_stderr": 0.03758517775404947
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.36507936507936506,
"acc_stderr": 0.024796060602699954,
"acc_norm": 0.36507936507936506,
"acc_norm_stderr": 0.024796060602699954
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.3819444444444444,
"acc_stderr": 0.04062990784146667,
"acc_norm": 0.3819444444444444,
"acc_norm_stderr": 0.04062990784146667
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.35,
"acc_stderr": 0.0479372485441102,
"acc_norm": 0.35,
"acc_norm_stderr": 0.0479372485441102
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.63,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.63,
"acc_norm_stderr": 0.048523658709391
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.5289017341040463,
"acc_stderr": 0.02687408588351835,
"acc_norm": 0.5289017341040463,
"acc_norm_stderr": 0.02687408588351835
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.5337423312883436,
"acc_stderr": 0.039194155450484096,
"acc_norm": 0.5337423312883436,
"acc_norm_stderr": 0.039194155450484096
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.49691358024691357,
"acc_stderr": 0.027820214158594384,
"acc_norm": 0.49691358024691357,
"acc_norm_stderr": 0.027820214158594384
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.26,
"acc_stderr": 0.04408440022768078,
"acc_norm": 0.26,
"acc_norm_stderr": 0.04408440022768078
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.5595854922279793,
"acc_stderr": 0.035827245300360945,
"acc_norm": 0.5595854922279793,
"acc_norm_stderr": 0.035827245300360945
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.04434600701584925,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.04434600701584925
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.48440366972477067,
"acc_stderr": 0.02142689153920805,
"acc_norm": 0.48440366972477067,
"acc_norm_stderr": 0.02142689153920805
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.042163702135578345,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.042163702135578345
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.5,
"acc_stderr": 0.028629916715693413,
"acc_norm": 0.5,
"acc_norm_stderr": 0.028629916715693413
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.44,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.44,
"acc_norm_stderr": 0.04988876515698589
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6694214876033058,
"acc_stderr": 0.04294340845212093,
"acc_norm": 0.6694214876033058,
"acc_norm_stderr": 0.04294340845212093
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.4276315789473684,
"acc_stderr": 0.04026097083296558,
"acc_norm": 0.4276315789473684,
"acc_norm_stderr": 0.04026097083296558
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.39215686274509803,
"acc_stderr": 0.019751726508762633,
"acc_norm": 0.39215686274509803,
"acc_norm_stderr": 0.019751726508762633
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.36524822695035464,
"acc_stderr": 0.028723863853281278,
"acc_norm": 0.36524822695035464,
"acc_norm_stderr": 0.028723863853281278
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.35714285714285715,
"acc_stderr": 0.04547960999764376,
"acc_norm": 0.35714285714285715,
"acc_norm_stderr": 0.04547960999764376
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.4027777777777778,
"acc_stderr": 0.03344887382997865,
"acc_norm": 0.4027777777777778,
"acc_norm_stderr": 0.03344887382997865
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2424581005586592,
"acc_stderr": 0.01433352205921789,
"acc_norm": 0.2424581005586592,
"acc_norm_stderr": 0.01433352205921789
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.41,
"acc_stderr": 0.049431107042371025,
"acc_norm": 0.41,
"acc_norm_stderr": 0.049431107042371025
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.56,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.56,
"acc_norm_stderr": 0.04988876515698589
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.34191176470588236,
"acc_stderr": 0.028814722422254187,
"acc_norm": 0.34191176470588236,
"acc_norm_stderr": 0.028814722422254187
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.4448979591836735,
"acc_stderr": 0.031814251181977865,
"acc_norm": 0.4448979591836735,
"acc_norm_stderr": 0.031814251181977865
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.6075949367088608,
"acc_stderr": 0.0317847187456473,
"acc_norm": 0.6075949367088608,
"acc_norm_stderr": 0.0317847187456473
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3435462842242503,
"acc_stderr": 0.012128961174190161,
"acc_norm": 0.3435462842242503,
"acc_norm_stderr": 0.012128961174190161
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.44607843137254904,
"acc_stderr": 0.034888454513049734,
"acc_norm": 0.44607843137254904,
"acc_norm_stderr": 0.034888454513049734
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.503030303030303,
"acc_stderr": 0.03904272341431855,
"acc_norm": 0.503030303030303,
"acc_norm_stderr": 0.03904272341431855
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.28151774785801714,
"mc1_stderr": 0.01574402724825605,
"mc2": 0.46644130422803615,
"mc2_stderr": 0.015463171251968822
},
"harness|ko_commongen_v2|2": {
"acc": 0.41912632821723733,
"acc_stderr": 0.016963995010862792,
"acc_norm": 0.4462809917355372,
"acc_norm_stderr": 0.017090852631668332
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "Ja3ck/Mistral-instruct-DPO-Y24-v2",
"model_sha": "15acc5b8edbcbfda168710c8764a4d13fc98fd05",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}