results / Edentns /DataVortexS-10.7B-dpo-v1.2 /result_2024-01-25 03:32:56.json
open-ko-llm-bot's picture
Add results for 2024-01-25 03:32:56
cc403d6 verified
raw
history blame
17.8 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.4786689419795222,
"acc_stderr": 0.014598087973127108,
"acc_norm": 0.5273037542662116,
"acc_norm_stderr": 0.014589589101986001
},
"harness|ko_hellaswag|10": {
"acc": 0.47739494124676357,
"acc_stderr": 0.004984679359375627,
"acc_norm": 0.6482772356104362,
"acc_norm_stderr": 0.0047653207849021086
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.6666666666666666,
"acc_stderr": 0.036155076303109344,
"acc_norm": 0.6666666666666666,
"acc_norm_stderr": 0.036155076303109344
},
"harness|ko_mmlu_management|5": {
"acc": 0.6407766990291263,
"acc_stderr": 0.04750458399041696,
"acc_norm": 0.6407766990291263,
"acc_norm_stderr": 0.04750458399041696
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.6206896551724138,
"acc_stderr": 0.01735126811754445,
"acc_norm": 0.6206896551724138,
"acc_norm_stderr": 0.01735126811754445
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.4222222222222222,
"acc_stderr": 0.04266763404099582,
"acc_norm": 0.4222222222222222,
"acc_norm_stderr": 0.04266763404099582
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.27,
"acc_stderr": 0.044619604333847394,
"acc_norm": 0.27,
"acc_norm_stderr": 0.044619604333847394
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.4127659574468085,
"acc_stderr": 0.03218471141400351,
"acc_norm": 0.4127659574468085,
"acc_norm_stderr": 0.03218471141400351
},
"harness|ko_mmlu_virology|5": {
"acc": 0.4397590361445783,
"acc_stderr": 0.03864139923699121,
"acc_norm": 0.4397590361445783,
"acc_norm_stderr": 0.03864139923699121
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.572347266881029,
"acc_stderr": 0.02809924077580957,
"acc_norm": 0.572347266881029,
"acc_norm_stderr": 0.02809924077580957
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.547085201793722,
"acc_stderr": 0.03340867501923324,
"acc_norm": 0.547085201793722,
"acc_norm_stderr": 0.03340867501923324
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.549618320610687,
"acc_stderr": 0.04363643698524779,
"acc_norm": 0.549618320610687,
"acc_norm_stderr": 0.04363643698524779
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.45,
"acc_stderr": 0.04999999999999999,
"acc_norm": 0.45,
"acc_norm_stderr": 0.04999999999999999
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.7323232323232324,
"acc_stderr": 0.03154449888270286,
"acc_norm": 0.7323232323232324,
"acc_norm_stderr": 0.03154449888270286
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.42758620689655175,
"acc_stderr": 0.041227371113703316,
"acc_norm": 0.42758620689655175,
"acc_norm_stderr": 0.041227371113703316
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.3235294117647059,
"acc_stderr": 0.04655010411319616,
"acc_norm": 0.3235294117647059,
"acc_norm_stderr": 0.04655010411319616
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.6680672268907563,
"acc_stderr": 0.03058869701378364,
"acc_norm": 0.6680672268907563,
"acc_norm_stderr": 0.03058869701378364
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.5871794871794872,
"acc_stderr": 0.024962683564331793,
"acc_norm": 0.5871794871794872,
"acc_norm_stderr": 0.024962683564331793
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.59,
"acc_stderr": 0.049431107042371025,
"acc_norm": 0.59,
"acc_norm_stderr": 0.049431107042371025
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.38,
"acc_stderr": 0.04878317312145633,
"acc_norm": 0.38,
"acc_norm_stderr": 0.04878317312145633
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.6296296296296297,
"acc_stderr": 0.04668408033024931,
"acc_norm": 0.6296296296296297,
"acc_norm_stderr": 0.04668408033024931
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.3793103448275862,
"acc_stderr": 0.034139638059062345,
"acc_norm": 0.3793103448275862,
"acc_norm_stderr": 0.034139638059062345
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.5645161290322581,
"acc_stderr": 0.028206225591502734,
"acc_norm": 0.5645161290322581,
"acc_norm_stderr": 0.028206225591502734
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.782051282051282,
"acc_stderr": 0.027046857630716663,
"acc_norm": 0.782051282051282,
"acc_norm_stderr": 0.027046857630716663
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.5320754716981132,
"acc_stderr": 0.030709486992556545,
"acc_norm": 0.5320754716981132,
"acc_norm_stderr": 0.030709486992556545
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.6,
"acc_stderr": 0.0469237132203465,
"acc_norm": 0.6,
"acc_norm_stderr": 0.0469237132203465
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.3888888888888889,
"acc_stderr": 0.029723278961476664,
"acc_norm": 0.3888888888888889,
"acc_norm_stderr": 0.029723278961476664
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.36423841059602646,
"acc_stderr": 0.03929111781242742,
"acc_norm": 0.36423841059602646,
"acc_norm_stderr": 0.03929111781242742
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.6069651741293532,
"acc_stderr": 0.0345368246603156,
"acc_norm": 0.6069651741293532,
"acc_norm_stderr": 0.0345368246603156
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.4797687861271676,
"acc_stderr": 0.03809342081273958,
"acc_norm": 0.4797687861271676,
"acc_norm_stderr": 0.03809342081273958
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.4523809523809524,
"acc_stderr": 0.02563425811555496,
"acc_norm": 0.4523809523809524,
"acc_norm_stderr": 0.02563425811555496
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.5555555555555556,
"acc_stderr": 0.04155319955593146,
"acc_norm": 0.5555555555555556,
"acc_norm_stderr": 0.04155319955593146
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.33,
"acc_stderr": 0.047258156262526045,
"acc_norm": 0.33,
"acc_norm_stderr": 0.047258156262526045
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.71,
"acc_stderr": 0.04560480215720683,
"acc_norm": 0.71,
"acc_norm_stderr": 0.04560480215720683
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.5809248554913294,
"acc_stderr": 0.02656417811142263,
"acc_norm": 0.5809248554913294,
"acc_norm_stderr": 0.02656417811142263
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.4723926380368098,
"acc_stderr": 0.0392237829061099,
"acc_norm": 0.4723926380368098,
"acc_norm_stderr": 0.0392237829061099
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.6172839506172839,
"acc_stderr": 0.027044538138402602,
"acc_norm": 0.6172839506172839,
"acc_norm_stderr": 0.027044538138402602
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.32,
"acc_stderr": 0.046882617226215034,
"acc_norm": 0.32,
"acc_norm_stderr": 0.046882617226215034
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.7409326424870466,
"acc_stderr": 0.03161877917935411,
"acc_norm": 0.7409326424870466,
"acc_norm_stderr": 0.03161877917935411
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.41228070175438597,
"acc_stderr": 0.046306532033665956,
"acc_norm": 0.41228070175438597,
"acc_norm_stderr": 0.046306532033665956
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.671559633027523,
"acc_stderr": 0.02013590279729841,
"acc_norm": 0.671559633027523,
"acc_norm_stderr": 0.02013590279729841
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.48412698412698413,
"acc_stderr": 0.04469881854072606,
"acc_norm": 0.48412698412698413,
"acc_norm_stderr": 0.04469881854072606
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.5686274509803921,
"acc_stderr": 0.02835895631342355,
"acc_norm": 0.5686274509803921,
"acc_norm_stderr": 0.02835895631342355
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.59,
"acc_stderr": 0.049431107042371025,
"acc_norm": 0.59,
"acc_norm_stderr": 0.049431107042371025
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6611570247933884,
"acc_stderr": 0.0432076780753667,
"acc_norm": 0.6611570247933884,
"acc_norm_stderr": 0.0432076780753667
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.5986842105263158,
"acc_stderr": 0.039889037033362836,
"acc_norm": 0.5986842105263158,
"acc_norm_stderr": 0.039889037033362836
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.49836601307189543,
"acc_stderr": 0.020227726838150117,
"acc_norm": 0.49836601307189543,
"acc_norm_stderr": 0.020227726838150117
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.3829787234042553,
"acc_stderr": 0.02899908090480619,
"acc_norm": 0.3829787234042553,
"acc_norm_stderr": 0.02899908090480619
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.375,
"acc_stderr": 0.04595091388086298,
"acc_norm": 0.375,
"acc_norm_stderr": 0.04595091388086298
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.5277777777777778,
"acc_stderr": 0.0340470532865388,
"acc_norm": 0.5277777777777778,
"acc_norm_stderr": 0.0340470532865388
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.25921787709497207,
"acc_stderr": 0.014655780837497717,
"acc_norm": 0.25921787709497207,
"acc_norm_stderr": 0.014655780837497717
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.51,
"acc_stderr": 0.05024183937956911,
"acc_norm": 0.51,
"acc_norm_stderr": 0.05024183937956911
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.67,
"acc_stderr": 0.04725815626252607,
"acc_norm": 0.67,
"acc_norm_stderr": 0.04725815626252607
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.5147058823529411,
"acc_stderr": 0.03035969707904611,
"acc_norm": 0.5147058823529411,
"acc_norm_stderr": 0.03035969707904611
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.5102040816326531,
"acc_stderr": 0.03200255347893782,
"acc_norm": 0.5102040816326531,
"acc_norm_stderr": 0.03200255347893782
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.759493670886076,
"acc_stderr": 0.02782078198114968,
"acc_norm": 0.759493670886076,
"acc_norm_stderr": 0.02782078198114968
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.42046936114732725,
"acc_stderr": 0.012607654553832707,
"acc_norm": 0.42046936114732725,
"acc_norm_stderr": 0.012607654553832707
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.6862745098039216,
"acc_stderr": 0.03256685484460388,
"acc_norm": 0.6862745098039216,
"acc_norm_stderr": 0.03256685484460388
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.696969696969697,
"acc_stderr": 0.035886248000917095,
"acc_norm": 0.696969696969697,
"acc_norm_stderr": 0.035886248000917095
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.41003671970624234,
"mc1_stderr": 0.017217844717449318,
"mc2": 0.583646136539335,
"mc2_stderr": 0.016204158336690234
},
"harness|ko_commongen_v2|2": {
"acc": 0.5242030696576151,
"acc_stderr": 0.017170202466520748,
"acc_norm": 0.5371900826446281,
"acc_norm_stderr": 0.017142736117643304
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "Edentns/DataVortexS-10.7B-dpo-v1.2",
"model_sha": "4620f8491ea4cc4b2d47f41ee7a6f32a6b1bfa01",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}