|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.4786689419795222, |
|
"acc_stderr": 0.014598087973127108, |
|
"acc_norm": 0.5273037542662116, |
|
"acc_norm_stderr": 0.014589589101986001 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.47739494124676357, |
|
"acc_stderr": 0.004984679359375627, |
|
"acc_norm": 0.6482772356104362, |
|
"acc_norm_stderr": 0.0047653207849021086 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.6666666666666666, |
|
"acc_stderr": 0.036155076303109344, |
|
"acc_norm": 0.6666666666666666, |
|
"acc_norm_stderr": 0.036155076303109344 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6407766990291263, |
|
"acc_stderr": 0.04750458399041696, |
|
"acc_norm": 0.6407766990291263, |
|
"acc_norm_stderr": 0.04750458399041696 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.6206896551724138, |
|
"acc_stderr": 0.01735126811754445, |
|
"acc_norm": 0.6206896551724138, |
|
"acc_norm_stderr": 0.01735126811754445 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4222222222222222, |
|
"acc_stderr": 0.04266763404099582, |
|
"acc_norm": 0.4222222222222222, |
|
"acc_norm_stderr": 0.04266763404099582 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.044619604333847394, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.044619604333847394 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.4127659574468085, |
|
"acc_stderr": 0.03218471141400351, |
|
"acc_norm": 0.4127659574468085, |
|
"acc_norm_stderr": 0.03218471141400351 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.4397590361445783, |
|
"acc_stderr": 0.03864139923699121, |
|
"acc_norm": 0.4397590361445783, |
|
"acc_norm_stderr": 0.03864139923699121 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.572347266881029, |
|
"acc_stderr": 0.02809924077580957, |
|
"acc_norm": 0.572347266881029, |
|
"acc_norm_stderr": 0.02809924077580957 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.547085201793722, |
|
"acc_stderr": 0.03340867501923324, |
|
"acc_norm": 0.547085201793722, |
|
"acc_norm_stderr": 0.03340867501923324 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.549618320610687, |
|
"acc_stderr": 0.04363643698524779, |
|
"acc_norm": 0.549618320610687, |
|
"acc_norm_stderr": 0.04363643698524779 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.45, |
|
"acc_stderr": 0.04999999999999999, |
|
"acc_norm": 0.45, |
|
"acc_norm_stderr": 0.04999999999999999 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.7323232323232324, |
|
"acc_stderr": 0.03154449888270286, |
|
"acc_norm": 0.7323232323232324, |
|
"acc_norm_stderr": 0.03154449888270286 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.42758620689655175, |
|
"acc_stderr": 0.041227371113703316, |
|
"acc_norm": 0.42758620689655175, |
|
"acc_norm_stderr": 0.041227371113703316 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.3235294117647059, |
|
"acc_stderr": 0.04655010411319616, |
|
"acc_norm": 0.3235294117647059, |
|
"acc_norm_stderr": 0.04655010411319616 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.6680672268907563, |
|
"acc_stderr": 0.03058869701378364, |
|
"acc_norm": 0.6680672268907563, |
|
"acc_norm_stderr": 0.03058869701378364 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.5871794871794872, |
|
"acc_stderr": 0.024962683564331793, |
|
"acc_norm": 0.5871794871794872, |
|
"acc_norm_stderr": 0.024962683564331793 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.59, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.59, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.04878317312145633, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.04878317312145633 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6296296296296297, |
|
"acc_stderr": 0.04668408033024931, |
|
"acc_norm": 0.6296296296296297, |
|
"acc_norm_stderr": 0.04668408033024931 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.3793103448275862, |
|
"acc_stderr": 0.034139638059062345, |
|
"acc_norm": 0.3793103448275862, |
|
"acc_norm_stderr": 0.034139638059062345 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.5645161290322581, |
|
"acc_stderr": 0.028206225591502734, |
|
"acc_norm": 0.5645161290322581, |
|
"acc_norm_stderr": 0.028206225591502734 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.782051282051282, |
|
"acc_stderr": 0.027046857630716663, |
|
"acc_norm": 0.782051282051282, |
|
"acc_norm_stderr": 0.027046857630716663 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.5320754716981132, |
|
"acc_stderr": 0.030709486992556545, |
|
"acc_norm": 0.5320754716981132, |
|
"acc_norm_stderr": 0.030709486992556545 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.6, |
|
"acc_stderr": 0.0469237132203465, |
|
"acc_norm": 0.6, |
|
"acc_norm_stderr": 0.0469237132203465 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3888888888888889, |
|
"acc_stderr": 0.029723278961476664, |
|
"acc_norm": 0.3888888888888889, |
|
"acc_norm_stderr": 0.029723278961476664 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.36423841059602646, |
|
"acc_stderr": 0.03929111781242742, |
|
"acc_norm": 0.36423841059602646, |
|
"acc_norm_stderr": 0.03929111781242742 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6069651741293532, |
|
"acc_stderr": 0.0345368246603156, |
|
"acc_norm": 0.6069651741293532, |
|
"acc_norm_stderr": 0.0345368246603156 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.4797687861271676, |
|
"acc_stderr": 0.03809342081273958, |
|
"acc_norm": 0.4797687861271676, |
|
"acc_norm_stderr": 0.03809342081273958 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.4523809523809524, |
|
"acc_stderr": 0.02563425811555496, |
|
"acc_norm": 0.4523809523809524, |
|
"acc_norm_stderr": 0.02563425811555496 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.5555555555555556, |
|
"acc_stderr": 0.04155319955593146, |
|
"acc_norm": 0.5555555555555556, |
|
"acc_norm_stderr": 0.04155319955593146 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.047258156262526045, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.71, |
|
"acc_stderr": 0.04560480215720683, |
|
"acc_norm": 0.71, |
|
"acc_norm_stderr": 0.04560480215720683 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5809248554913294, |
|
"acc_stderr": 0.02656417811142263, |
|
"acc_norm": 0.5809248554913294, |
|
"acc_norm_stderr": 0.02656417811142263 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.4723926380368098, |
|
"acc_stderr": 0.0392237829061099, |
|
"acc_norm": 0.4723926380368098, |
|
"acc_norm_stderr": 0.0392237829061099 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.6172839506172839, |
|
"acc_stderr": 0.027044538138402602, |
|
"acc_norm": 0.6172839506172839, |
|
"acc_norm_stderr": 0.027044538138402602 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.7409326424870466, |
|
"acc_stderr": 0.03161877917935411, |
|
"acc_norm": 0.7409326424870466, |
|
"acc_norm_stderr": 0.03161877917935411 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.41228070175438597, |
|
"acc_stderr": 0.046306532033665956, |
|
"acc_norm": 0.41228070175438597, |
|
"acc_norm_stderr": 0.046306532033665956 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.671559633027523, |
|
"acc_stderr": 0.02013590279729841, |
|
"acc_norm": 0.671559633027523, |
|
"acc_norm_stderr": 0.02013590279729841 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.48412698412698413, |
|
"acc_stderr": 0.04469881854072606, |
|
"acc_norm": 0.48412698412698413, |
|
"acc_norm_stderr": 0.04469881854072606 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5686274509803921, |
|
"acc_stderr": 0.02835895631342355, |
|
"acc_norm": 0.5686274509803921, |
|
"acc_norm_stderr": 0.02835895631342355 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.59, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.59, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6611570247933884, |
|
"acc_stderr": 0.0432076780753667, |
|
"acc_norm": 0.6611570247933884, |
|
"acc_norm_stderr": 0.0432076780753667 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5986842105263158, |
|
"acc_stderr": 0.039889037033362836, |
|
"acc_norm": 0.5986842105263158, |
|
"acc_norm_stderr": 0.039889037033362836 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.49836601307189543, |
|
"acc_stderr": 0.020227726838150117, |
|
"acc_norm": 0.49836601307189543, |
|
"acc_norm_stderr": 0.020227726838150117 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3829787234042553, |
|
"acc_stderr": 0.02899908090480619, |
|
"acc_norm": 0.3829787234042553, |
|
"acc_norm_stderr": 0.02899908090480619 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.375, |
|
"acc_stderr": 0.04595091388086298, |
|
"acc_norm": 0.375, |
|
"acc_norm_stderr": 0.04595091388086298 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.5277777777777778, |
|
"acc_stderr": 0.0340470532865388, |
|
"acc_norm": 0.5277777777777778, |
|
"acc_norm_stderr": 0.0340470532865388 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.25921787709497207, |
|
"acc_stderr": 0.014655780837497717, |
|
"acc_norm": 0.25921787709497207, |
|
"acc_norm_stderr": 0.014655780837497717 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.51, |
|
"acc_stderr": 0.05024183937956911, |
|
"acc_norm": 0.51, |
|
"acc_norm_stderr": 0.05024183937956911 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.67, |
|
"acc_stderr": 0.04725815626252607, |
|
"acc_norm": 0.67, |
|
"acc_norm_stderr": 0.04725815626252607 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.5147058823529411, |
|
"acc_stderr": 0.03035969707904611, |
|
"acc_norm": 0.5147058823529411, |
|
"acc_norm_stderr": 0.03035969707904611 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.5102040816326531, |
|
"acc_stderr": 0.03200255347893782, |
|
"acc_norm": 0.5102040816326531, |
|
"acc_norm_stderr": 0.03200255347893782 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.759493670886076, |
|
"acc_stderr": 0.02782078198114968, |
|
"acc_norm": 0.759493670886076, |
|
"acc_norm_stderr": 0.02782078198114968 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.42046936114732725, |
|
"acc_stderr": 0.012607654553832707, |
|
"acc_norm": 0.42046936114732725, |
|
"acc_norm_stderr": 0.012607654553832707 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.6862745098039216, |
|
"acc_stderr": 0.03256685484460388, |
|
"acc_norm": 0.6862745098039216, |
|
"acc_norm_stderr": 0.03256685484460388 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.696969696969697, |
|
"acc_stderr": 0.035886248000917095, |
|
"acc_norm": 0.696969696969697, |
|
"acc_norm_stderr": 0.035886248000917095 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.41003671970624234, |
|
"mc1_stderr": 0.017217844717449318, |
|
"mc2": 0.583646136539335, |
|
"mc2_stderr": 0.016204158336690234 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5242030696576151, |
|
"acc_stderr": 0.017170202466520748, |
|
"acc_norm": 0.5371900826446281, |
|
"acc_norm_stderr": 0.017142736117643304 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "Edentns/DataVortexS-10.7B-dpo-v1.2", |
|
"model_sha": "4620f8491ea4cc4b2d47f41ee7a6f32a6b1bfa01", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |