|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.48378839590443684, |
|
"acc_stderr": 0.014603708567414947, |
|
"acc_norm": 0.5255972696245734, |
|
"acc_norm_stderr": 0.014592230885298962 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.4849631547500498, |
|
"acc_stderr": 0.004987524454849698, |
|
"acc_norm": 0.6667994423421629, |
|
"acc_norm_stderr": 0.004703942346762255 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.6081871345029239, |
|
"acc_stderr": 0.037439798259263996, |
|
"acc_norm": 0.6081871345029239, |
|
"acc_norm_stderr": 0.037439798259263996 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.7087378640776699, |
|
"acc_stderr": 0.044986763205729224, |
|
"acc_norm": 0.7087378640776699, |
|
"acc_norm_stderr": 0.044986763205729224 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.6309067688378033, |
|
"acc_stderr": 0.01725628310912463, |
|
"acc_norm": 0.6309067688378033, |
|
"acc_norm_stderr": 0.01725628310912463 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.362962962962963, |
|
"acc_stderr": 0.04153948404742401, |
|
"acc_norm": 0.362962962962963, |
|
"acc_norm_stderr": 0.04153948404742401 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.04020151261036846, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.04020151261036846 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.43829787234042555, |
|
"acc_stderr": 0.032436186361081025, |
|
"acc_norm": 0.43829787234042555, |
|
"acc_norm_stderr": 0.032436186361081025 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.42168674698795183, |
|
"acc_stderr": 0.03844453181770917, |
|
"acc_norm": 0.42168674698795183, |
|
"acc_norm_stderr": 0.03844453181770917 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.5691318327974276, |
|
"acc_stderr": 0.028125340983972714, |
|
"acc_norm": 0.5691318327974276, |
|
"acc_norm_stderr": 0.028125340983972714 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.515695067264574, |
|
"acc_stderr": 0.0335412657542081, |
|
"acc_norm": 0.515695067264574, |
|
"acc_norm_stderr": 0.0335412657542081 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.549618320610687, |
|
"acc_stderr": 0.04363643698524779, |
|
"acc_norm": 0.549618320610687, |
|
"acc_norm_stderr": 0.04363643698524779 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.44, |
|
"acc_stderr": 0.0498887651569859, |
|
"acc_norm": 0.44, |
|
"acc_norm_stderr": 0.0498887651569859 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.6868686868686869, |
|
"acc_stderr": 0.033042050878136525, |
|
"acc_norm": 0.6868686868686869, |
|
"acc_norm_stderr": 0.033042050878136525 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.47586206896551725, |
|
"acc_stderr": 0.0416180850350153, |
|
"acc_norm": 0.47586206896551725, |
|
"acc_norm_stderr": 0.0416180850350153 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.04690650298201943, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.04690650298201943 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.6134453781512605, |
|
"acc_stderr": 0.03163145807552378, |
|
"acc_norm": 0.6134453781512605, |
|
"acc_norm_stderr": 0.03163145807552378 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.5282051282051282, |
|
"acc_stderr": 0.02531063925493386, |
|
"acc_norm": 0.5282051282051282, |
|
"acc_norm_stderr": 0.02531063925493386 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.59, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.59, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6018518518518519, |
|
"acc_stderr": 0.047323326159788126, |
|
"acc_norm": 0.6018518518518519, |
|
"acc_norm_stderr": 0.047323326159788126 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.43842364532019706, |
|
"acc_stderr": 0.03491207857486519, |
|
"acc_norm": 0.43842364532019706, |
|
"acc_norm_stderr": 0.03491207857486519 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.567741935483871, |
|
"acc_stderr": 0.028181739720019406, |
|
"acc_norm": 0.567741935483871, |
|
"acc_norm_stderr": 0.028181739720019406 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7350427350427351, |
|
"acc_stderr": 0.028911208802749472, |
|
"acc_norm": 0.7350427350427351, |
|
"acc_norm_stderr": 0.028911208802749472 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.5207547169811321, |
|
"acc_stderr": 0.03074634997572347, |
|
"acc_norm": 0.5207547169811321, |
|
"acc_norm_stderr": 0.03074634997572347 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5636363636363636, |
|
"acc_stderr": 0.04750185058907296, |
|
"acc_norm": 0.5636363636363636, |
|
"acc_norm_stderr": 0.04750185058907296 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.35555555555555557, |
|
"acc_stderr": 0.029185714949857406, |
|
"acc_norm": 0.35555555555555557, |
|
"acc_norm_stderr": 0.029185714949857406 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3443708609271523, |
|
"acc_stderr": 0.038796870240733264, |
|
"acc_norm": 0.3443708609271523, |
|
"acc_norm_stderr": 0.038796870240733264 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6169154228855721, |
|
"acc_stderr": 0.0343751933733825, |
|
"acc_norm": 0.6169154228855721, |
|
"acc_norm_stderr": 0.0343751933733825 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.45664739884393063, |
|
"acc_stderr": 0.03798106566014498, |
|
"acc_norm": 0.45664739884393063, |
|
"acc_norm_stderr": 0.03798106566014498 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.4523809523809524, |
|
"acc_stderr": 0.025634258115554965, |
|
"acc_norm": 0.4523809523809524, |
|
"acc_norm_stderr": 0.025634258115554965 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.5138888888888888, |
|
"acc_stderr": 0.041795966175810016, |
|
"acc_norm": 0.5138888888888888, |
|
"acc_norm_stderr": 0.041795966175810016 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.39, |
|
"acc_stderr": 0.04902071300001975, |
|
"acc_norm": 0.39, |
|
"acc_norm_stderr": 0.04902071300001975 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.71, |
|
"acc_stderr": 0.04560480215720684, |
|
"acc_norm": 0.71, |
|
"acc_norm_stderr": 0.04560480215720684 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.546242774566474, |
|
"acc_stderr": 0.026803720583206184, |
|
"acc_norm": 0.546242774566474, |
|
"acc_norm_stderr": 0.026803720583206184 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.44785276073619634, |
|
"acc_stderr": 0.03906947479456601, |
|
"acc_norm": 0.44785276073619634, |
|
"acc_norm_stderr": 0.03906947479456601 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.5617283950617284, |
|
"acc_stderr": 0.027607914087400477, |
|
"acc_norm": 0.5617283950617284, |
|
"acc_norm_stderr": 0.027607914087400477 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.047258156262526045, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.6839378238341969, |
|
"acc_stderr": 0.033553973696861736, |
|
"acc_norm": 0.6839378238341969, |
|
"acc_norm_stderr": 0.033553973696861736 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.5263157894736842, |
|
"acc_stderr": 0.046970851366478626, |
|
"acc_norm": 0.5263157894736842, |
|
"acc_norm_stderr": 0.046970851366478626 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.6293577981651376, |
|
"acc_stderr": 0.02070745816435298, |
|
"acc_norm": 0.6293577981651376, |
|
"acc_norm_stderr": 0.02070745816435298 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.4126984126984127, |
|
"acc_stderr": 0.04403438954768176, |
|
"acc_norm": 0.4126984126984127, |
|
"acc_norm_stderr": 0.04403438954768176 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5326797385620915, |
|
"acc_stderr": 0.02856869975222587, |
|
"acc_norm": 0.5326797385620915, |
|
"acc_norm_stderr": 0.02856869975222587 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.52, |
|
"acc_stderr": 0.05021167315686779, |
|
"acc_norm": 0.52, |
|
"acc_norm_stderr": 0.05021167315686779 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.628099173553719, |
|
"acc_stderr": 0.044120158066245044, |
|
"acc_norm": 0.628099173553719, |
|
"acc_norm_stderr": 0.044120158066245044 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5855263157894737, |
|
"acc_stderr": 0.04008973785779206, |
|
"acc_norm": 0.5855263157894737, |
|
"acc_norm_stderr": 0.04008973785779206 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.48366013071895425, |
|
"acc_stderr": 0.020217030653186453, |
|
"acc_norm": 0.48366013071895425, |
|
"acc_norm_stderr": 0.020217030653186453 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.375886524822695, |
|
"acc_stderr": 0.028893955412115886, |
|
"acc_norm": 0.375886524822695, |
|
"acc_norm_stderr": 0.028893955412115886 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.3482142857142857, |
|
"acc_stderr": 0.04521829902833585, |
|
"acc_norm": 0.3482142857142857, |
|
"acc_norm_stderr": 0.04521829902833585 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.4861111111111111, |
|
"acc_stderr": 0.03408655867977749, |
|
"acc_norm": 0.4861111111111111, |
|
"acc_norm_stderr": 0.03408655867977749 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2871508379888268, |
|
"acc_stderr": 0.01513160884996376, |
|
"acc_norm": 0.2871508379888268, |
|
"acc_norm_stderr": 0.01513160884996376 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.43, |
|
"acc_stderr": 0.049756985195624284, |
|
"acc_norm": 0.43, |
|
"acc_norm_stderr": 0.049756985195624284 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.63, |
|
"acc_stderr": 0.048523658709391, |
|
"acc_norm": 0.63, |
|
"acc_norm_stderr": 0.048523658709391 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.49264705882352944, |
|
"acc_stderr": 0.030369552523902173, |
|
"acc_norm": 0.49264705882352944, |
|
"acc_norm_stderr": 0.030369552523902173 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.5265306122448979, |
|
"acc_stderr": 0.03196412734523272, |
|
"acc_norm": 0.5265306122448979, |
|
"acc_norm_stderr": 0.03196412734523272 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.7552742616033755, |
|
"acc_stderr": 0.027985699387036416, |
|
"acc_norm": 0.7552742616033755, |
|
"acc_norm_stderr": 0.027985699387036416 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.4061277705345502, |
|
"acc_stderr": 0.012543154588412927, |
|
"acc_norm": 0.4061277705345502, |
|
"acc_norm_stderr": 0.012543154588412927 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.6323529411764706, |
|
"acc_stderr": 0.03384132045674118, |
|
"acc_norm": 0.6323529411764706, |
|
"acc_norm_stderr": 0.03384132045674118 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.6787878787878788, |
|
"acc_stderr": 0.03646204963253812, |
|
"acc_norm": 0.6787878787878788, |
|
"acc_norm_stderr": 0.03646204963253812 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.4112607099143207, |
|
"mc1_stderr": 0.017225627083660856, |
|
"mc2": 0.5927070170654897, |
|
"mc2_stderr": 0.0162035946763131 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5962219598583235, |
|
"acc_stderr": 0.016869031540298625, |
|
"acc_norm": 0.6103896103896104, |
|
"acc_norm_stderr": 0.0167661616718935 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "Edentns/DataVortexS-10.7B-dpo-v1.8", |
|
"model_sha": "91c4b182d2c6d514b5f0205001e7ca4e37cfbe60", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |