|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.25426621160409557, |
|
"acc_stderr": 0.012724999945157741, |
|
"acc_norm": 0.31143344709897613, |
|
"acc_norm_stderr": 0.013532472099850944 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3311093407687712, |
|
"acc_stderr": 0.004696505101217403, |
|
"acc_norm": 0.40290778729336785, |
|
"acc_norm_stderr": 0.004894801119898608 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.27485380116959063, |
|
"acc_stderr": 0.03424042924691583, |
|
"acc_norm": 0.27485380116959063, |
|
"acc_norm_stderr": 0.03424042924691583 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.2621359223300971, |
|
"acc_stderr": 0.04354631077260597, |
|
"acc_norm": 0.2621359223300971, |
|
"acc_norm_stderr": 0.04354631077260597 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.32567049808429116, |
|
"acc_stderr": 0.016757989458549675, |
|
"acc_norm": 0.32567049808429116, |
|
"acc_norm_stderr": 0.016757989458549675 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.2962962962962963, |
|
"acc_stderr": 0.03944624162501117, |
|
"acc_norm": 0.2962962962962963, |
|
"acc_norm_stderr": 0.03944624162501117 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.33617021276595743, |
|
"acc_stderr": 0.030881618520676942, |
|
"acc_norm": 0.33617021276595743, |
|
"acc_norm_stderr": 0.030881618520676942 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.2710843373493976, |
|
"acc_stderr": 0.03460579907553026, |
|
"acc_norm": 0.2710843373493976, |
|
"acc_norm_stderr": 0.03460579907553026 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.3022508038585209, |
|
"acc_stderr": 0.026082700695399662, |
|
"acc_norm": 0.3022508038585209, |
|
"acc_norm_stderr": 0.026082700695399662 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.36771300448430494, |
|
"acc_stderr": 0.03236198350928275, |
|
"acc_norm": 0.36771300448430494, |
|
"acc_norm_stderr": 0.03236198350928275 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.3053435114503817, |
|
"acc_stderr": 0.0403931497872456, |
|
"acc_norm": 0.3053435114503817, |
|
"acc_norm_stderr": 0.0403931497872456 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252606, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252606 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.2777777777777778, |
|
"acc_stderr": 0.03191178226713545, |
|
"acc_norm": 0.2777777777777778, |
|
"acc_norm_stderr": 0.03191178226713545 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.2482758620689655, |
|
"acc_stderr": 0.03600105692727772, |
|
"acc_norm": 0.2482758620689655, |
|
"acc_norm_stderr": 0.03600105692727772 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.27450980392156865, |
|
"acc_stderr": 0.044405219061793275, |
|
"acc_norm": 0.27450980392156865, |
|
"acc_norm_stderr": 0.044405219061793275 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.3403361344537815, |
|
"acc_stderr": 0.03077805742293167, |
|
"acc_norm": 0.3403361344537815, |
|
"acc_norm_stderr": 0.03077805742293167 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.36666666666666664, |
|
"acc_stderr": 0.024433016466052455, |
|
"acc_norm": 0.36666666666666664, |
|
"acc_norm_stderr": 0.024433016466052455 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.04020151261036845, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.04020151261036845 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.045604802157206845, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.045604802157206845 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.2962962962962963, |
|
"acc_stderr": 0.04414343666854932, |
|
"acc_norm": 0.2962962962962963, |
|
"acc_norm_stderr": 0.04414343666854932 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.18226600985221675, |
|
"acc_stderr": 0.02716334085964515, |
|
"acc_norm": 0.18226600985221675, |
|
"acc_norm_stderr": 0.02716334085964515 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.2838709677419355, |
|
"acc_stderr": 0.025649381063029258, |
|
"acc_norm": 0.2838709677419355, |
|
"acc_norm_stderr": 0.025649381063029258 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.23931623931623933, |
|
"acc_stderr": 0.027951826808924333, |
|
"acc_norm": 0.23931623931623933, |
|
"acc_norm_stderr": 0.027951826808924333 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.24528301886792453, |
|
"acc_stderr": 0.02648035717989569, |
|
"acc_norm": 0.24528301886792453, |
|
"acc_norm_stderr": 0.02648035717989569 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.3090909090909091, |
|
"acc_stderr": 0.044262946482000985, |
|
"acc_norm": 0.3090909090909091, |
|
"acc_norm_stderr": 0.044262946482000985 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.24074074074074073, |
|
"acc_stderr": 0.026067159222275798, |
|
"acc_norm": 0.24074074074074073, |
|
"acc_norm_stderr": 0.026067159222275798 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2980132450331126, |
|
"acc_stderr": 0.037345356767871984, |
|
"acc_norm": 0.2980132450331126, |
|
"acc_norm_stderr": 0.037345356767871984 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.24875621890547264, |
|
"acc_stderr": 0.030567675938916718, |
|
"acc_norm": 0.24875621890547264, |
|
"acc_norm_stderr": 0.030567675938916718 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.2543352601156069, |
|
"acc_stderr": 0.0332055644308557, |
|
"acc_norm": 0.2543352601156069, |
|
"acc_norm_stderr": 0.0332055644308557 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.2619047619047619, |
|
"acc_stderr": 0.022644212615525218, |
|
"acc_norm": 0.2619047619047619, |
|
"acc_norm_stderr": 0.022644212615525218 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.24305555555555555, |
|
"acc_stderr": 0.03586879280080341, |
|
"acc_norm": 0.24305555555555555, |
|
"acc_norm_stderr": 0.03586879280080341 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.040201512610368445, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.040201512610368445 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.39, |
|
"acc_stderr": 0.04902071300001974, |
|
"acc_norm": 0.39, |
|
"acc_norm_stderr": 0.04902071300001974 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.2398843930635838, |
|
"acc_stderr": 0.022989592543123567, |
|
"acc_norm": 0.2398843930635838, |
|
"acc_norm_stderr": 0.022989592543123567 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.2331288343558282, |
|
"acc_stderr": 0.0332201579577674, |
|
"acc_norm": 0.2331288343558282, |
|
"acc_norm_stderr": 0.0332201579577674 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.30246913580246915, |
|
"acc_stderr": 0.025557653981868055, |
|
"acc_norm": 0.30246913580246915, |
|
"acc_norm_stderr": 0.025557653981868055 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.23, |
|
"acc_stderr": 0.04229525846816505, |
|
"acc_norm": 0.23, |
|
"acc_norm_stderr": 0.04229525846816505 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.33678756476683935, |
|
"acc_stderr": 0.03410780251836183, |
|
"acc_norm": 0.33678756476683935, |
|
"acc_norm_stderr": 0.03410780251836183 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.24561403508771928, |
|
"acc_stderr": 0.04049339297748141, |
|
"acc_norm": 0.24561403508771928, |
|
"acc_norm_stderr": 0.04049339297748141 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.28623853211009176, |
|
"acc_stderr": 0.019379436628919982, |
|
"acc_norm": 0.28623853211009176, |
|
"acc_norm_stderr": 0.019379436628919982 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.15873015873015872, |
|
"acc_stderr": 0.03268454013011744, |
|
"acc_norm": 0.15873015873015872, |
|
"acc_norm_stderr": 0.03268454013011744 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.28104575163398693, |
|
"acc_stderr": 0.025738854797818737, |
|
"acc_norm": 0.28104575163398693, |
|
"acc_norm_stderr": 0.025738854797818737 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.2727272727272727, |
|
"acc_stderr": 0.04065578140908705, |
|
"acc_norm": 0.2727272727272727, |
|
"acc_norm_stderr": 0.04065578140908705 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.21052631578947367, |
|
"acc_stderr": 0.03317672787533157, |
|
"acc_norm": 0.21052631578947367, |
|
"acc_norm_stderr": 0.03317672787533157 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.21895424836601307, |
|
"acc_stderr": 0.016729937565537534, |
|
"acc_norm": 0.21895424836601307, |
|
"acc_norm_stderr": 0.016729937565537534 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.2730496453900709, |
|
"acc_stderr": 0.026577860943307857, |
|
"acc_norm": 0.2730496453900709, |
|
"acc_norm_stderr": 0.026577860943307857 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04109974682633932, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04109974682633932 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.4675925925925926, |
|
"acc_stderr": 0.03402801581358966, |
|
"acc_norm": 0.4675925925925926, |
|
"acc_norm_stderr": 0.03402801581358966 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.27039106145251396, |
|
"acc_stderr": 0.014854993938010088, |
|
"acc_norm": 0.27039106145251396, |
|
"acc_norm_stderr": 0.014854993938010088 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621504, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621504 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252605, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252605 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.4485294117647059, |
|
"acc_stderr": 0.030211479609121593, |
|
"acc_norm": 0.4485294117647059, |
|
"acc_norm_stderr": 0.030211479609121593 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.4163265306122449, |
|
"acc_stderr": 0.03155782816556164, |
|
"acc_norm": 0.4163265306122449, |
|
"acc_norm_stderr": 0.03155782816556164 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.26582278481012656, |
|
"acc_stderr": 0.02875679962965834, |
|
"acc_norm": 0.26582278481012656, |
|
"acc_norm_stderr": 0.02875679962965834 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.2516297262059974, |
|
"acc_stderr": 0.011083276280441904, |
|
"acc_norm": 0.2516297262059974, |
|
"acc_norm_stderr": 0.011083276280441904 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.2647058823529412, |
|
"acc_stderr": 0.030964517926923393, |
|
"acc_norm": 0.2647058823529412, |
|
"acc_norm_stderr": 0.030964517926923393 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.21818181818181817, |
|
"acc_stderr": 0.03225078108306289, |
|
"acc_norm": 0.21818181818181817, |
|
"acc_norm_stderr": 0.03225078108306289 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.27050183598531213, |
|
"mc1_stderr": 0.015550778332842892, |
|
"mc2": 0.4372897231981029, |
|
"mc2_stderr": 0.01541061408460767 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.22904368358913813, |
|
"acc_stderr": 0.01444737227725382, |
|
"acc_norm": 0.2715466351829988, |
|
"acc_norm_stderr": 0.015291071117310378 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "JYKIM-AI/Mistral-7B-SFT", |
|
"model_sha": "ea2e2395fcf295d293c236b0228ece970cca5aba", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |