{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.24146757679180889, "acc_stderr": 0.01250656483973943, "acc_norm": 0.28668941979522183, "acc_norm_stderr": 0.013214986329274779 }, "harness|ko_hellaswag|10": { "acc": 0.3467436765584545, "acc_stderr": 0.004749606196363337, "acc_norm": 0.4356701852220673, "acc_norm_stderr": 0.004948310399746081 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.26900584795321636, "acc_stderr": 0.03401052620104089, "acc_norm": 0.26900584795321636, "acc_norm_stderr": 0.03401052620104089 }, "harness|ko_mmlu_management|5": { "acc": 0.20388349514563106, "acc_stderr": 0.039891398595317706, "acc_norm": 0.20388349514563106, "acc_norm_stderr": 0.039891398595317706 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.27458492975734355, "acc_stderr": 0.01595982993308404, "acc_norm": 0.27458492975734355, "acc_norm_stderr": 0.01595982993308404 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.24444444444444444, "acc_stderr": 0.037125378336148665, "acc_norm": 0.24444444444444444, "acc_norm_stderr": 0.037125378336148665 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.22, "acc_stderr": 0.04163331998932268, "acc_norm": 0.22, "acc_norm_stderr": 0.04163331998932268 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.2, "acc_stderr": 0.0261488180184245, "acc_norm": 0.2, "acc_norm_stderr": 0.0261488180184245 }, "harness|ko_mmlu_virology|5": { "acc": 0.2469879518072289, "acc_stderr": 0.03357351982064537, "acc_norm": 0.2469879518072289, "acc_norm_stderr": 0.03357351982064537 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.2958199356913183, "acc_stderr": 0.025922371788818774, "acc_norm": 0.2958199356913183, "acc_norm_stderr": 0.025922371788818774 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.21524663677130046, "acc_stderr": 0.027584066602208274, "acc_norm": 0.21524663677130046, "acc_norm_stderr": 0.027584066602208274 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.22900763358778625, "acc_stderr": 0.036853466317118506, "acc_norm": 0.22900763358778625, "acc_norm_stderr": 0.036853466317118506 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.22727272727272727, "acc_stderr": 0.0298575156733864, "acc_norm": 0.22727272727272727, "acc_norm_stderr": 0.0298575156733864 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.2620689655172414, "acc_stderr": 0.036646663372252565, "acc_norm": 0.2620689655172414, "acc_norm_stderr": 0.036646663372252565 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.21568627450980393, "acc_stderr": 0.04092563958237654, "acc_norm": 0.21568627450980393, "acc_norm_stderr": 0.04092563958237654 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.3403361344537815, "acc_stderr": 0.03077805742293167, "acc_norm": 0.3403361344537815, "acc_norm_stderr": 0.03077805742293167 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.2846153846153846, "acc_stderr": 0.022878322799706304, "acc_norm": 0.2846153846153846, "acc_norm_stderr": 0.022878322799706304 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.21, "acc_stderr": 0.040936018074033256, "acc_norm": 0.21, "acc_norm_stderr": 0.040936018074033256 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.22, "acc_stderr": 0.04163331998932269, "acc_norm": 0.22, "acc_norm_stderr": 0.04163331998932269 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.21296296296296297, "acc_stderr": 0.03957835471980981, "acc_norm": 0.21296296296296297, "acc_norm_stderr": 0.03957835471980981 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.2512315270935961, "acc_stderr": 0.030516530732694436, "acc_norm": 0.2512315270935961, "acc_norm_stderr": 0.030516530732694436 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.2645161290322581, "acc_stderr": 0.02509189237885928, "acc_norm": 0.2645161290322581, "acc_norm_stderr": 0.02509189237885928 }, "harness|ko_mmlu_marketing|5": { "acc": 0.2094017094017094, "acc_stderr": 0.026655699653922737, "acc_norm": 0.2094017094017094, "acc_norm_stderr": 0.026655699653922737 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.21509433962264152, "acc_stderr": 0.025288394502891356, "acc_norm": 0.21509433962264152, "acc_norm_stderr": 0.025288394502891356 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.15454545454545454, "acc_stderr": 0.03462262571262667, "acc_norm": 0.15454545454545454, "acc_norm_stderr": 0.03462262571262667 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.25925925925925924, "acc_stderr": 0.02671924078371217, "acc_norm": 0.25925925925925924, "acc_norm_stderr": 0.02671924078371217 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2781456953642384, "acc_stderr": 0.03658603262763744, "acc_norm": 0.2781456953642384, "acc_norm_stderr": 0.03658603262763744 }, "harness|ko_mmlu_sociology|5": { "acc": 0.25870646766169153, "acc_stderr": 0.030965903123573026, "acc_norm": 0.25870646766169153, "acc_norm_stderr": 0.030965903123573026 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.2543352601156069, "acc_stderr": 0.0332055644308557, "acc_norm": 0.2543352601156069, "acc_norm_stderr": 0.0332055644308557 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2275132275132275, "acc_stderr": 0.021591269407823778, "acc_norm": 0.2275132275132275, "acc_norm_stderr": 0.021591269407823778 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.20833333333333334, "acc_stderr": 0.033961162058453336, "acc_norm": 0.20833333333333334, "acc_norm_stderr": 0.033961162058453336 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.19, "acc_stderr": 0.03942772444036623, "acc_norm": 0.19, "acc_norm_stderr": 0.03942772444036623 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.23, "acc_stderr": 0.04229525846816505, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816505 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.21965317919075145, "acc_stderr": 0.022289638852617904, "acc_norm": 0.21965317919075145, "acc_norm_stderr": 0.022289638852617904 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.3128834355828221, "acc_stderr": 0.036429145782924034, "acc_norm": 0.3128834355828221, "acc_norm_stderr": 0.036429145782924034 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.23148148148148148, "acc_stderr": 0.02346842983245116, "acc_norm": 0.23148148148148148, "acc_norm_stderr": 0.02346842983245116 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.26, "acc_stderr": 0.04408440022768078, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768078 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.31088082901554404, "acc_stderr": 0.03340361906276587, "acc_norm": 0.31088082901554404, "acc_norm_stderr": 0.03340361906276587 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.21052631578947367, "acc_stderr": 0.03835153954399421, "acc_norm": 0.21052631578947367, "acc_norm_stderr": 0.03835153954399421 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.24770642201834864, "acc_stderr": 0.018508143602547805, "acc_norm": 0.24770642201834864, "acc_norm_stderr": 0.018508143602547805 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.15873015873015872, "acc_stderr": 0.03268454013011744, "acc_norm": 0.15873015873015872, "acc_norm_stderr": 0.03268454013011744 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.2549019607843137, "acc_stderr": 0.02495418432487991, "acc_norm": 0.2549019607843137, "acc_norm_stderr": 0.02495418432487991 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_international_law|5": { "acc": 0.256198347107438, "acc_stderr": 0.03984979653302871, "acc_norm": 0.256198347107438, "acc_norm_stderr": 0.03984979653302871 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.21052631578947367, "acc_stderr": 0.03317672787533156, "acc_norm": 0.21052631578947367, "acc_norm_stderr": 0.03317672787533156 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.2173202614379085, "acc_stderr": 0.016684820929148598, "acc_norm": 0.2173202614379085, "acc_norm_stderr": 0.016684820929148598 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.2375886524822695, "acc_stderr": 0.025389512552729906, "acc_norm": 0.2375886524822695, "acc_norm_stderr": 0.025389512552729906 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.2857142857142857, "acc_stderr": 0.042878587513404565, "acc_norm": 0.2857142857142857, "acc_norm_stderr": 0.042878587513404565 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4027777777777778, "acc_stderr": 0.03344887382997866, "acc_norm": 0.4027777777777778, "acc_norm_stderr": 0.03344887382997866 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24692737430167597, "acc_stderr": 0.014422292204808852, "acc_norm": 0.24692737430167597, "acc_norm_stderr": 0.014422292204808852 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.39, "acc_stderr": 0.04902071300001974, "acc_norm": 0.39, "acc_norm_stderr": 0.04902071300001974 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.28, "acc_stderr": 0.04512608598542128, "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542128 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.34191176470588236, "acc_stderr": 0.02881472242225417, "acc_norm": 0.34191176470588236, "acc_norm_stderr": 0.02881472242225417 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.2612244897959184, "acc_stderr": 0.028123429335142783, "acc_norm": 0.2612244897959184, "acc_norm_stderr": 0.028123429335142783 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.2616033755274262, "acc_stderr": 0.028609516716994934, "acc_norm": 0.2616033755274262, "acc_norm_stderr": 0.028609516716994934 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.22685788787483702, "acc_stderr": 0.010696348133569924, "acc_norm": 0.22685788787483702, "acc_norm_stderr": 0.010696348133569924 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.2549019607843137, "acc_stderr": 0.030587591351604243, "acc_norm": 0.2549019607843137, "acc_norm_stderr": 0.030587591351604243 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.22424242424242424, "acc_stderr": 0.03256866661681102, "acc_norm": 0.22424242424242424, "acc_norm_stderr": 0.03256866661681102 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.22399020807833536, "mc1_stderr": 0.014594964329474205, "mc2": 0.38068228949426847, "mc2_stderr": 0.014620809751439413 }, "harness|ko_commongen_v2|2": { "acc": 0.4037780401416765, "acc_stderr": 0.016869031540298632, "acc_norm": 0.5407319952774499, "acc_norm_stderr": 0.017133218276537673 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "beomi/KoRWKV-6B", "model_sha": "541600070459baf0f1be9560181d5ceb77794085", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }