{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.2790102389078498, "acc_stderr": 0.013106784883601348, "acc_norm": 0.318259385665529, "acc_norm_stderr": 0.013611993916971453 }, "harness|ko_hellaswag|10": { "acc": 0.3414658434574786, "acc_stderr": 0.0047323221721537485, "acc_norm": 0.41884086835291773, "acc_norm_stderr": 0.004923609207861538 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.45614035087719296, "acc_stderr": 0.03820042586602967, "acc_norm": 0.45614035087719296, "acc_norm_stderr": 0.03820042586602967 }, "harness|ko_mmlu_management|5": { "acc": 0.3786407766990291, "acc_stderr": 0.048026946982589726, "acc_norm": 0.3786407766990291, "acc_norm_stderr": 0.048026946982589726 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.3933588761174968, "acc_stderr": 0.017468556724503162, "acc_norm": 0.3933588761174968, "acc_norm_stderr": 0.017468556724503162 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.3333333333333333, "acc_stderr": 0.04072314811876837, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.04072314811876837 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.34893617021276596, "acc_stderr": 0.031158522131357797, "acc_norm": 0.34893617021276596, "acc_norm_stderr": 0.031158522131357797 }, "harness|ko_mmlu_virology|5": { "acc": 0.30120481927710846, "acc_stderr": 0.03571609230053481, "acc_norm": 0.30120481927710846, "acc_norm_stderr": 0.03571609230053481 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.3408360128617363, "acc_stderr": 0.026920841260776162, "acc_norm": 0.3408360128617363, "acc_norm_stderr": 0.026920841260776162 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.37668161434977576, "acc_stderr": 0.032521134899291884, "acc_norm": 0.37668161434977576, "acc_norm_stderr": 0.032521134899291884 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.44274809160305345, "acc_stderr": 0.0435644720266507, "acc_norm": 0.44274809160305345, "acc_norm_stderr": 0.0435644720266507 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.37373737373737376, "acc_stderr": 0.03446897738659333, "acc_norm": 0.37373737373737376, "acc_norm_stderr": 0.03446897738659333 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.33793103448275863, "acc_stderr": 0.03941707632064889, "acc_norm": 0.33793103448275863, "acc_norm_stderr": 0.03941707632064889 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.17647058823529413, "acc_stderr": 0.0379328118530781, "acc_norm": 0.17647058823529413, "acc_norm_stderr": 0.0379328118530781 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.3445378151260504, "acc_stderr": 0.030868682604121633, "acc_norm": 0.3445378151260504, "acc_norm_stderr": 0.030868682604121633 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.3871794871794872, "acc_stderr": 0.024697216930878944, "acc_norm": 0.3871794871794872, "acc_norm_stderr": 0.024697216930878944 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.33, "acc_stderr": 0.047258156262526045, "acc_norm": 0.33, "acc_norm_stderr": 0.047258156262526045 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.4074074074074074, "acc_stderr": 0.04750077341199985, "acc_norm": 0.4074074074074074, "acc_norm_stderr": 0.04750077341199985 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.28078817733990147, "acc_stderr": 0.03161856335358611, "acc_norm": 0.28078817733990147, "acc_norm_stderr": 0.03161856335358611 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.3580645161290323, "acc_stderr": 0.02727389059430064, "acc_norm": 0.3580645161290323, "acc_norm_stderr": 0.02727389059430064 }, "harness|ko_mmlu_marketing|5": { "acc": 0.5555555555555556, "acc_stderr": 0.03255326307272486, "acc_norm": 0.5555555555555556, "acc_norm_stderr": 0.03255326307272486 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.33584905660377357, "acc_stderr": 0.029067220146644826, "acc_norm": 0.33584905660377357, "acc_norm_stderr": 0.029067220146644826 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.43636363636363634, "acc_stderr": 0.04750185058907297, "acc_norm": 0.43636363636363634, "acc_norm_stderr": 0.04750185058907297 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.2777777777777778, "acc_stderr": 0.027309140588230193, "acc_norm": 0.2777777777777778, "acc_norm_stderr": 0.027309140588230193 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2052980132450331, "acc_stderr": 0.03297986648473834, "acc_norm": 0.2052980132450331, "acc_norm_stderr": 0.03297986648473834 }, "harness|ko_mmlu_sociology|5": { "acc": 0.48756218905472637, "acc_stderr": 0.035344398485395806, "acc_norm": 0.48756218905472637, "acc_norm_stderr": 0.035344398485395806 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.34104046242774566, "acc_stderr": 0.036146654241808254, "acc_norm": 0.34104046242774566, "acc_norm_stderr": 0.036146654241808254 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2619047619047619, "acc_stderr": 0.022644212615525218, "acc_norm": 0.2619047619047619, "acc_norm_stderr": 0.022644212615525218 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3472222222222222, "acc_stderr": 0.039812405437178615, "acc_norm": 0.3472222222222222, "acc_norm_stderr": 0.039812405437178615 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.27, "acc_stderr": 0.044619604333847394, "acc_norm": 0.27, "acc_norm_stderr": 0.044619604333847394 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.51, "acc_stderr": 0.05024183937956911, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956911 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.36127167630057805, "acc_stderr": 0.02586220185227789, "acc_norm": 0.36127167630057805, "acc_norm_stderr": 0.02586220185227789 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.3619631901840491, "acc_stderr": 0.037757007291414416, "acc_norm": 0.3619631901840491, "acc_norm_stderr": 0.037757007291414416 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.39197530864197533, "acc_stderr": 0.027163686038271215, "acc_norm": 0.39197530864197533, "acc_norm_stderr": 0.027163686038271215 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.26, "acc_stderr": 0.04408440022768079, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768079 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.46113989637305697, "acc_stderr": 0.03597524411734578, "acc_norm": 0.46113989637305697, "acc_norm_stderr": 0.03597524411734578 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2631578947368421, "acc_stderr": 0.0414243971948936, "acc_norm": 0.2631578947368421, "acc_norm_stderr": 0.0414243971948936 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.3688073394495413, "acc_stderr": 0.020686227560729548, "acc_norm": 0.3688073394495413, "acc_norm_stderr": 0.020686227560729548 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.29365079365079366, "acc_stderr": 0.040735243221471255, "acc_norm": 0.29365079365079366, "acc_norm_stderr": 0.040735243221471255 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.4084967320261438, "acc_stderr": 0.028146405993096358, "acc_norm": 0.4084967320261438, "acc_norm_stderr": 0.028146405993096358 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_international_law|5": { "acc": 0.5371900826446281, "acc_stderr": 0.04551711196104218, "acc_norm": 0.5371900826446281, "acc_norm_stderr": 0.04551711196104218 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.34868421052631576, "acc_stderr": 0.03878139888797611, "acc_norm": 0.34868421052631576, "acc_norm_stderr": 0.03878139888797611 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.32516339869281047, "acc_stderr": 0.01895088677080631, "acc_norm": 0.32516339869281047, "acc_norm_stderr": 0.01895088677080631 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.2978723404255319, "acc_stderr": 0.027281608344469414, "acc_norm": 0.2978723404255319, "acc_norm_stderr": 0.027281608344469414 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.3125, "acc_stderr": 0.043994650575715215, "acc_norm": 0.3125, "acc_norm_stderr": 0.043994650575715215 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.2916666666666667, "acc_stderr": 0.030998666304560538, "acc_norm": 0.2916666666666667, "acc_norm_stderr": 0.030998666304560538 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2446927374301676, "acc_stderr": 0.014378169884098424, "acc_norm": 0.2446927374301676, "acc_norm_stderr": 0.014378169884098424 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.27, "acc_stderr": 0.044619604333847394, "acc_norm": 0.27, "acc_norm_stderr": 0.044619604333847394 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.3639705882352941, "acc_stderr": 0.02922719246003203, "acc_norm": 0.3639705882352941, "acc_norm_stderr": 0.02922719246003203 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.3673469387755102, "acc_stderr": 0.030862144921087558, "acc_norm": 0.3673469387755102, "acc_norm_stderr": 0.030862144921087558 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.43037974683544306, "acc_stderr": 0.03223017195937599, "acc_norm": 0.43037974683544306, "acc_norm_stderr": 0.03223017195937599 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.26140808344198174, "acc_stderr": 0.011222528169771316, "acc_norm": 0.26140808344198174, "acc_norm_stderr": 0.011222528169771316 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.37254901960784315, "acc_stderr": 0.03393388584958406, "acc_norm": 0.37254901960784315, "acc_norm_stderr": 0.03393388584958406 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.41818181818181815, "acc_stderr": 0.038517163193983954, "acc_norm": 0.41818181818181815, "acc_norm_stderr": 0.038517163193983954 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2962056303549572, "mc1_stderr": 0.015983595101811392, "mc2": 0.4712238434154724, "mc2_stderr": 0.016160223034293618 }, "harness|ko_commongen_v2|2": { "acc": 0.3215962441314554, "acc_stderr": 0.01601160345597796, "acc_norm": 0.3544600938967136, "acc_norm_stderr": 0.016397605788502075 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "NousResearch/Nous-Capybara-7B", "model_sha": "49d96bb4659cad308f204ca946962cdcd8958ca6", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }