{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.3319112627986348, "acc_stderr": 0.013760988200880533, "acc_norm": 0.38993174061433444, "acc_norm_stderr": 0.014252959848892884 }, "harness|ko_hellaswag|10": { "acc": 0.3857797251543517, "acc_stderr": 0.004857840934549158, "acc_norm": 0.5046803425612428, "acc_norm_stderr": 0.004989562798280523 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.38011695906432746, "acc_stderr": 0.037229657413855394, "acc_norm": 0.38011695906432746, "acc_norm_stderr": 0.037229657413855394 }, "harness|ko_mmlu_management|5": { "acc": 0.33980582524271846, "acc_stderr": 0.04689765937278134, "acc_norm": 0.33980582524271846, "acc_norm_stderr": 0.04689765937278134 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.41762452107279696, "acc_stderr": 0.01763563732695152, "acc_norm": 0.41762452107279696, "acc_norm_stderr": 0.01763563732695152 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.37037037037037035, "acc_stderr": 0.04171654161354544, "acc_norm": 0.37037037037037035, "acc_norm_stderr": 0.04171654161354544 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.33, "acc_stderr": 0.047258156262526045, "acc_norm": 0.33, "acc_norm_stderr": 0.047258156262526045 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.2936170212765957, "acc_stderr": 0.02977164271249123, "acc_norm": 0.2936170212765957, "acc_norm_stderr": 0.02977164271249123 }, "harness|ko_mmlu_virology|5": { "acc": 0.35542168674698793, "acc_stderr": 0.03726214354322415, "acc_norm": 0.35542168674698793, "acc_norm_stderr": 0.03726214354322415 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.43086816720257237, "acc_stderr": 0.028125340983972714, "acc_norm": 0.43086816720257237, "acc_norm_stderr": 0.028125340983972714 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.42152466367713004, "acc_stderr": 0.03314190222110658, "acc_norm": 0.42152466367713004, "acc_norm_stderr": 0.03314190222110658 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.45038167938931295, "acc_stderr": 0.04363643698524779, "acc_norm": 0.45038167938931295, "acc_norm_stderr": 0.04363643698524779 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.41414141414141414, "acc_stderr": 0.03509438348879629, "acc_norm": 0.41414141414141414, "acc_norm_stderr": 0.03509438348879629 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.3724137931034483, "acc_stderr": 0.0402873153294756, "acc_norm": 0.3724137931034483, "acc_norm_stderr": 0.0402873153294756 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.21568627450980393, "acc_stderr": 0.04092563958237655, "acc_norm": 0.21568627450980393, "acc_norm_stderr": 0.04092563958237655 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.3697478991596639, "acc_stderr": 0.031357095996135904, "acc_norm": 0.3697478991596639, "acc_norm_stderr": 0.031357095996135904 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.26153846153846155, "acc_stderr": 0.022282141204204433, "acc_norm": 0.26153846153846155, "acc_norm_stderr": 0.022282141204204433 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.41, "acc_stderr": 0.04943110704237102, "acc_norm": 0.41, "acc_norm_stderr": 0.04943110704237102 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.37962962962962965, "acc_stderr": 0.04691521224077742, "acc_norm": 0.37962962962962965, "acc_norm_stderr": 0.04691521224077742 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.2315270935960591, "acc_stderr": 0.02967833314144444, "acc_norm": 0.2315270935960591, "acc_norm_stderr": 0.02967833314144444 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.36774193548387096, "acc_stderr": 0.02743086657997347, "acc_norm": 0.36774193548387096, "acc_norm_stderr": 0.02743086657997347 }, "harness|ko_mmlu_marketing|5": { "acc": 0.5042735042735043, "acc_stderr": 0.03275489264382132, "acc_norm": 0.5042735042735043, "acc_norm_stderr": 0.03275489264382132 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.33962264150943394, "acc_stderr": 0.029146904747798352, "acc_norm": 0.33962264150943394, "acc_norm_stderr": 0.029146904747798352 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.37272727272727274, "acc_stderr": 0.046313813194254635, "acc_norm": 0.37272727272727274, "acc_norm_stderr": 0.046313813194254635 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.24814814814814815, "acc_stderr": 0.0263357394040558, "acc_norm": 0.24814814814814815, "acc_norm_stderr": 0.0263357394040558 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.23178807947019867, "acc_stderr": 0.03445406271987054, "acc_norm": 0.23178807947019867, "acc_norm_stderr": 0.03445406271987054 }, "harness|ko_mmlu_sociology|5": { "acc": 0.43283582089552236, "acc_stderr": 0.0350349092367328, "acc_norm": 0.43283582089552236, "acc_norm_stderr": 0.0350349092367328 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.30057803468208094, "acc_stderr": 0.03496101481191181, "acc_norm": 0.30057803468208094, "acc_norm_stderr": 0.03496101481191181 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2328042328042328, "acc_stderr": 0.021765961672154537, "acc_norm": 0.2328042328042328, "acc_norm_stderr": 0.021765961672154537 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3333333333333333, "acc_stderr": 0.039420826399272135, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.039420826399272135 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.24, "acc_stderr": 0.042923469599092816, "acc_norm": 0.24, "acc_norm_stderr": 0.042923469599092816 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.4, "acc_stderr": 0.049236596391733084, "acc_norm": 0.4, "acc_norm_stderr": 0.049236596391733084 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.45664739884393063, "acc_stderr": 0.02681771813034892, "acc_norm": 0.45664739884393063, "acc_norm_stderr": 0.02681771813034892 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.3312883435582822, "acc_stderr": 0.03697983910025588, "acc_norm": 0.3312883435582822, "acc_norm_stderr": 0.03697983910025588 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.39197530864197533, "acc_stderr": 0.02716368603827123, "acc_norm": 0.39197530864197533, "acc_norm_stderr": 0.02716368603827123 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.33, "acc_stderr": 0.04725815626252605, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252605 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.44041450777202074, "acc_stderr": 0.035827245300360945, "acc_norm": 0.44041450777202074, "acc_norm_stderr": 0.035827245300360945 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2543859649122807, "acc_stderr": 0.04096985139843671, "acc_norm": 0.2543859649122807, "acc_norm_stderr": 0.04096985139843671 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.3798165137614679, "acc_stderr": 0.020808825617866244, "acc_norm": 0.3798165137614679, "acc_norm_stderr": 0.020808825617866244 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.24603174603174602, "acc_stderr": 0.03852273364924318, "acc_norm": 0.24603174603174602, "acc_norm_stderr": 0.03852273364924318 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.4117647058823529, "acc_stderr": 0.02818059632825929, "acc_norm": 0.4117647058823529, "acc_norm_stderr": 0.02818059632825929 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.34, "acc_stderr": 0.04760952285695235, "acc_norm": 0.34, "acc_norm_stderr": 0.04760952285695235 }, "harness|ko_mmlu_international_law|5": { "acc": 0.48760330578512395, "acc_stderr": 0.04562951548180765, "acc_norm": 0.48760330578512395, "acc_norm_stderr": 0.04562951548180765 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.3026315789473684, "acc_stderr": 0.03738520676119667, "acc_norm": 0.3026315789473684, "acc_norm_stderr": 0.03738520676119667 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.32189542483660133, "acc_stderr": 0.018901015322093095, "acc_norm": 0.32189542483660133, "acc_norm_stderr": 0.018901015322093095 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3191489361702128, "acc_stderr": 0.027807990141320203, "acc_norm": 0.3191489361702128, "acc_norm_stderr": 0.027807990141320203 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.2767857142857143, "acc_stderr": 0.042466243366976256, "acc_norm": 0.2767857142857143, "acc_norm_stderr": 0.042466243366976256 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.35185185185185186, "acc_stderr": 0.032568505702936464, "acc_norm": 0.35185185185185186, "acc_norm_stderr": 0.032568505702936464 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24134078212290502, "acc_stderr": 0.014310999547961443, "acc_norm": 0.24134078212290502, "acc_norm_stderr": 0.014310999547961443 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.28, "acc_stderr": 0.045126085985421276, "acc_norm": 0.28, "acc_norm_stderr": 0.045126085985421276 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.28, "acc_stderr": 0.04512608598542127, "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542127 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4264705882352941, "acc_stderr": 0.03004261583271486, "acc_norm": 0.4264705882352941, "acc_norm_stderr": 0.03004261583271486 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.39591836734693875, "acc_stderr": 0.03130802899065686, "acc_norm": 0.39591836734693875, "acc_norm_stderr": 0.03130802899065686 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.42616033755274263, "acc_stderr": 0.03219035703131774, "acc_norm": 0.42616033755274263, "acc_norm_stderr": 0.03219035703131774 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.33376792698826596, "acc_stderr": 0.012043812655846146, "acc_norm": 0.33376792698826596, "acc_norm_stderr": 0.012043812655846146 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.3627450980392157, "acc_stderr": 0.033744993563193555, "acc_norm": 0.3627450980392157, "acc_norm_stderr": 0.033744993563193555 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.3575757575757576, "acc_stderr": 0.03742597043806587, "acc_norm": 0.3575757575757576, "acc_norm_stderr": 0.03742597043806587 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2484700122399021, "mc1_stderr": 0.015127427096520688, "mc2": 0.3821911392219441, "mc2_stderr": 0.014928316371274168 }, "harness|ko_commongen_v2|2": { "acc": 0.5692488262910798, "acc_stderr": 0.01697459912173145, "acc_norm": 0.6326291079812206, "acc_norm_stderr": 0.016525787977931604 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "DopeorNope/COLA_LO-7B", "model_sha": "4cccb5249ea36f58588c32fe58c6f104f89f0487", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }