{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.45051194539249145, "acc_stderr": 0.014539646098471627, "acc_norm": 0.5008532423208191, "acc_norm_stderr": 0.014611369529813269 }, "harness|ko_hellaswag|10": { "acc": 0.4230233021310496, "acc_stderr": 0.004930293787545619, "acc_norm": 0.5676160127464649, "acc_norm_stderr": 0.0049439450696114546 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.6374269005847953, "acc_stderr": 0.0368713061556206, "acc_norm": 0.6374269005847953, "acc_norm_stderr": 0.0368713061556206 }, "harness|ko_mmlu_management|5": { "acc": 0.6019417475728155, "acc_stderr": 0.048467482539772386, "acc_norm": 0.6019417475728155, "acc_norm_stderr": 0.048467482539772386 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.6168582375478927, "acc_stderr": 0.01738477419488562, "acc_norm": 0.6168582375478927, "acc_norm_stderr": 0.01738477419488562 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.48148148148148145, "acc_stderr": 0.04316378599511326, "acc_norm": 0.48148148148148145, "acc_norm_stderr": 0.04316378599511326 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.37, "acc_stderr": 0.048523658709391, "acc_norm": 0.37, "acc_norm_stderr": 0.048523658709391 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.4595744680851064, "acc_stderr": 0.032579014820998356, "acc_norm": 0.4595744680851064, "acc_norm_stderr": 0.032579014820998356 }, "harness|ko_mmlu_virology|5": { "acc": 0.46987951807228917, "acc_stderr": 0.03885425420866766, "acc_norm": 0.46987951807228917, "acc_norm_stderr": 0.03885425420866766 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5594855305466238, "acc_stderr": 0.028196400574197422, "acc_norm": 0.5594855305466238, "acc_norm_stderr": 0.028196400574197422 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.5739910313901345, "acc_stderr": 0.03318833286217281, "acc_norm": 0.5739910313901345, "acc_norm_stderr": 0.03318833286217281 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.5725190839694656, "acc_stderr": 0.04338920305792401, "acc_norm": 0.5725190839694656, "acc_norm_stderr": 0.04338920305792401 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.53, "acc_stderr": 0.050161355804659205, "acc_norm": 0.53, "acc_norm_stderr": 0.050161355804659205 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.6666666666666666, "acc_stderr": 0.03358618145732524, "acc_norm": 0.6666666666666666, "acc_norm_stderr": 0.03358618145732524 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.45517241379310347, "acc_stderr": 0.04149886942192117, "acc_norm": 0.45517241379310347, "acc_norm_stderr": 0.04149886942192117 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.2647058823529412, "acc_stderr": 0.04389869956808778, "acc_norm": 0.2647058823529412, "acc_norm_stderr": 0.04389869956808778 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.5798319327731093, "acc_stderr": 0.032061837832361516, "acc_norm": 0.5798319327731093, "acc_norm_stderr": 0.032061837832361516 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.5333333333333333, "acc_stderr": 0.025294608023986462, "acc_norm": 0.5333333333333333, "acc_norm_stderr": 0.025294608023986462 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.62, "acc_stderr": 0.048783173121456316, "acc_norm": 0.62, "acc_norm_stderr": 0.048783173121456316 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.6018518518518519, "acc_stderr": 0.04732332615978814, "acc_norm": 0.6018518518518519, "acc_norm_stderr": 0.04732332615978814 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.4482758620689655, "acc_stderr": 0.034991131376767445, "acc_norm": 0.4482758620689655, "acc_norm_stderr": 0.034991131376767445 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.6, "acc_stderr": 0.027869320571664632, "acc_norm": 0.6, "acc_norm_stderr": 0.027869320571664632 }, "harness|ko_mmlu_marketing|5": { "acc": 0.7264957264957265, "acc_stderr": 0.029202540153431187, "acc_norm": 0.7264957264957265, "acc_norm_stderr": 0.029202540153431187 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.5471698113207547, "acc_stderr": 0.03063562795796182, "acc_norm": 0.5471698113207547, "acc_norm_stderr": 0.03063562795796182 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5909090909090909, "acc_stderr": 0.04709306978661895, "acc_norm": 0.5909090909090909, "acc_norm_stderr": 0.04709306978661895 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.3296296296296296, "acc_stderr": 0.028661201116524572, "acc_norm": 0.3296296296296296, "acc_norm_stderr": 0.028661201116524572 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.31788079470198677, "acc_stderr": 0.038020397601079024, "acc_norm": 0.31788079470198677, "acc_norm_stderr": 0.038020397601079024 }, "harness|ko_mmlu_sociology|5": { "acc": 0.6965174129353234, "acc_stderr": 0.03251006816458618, "acc_norm": 0.6965174129353234, "acc_norm_stderr": 0.03251006816458618 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.5433526011560693, "acc_stderr": 0.03798106566014498, "acc_norm": 0.5433526011560693, "acc_norm_stderr": 0.03798106566014498 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.37566137566137564, "acc_stderr": 0.02494236893115979, "acc_norm": 0.37566137566137564, "acc_norm_stderr": 0.02494236893115979 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.4583333333333333, "acc_stderr": 0.04166666666666665, "acc_norm": 0.4583333333333333, "acc_norm_stderr": 0.04166666666666665 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.42, "acc_stderr": 0.049604496374885836, "acc_norm": 0.42, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.72, "acc_stderr": 0.04512608598542128, "acc_norm": 0.72, "acc_norm_stderr": 0.04512608598542128 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5491329479768786, "acc_stderr": 0.026788811931562757, "acc_norm": 0.5491329479768786, "acc_norm_stderr": 0.026788811931562757 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.5214723926380368, "acc_stderr": 0.03924746876751129, "acc_norm": 0.5214723926380368, "acc_norm_stderr": 0.03924746876751129 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.5401234567901234, "acc_stderr": 0.027731022753539274, "acc_norm": 0.5401234567901234, "acc_norm_stderr": 0.027731022753539274 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.35, "acc_stderr": 0.0479372485441102, "acc_norm": 0.35, "acc_norm_stderr": 0.0479372485441102 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.689119170984456, "acc_stderr": 0.03340361906276586, "acc_norm": 0.689119170984456, "acc_norm_stderr": 0.03340361906276586 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2894736842105263, "acc_stderr": 0.04266339443159394, "acc_norm": 0.2894736842105263, "acc_norm_stderr": 0.04266339443159394 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.6770642201834862, "acc_stderr": 0.020048115923415342, "acc_norm": 0.6770642201834862, "acc_norm_stderr": 0.020048115923415342 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.3253968253968254, "acc_stderr": 0.04190596438871136, "acc_norm": 0.3253968253968254, "acc_norm_stderr": 0.04190596438871136 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5718954248366013, "acc_stderr": 0.028332397483664274, "acc_norm": 0.5718954248366013, "acc_norm_stderr": 0.028332397483664274 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.49, "acc_stderr": 0.05024183937956912, "acc_norm": 0.49, "acc_norm_stderr": 0.05024183937956912 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6776859504132231, "acc_stderr": 0.042664163633521685, "acc_norm": 0.6776859504132231, "acc_norm_stderr": 0.042664163633521685 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.5328947368421053, "acc_stderr": 0.04060127035236395, "acc_norm": 0.5328947368421053, "acc_norm_stderr": 0.04060127035236395 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.49673202614379086, "acc_stderr": 0.020227402794434867, "acc_norm": 0.49673202614379086, "acc_norm_stderr": 0.020227402794434867 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.35106382978723405, "acc_stderr": 0.028473501272963775, "acc_norm": 0.35106382978723405, "acc_norm_stderr": 0.028473501272963775 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.38392857142857145, "acc_stderr": 0.046161430750285455, "acc_norm": 0.38392857142857145, "acc_norm_stderr": 0.046161430750285455 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.44907407407407407, "acc_stderr": 0.03392238405321617, "acc_norm": 0.44907407407407407, "acc_norm_stderr": 0.03392238405321617 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.30726256983240224, "acc_stderr": 0.01543015884646961, "acc_norm": 0.30726256983240224, "acc_norm_stderr": 0.01543015884646961 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.46, "acc_stderr": 0.05009082659620333, "acc_norm": 0.46, "acc_norm_stderr": 0.05009082659620333 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.62, "acc_stderr": 0.04878317312145632, "acc_norm": 0.62, "acc_norm_stderr": 0.04878317312145632 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4485294117647059, "acc_stderr": 0.0302114796091216, "acc_norm": 0.4485294117647059, "acc_norm_stderr": 0.0302114796091216 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.6489795918367347, "acc_stderr": 0.030555316755573644, "acc_norm": 0.6489795918367347, "acc_norm_stderr": 0.030555316755573644 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.6919831223628692, "acc_stderr": 0.030052389335605695, "acc_norm": 0.6919831223628692, "acc_norm_stderr": 0.030052389335605695 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3559322033898305, "acc_stderr": 0.01222864553727757, "acc_norm": 0.3559322033898305, "acc_norm_stderr": 0.01222864553727757 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.6029411764705882, "acc_stderr": 0.034341311647191286, "acc_norm": 0.6029411764705882, "acc_norm_stderr": 0.034341311647191286 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.6242424242424243, "acc_stderr": 0.03781887353205983, "acc_norm": 0.6242424242424243, "acc_norm_stderr": 0.03781887353205983 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2827417380660955, "mc1_stderr": 0.01576477083677731, "mc2": 0.43390504265082586, "mc2_stderr": 0.015336718297088065 }, "harness|ko_commongen_v2|2": { "acc": 0.5017709563164109, "acc_stderr": 0.017190246276231853, "acc_norm": 0.5584415584415584, "acc_norm_stderr": 0.01707252587556311 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Changgil/K2S3-v0.1", "model_sha": "d544e389f091983bb4f11314edb526d81753c919", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }