{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.39419795221843, "acc_stderr": 0.01428052266746733, "acc_norm": 0.4513651877133106, "acc_norm_stderr": 0.014542104569955262 }, "harness|ko_hellaswag|10": { "acc": 0.4221270663214499, "acc_stderr": 0.004928891895874289, "acc_norm": 0.5546703843855806, "acc_norm_stderr": 0.00495986429917813 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.5380116959064327, "acc_stderr": 0.03823727092882307, "acc_norm": 0.5380116959064327, "acc_norm_stderr": 0.03823727092882307 }, "harness|ko_mmlu_management|5": { "acc": 0.5436893203883495, "acc_stderr": 0.049318019942204146, "acc_norm": 0.5436893203883495, "acc_norm_stderr": 0.049318019942204146 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.5044699872286079, "acc_stderr": 0.017879248970584356, "acc_norm": 0.5044699872286079, "acc_norm_stderr": 0.017879248970584356 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4074074074074074, "acc_stderr": 0.042446332383532286, "acc_norm": 0.4074074074074074, "acc_norm_stderr": 0.042446332383532286 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.32, "acc_stderr": 0.04688261722621503, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621503 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.3659574468085106, "acc_stderr": 0.0314895582974553, "acc_norm": 0.3659574468085106, "acc_norm_stderr": 0.0314895582974553 }, "harness|ko_mmlu_virology|5": { "acc": 0.42168674698795183, "acc_stderr": 0.03844453181770917, "acc_norm": 0.42168674698795183, "acc_norm_stderr": 0.03844453181770917 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.4533762057877814, "acc_stderr": 0.028274359854894255, "acc_norm": 0.4533762057877814, "acc_norm_stderr": 0.028274359854894255 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.3632286995515695, "acc_stderr": 0.03227790442850499, "acc_norm": 0.3632286995515695, "acc_norm_stderr": 0.03227790442850499 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.5038167938931297, "acc_stderr": 0.043851623256015534, "acc_norm": 0.5038167938931297, "acc_norm_stderr": 0.043851623256015534 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5858585858585859, "acc_stderr": 0.03509438348879629, "acc_norm": 0.5858585858585859, "acc_norm_stderr": 0.03509438348879629 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.41379310344827586, "acc_stderr": 0.041042692118062316, "acc_norm": 0.41379310344827586, "acc_norm_stderr": 0.041042692118062316 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.24509803921568626, "acc_stderr": 0.042801058373643966, "acc_norm": 0.24509803921568626, "acc_norm_stderr": 0.042801058373643966 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.42436974789915966, "acc_stderr": 0.03210479051015776, "acc_norm": 0.42436974789915966, "acc_norm_stderr": 0.03210479051015776 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.39487179487179486, "acc_stderr": 0.02478431694215636, "acc_norm": 0.39487179487179486, "acc_norm_stderr": 0.02478431694215636 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.48, "acc_stderr": 0.050211673156867795, "acc_norm": 0.48, "acc_norm_stderr": 0.050211673156867795 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.36, "acc_stderr": 0.04824181513244218, "acc_norm": 0.36, "acc_norm_stderr": 0.04824181513244218 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.4444444444444444, "acc_stderr": 0.04803752235190193, "acc_norm": 0.4444444444444444, "acc_norm_stderr": 0.04803752235190193 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3448275862068966, "acc_stderr": 0.03344283744280458, "acc_norm": 0.3448275862068966, "acc_norm_stderr": 0.03344283744280458 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.44193548387096776, "acc_stderr": 0.02825155790684973, "acc_norm": 0.44193548387096776, "acc_norm_stderr": 0.02825155790684973 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6581196581196581, "acc_stderr": 0.031075028526507748, "acc_norm": 0.6581196581196581, "acc_norm_stderr": 0.031075028526507748 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.47924528301886793, "acc_stderr": 0.030746349975723463, "acc_norm": 0.47924528301886793, "acc_norm_stderr": 0.030746349975723463 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.4818181818181818, "acc_stderr": 0.04785964010794916, "acc_norm": 0.4818181818181818, "acc_norm_stderr": 0.04785964010794916 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.24444444444444444, "acc_stderr": 0.026202766534652148, "acc_norm": 0.24444444444444444, "acc_norm_stderr": 0.026202766534652148 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2781456953642384, "acc_stderr": 0.03658603262763743, "acc_norm": 0.2781456953642384, "acc_norm_stderr": 0.03658603262763743 }, "harness|ko_mmlu_sociology|5": { "acc": 0.5970149253731343, "acc_stderr": 0.034683432951111266, "acc_norm": 0.5970149253731343, "acc_norm_stderr": 0.034683432951111266 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.35260115606936415, "acc_stderr": 0.036430371689585496, "acc_norm": 0.35260115606936415, "acc_norm_stderr": 0.036430371689585496 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.30423280423280424, "acc_stderr": 0.023695415009463084, "acc_norm": 0.30423280423280424, "acc_norm_stderr": 0.023695415009463084 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3194444444444444, "acc_stderr": 0.038990736873573344, "acc_norm": 0.3194444444444444, "acc_norm_stderr": 0.038990736873573344 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.64, "acc_stderr": 0.048241815132442176, "acc_norm": 0.64, "acc_norm_stderr": 0.048241815132442176 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.45375722543352603, "acc_stderr": 0.02680372058320618, "acc_norm": 0.45375722543352603, "acc_norm_stderr": 0.02680372058320618 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.4171779141104294, "acc_stderr": 0.038741028598180814, "acc_norm": 0.4171779141104294, "acc_norm_stderr": 0.038741028598180814 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.4567901234567901, "acc_stderr": 0.027716661650194045, "acc_norm": 0.4567901234567901, "acc_norm_stderr": 0.027716661650194045 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.33, "acc_stderr": 0.04725815626252606, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252606 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.47150259067357514, "acc_stderr": 0.03602573571288442, "acc_norm": 0.47150259067357514, "acc_norm_stderr": 0.03602573571288442 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2719298245614035, "acc_stderr": 0.04185774424022057, "acc_norm": 0.2719298245614035, "acc_norm_stderr": 0.04185774424022057 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.4917431192660551, "acc_stderr": 0.021434399918214338, "acc_norm": 0.4917431192660551, "acc_norm_stderr": 0.021434399918214338 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.30952380952380953, "acc_stderr": 0.04134913018303316, "acc_norm": 0.30952380952380953, "acc_norm_stderr": 0.04134913018303316 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.45098039215686275, "acc_stderr": 0.02849199358617157, "acc_norm": 0.45098039215686275, "acc_norm_stderr": 0.02849199358617157 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.39, "acc_stderr": 0.04902071300001975, "acc_norm": 0.39, "acc_norm_stderr": 0.04902071300001975 }, "harness|ko_mmlu_international_law|5": { "acc": 0.5785123966942148, "acc_stderr": 0.04507732278775087, "acc_norm": 0.5785123966942148, "acc_norm_stderr": 0.04507732278775087 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.4407894736842105, "acc_stderr": 0.04040311062490436, "acc_norm": 0.4407894736842105, "acc_norm_stderr": 0.04040311062490436 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.3300653594771242, "acc_stderr": 0.019023726160724553, "acc_norm": 0.3300653594771242, "acc_norm_stderr": 0.019023726160724553 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3262411347517731, "acc_stderr": 0.02796845304356317, "acc_norm": 0.3262411347517731, "acc_norm_stderr": 0.02796845304356317 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.23214285714285715, "acc_stderr": 0.040073418097558065, "acc_norm": 0.23214285714285715, "acc_norm_stderr": 0.040073418097558065 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.2777777777777778, "acc_stderr": 0.0305467452649532, "acc_norm": 0.2777777777777778, "acc_norm_stderr": 0.0305467452649532 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.2424581005586592, "acc_stderr": 0.01433352205921789, "acc_norm": 0.2424581005586592, "acc_norm_stderr": 0.01433352205921789 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.46, "acc_stderr": 0.05009082659620333, "acc_norm": 0.46, "acc_norm_stderr": 0.05009082659620333 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.27205882352941174, "acc_stderr": 0.027033041151681456, "acc_norm": 0.27205882352941174, "acc_norm_stderr": 0.027033041151681456 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.49387755102040815, "acc_stderr": 0.03200682020163908, "acc_norm": 0.49387755102040815, "acc_norm_stderr": 0.03200682020163908 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.5189873417721519, "acc_stderr": 0.03252375148090447, "acc_norm": 0.5189873417721519, "acc_norm_stderr": 0.03252375148090447 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.29726205997392435, "acc_stderr": 0.01167334617308604, "acc_norm": 0.29726205997392435, "acc_norm_stderr": 0.01167334617308604 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.4264705882352941, "acc_stderr": 0.034711579079534254, "acc_norm": 0.4264705882352941, "acc_norm_stderr": 0.034711579079534254 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.503030303030303, "acc_stderr": 0.03904272341431855, "acc_norm": 0.503030303030303, "acc_norm_stderr": 0.03904272341431855 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.30599755201958384, "mc1_stderr": 0.016132229728155062, "mc2": 0.47405944536561195, "mc2_stderr": 0.015299344788205708 }, "harness|ko_commongen_v2|2": { "acc": 0.5206611570247934, "acc_stderr": 0.01717567127983645, "acc_norm": 0.5726092089728453, "acc_norm_stderr": 0.017008129844823156 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "krevas/LDCC-Instruct-Llama-2-ko-13B-v4.2.4", "model_sha": "80bc09f73579e2bc021f084424f761f81e4f808d", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }