{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.3447098976109215, "acc_stderr": 0.013888816286782112, "acc_norm": 0.3984641638225256, "acc_norm_stderr": 0.014306946052735565 }, "harness|ko_hellaswag|10": { "acc": 0.398725353515236, "acc_stderr": 0.004886353563571851, "acc_norm": 0.5272854013144792, "acc_norm_stderr": 0.004982346155911131 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.4619883040935672, "acc_stderr": 0.03823727092882307, "acc_norm": 0.4619883040935672, "acc_norm_stderr": 0.03823727092882307 }, "harness|ko_mmlu_management|5": { "acc": 0.5631067961165048, "acc_stderr": 0.04911147107365777, "acc_norm": 0.5631067961165048, "acc_norm_stderr": 0.04911147107365777 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.5555555555555556, "acc_stderr": 0.017769250583533253, "acc_norm": 0.5555555555555556, "acc_norm_stderr": 0.017769250583533253 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.45925925925925926, "acc_stderr": 0.04304979692464244, "acc_norm": 0.45925925925925926, "acc_norm_stderr": 0.04304979692464244 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.32, "acc_stderr": 0.04688261722621503, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621503 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.40425531914893614, "acc_stderr": 0.03208115750788684, "acc_norm": 0.40425531914893614, "acc_norm_stderr": 0.03208115750788684 }, "harness|ko_mmlu_virology|5": { "acc": 0.37349397590361444, "acc_stderr": 0.037658451171688624, "acc_norm": 0.37349397590361444, "acc_norm_stderr": 0.037658451171688624 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5144694533762058, "acc_stderr": 0.02838619808417768, "acc_norm": 0.5144694533762058, "acc_norm_stderr": 0.02838619808417768 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.49327354260089684, "acc_stderr": 0.033554765962343545, "acc_norm": 0.49327354260089684, "acc_norm_stderr": 0.033554765962343545 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.4580152671755725, "acc_stderr": 0.04369802690578756, "acc_norm": 0.4580152671755725, "acc_norm_stderr": 0.04369802690578756 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5606060606060606, "acc_stderr": 0.03536085947529482, "acc_norm": 0.5606060606060606, "acc_norm_stderr": 0.03536085947529482 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.4689655172413793, "acc_stderr": 0.04158632762097828, "acc_norm": 0.4689655172413793, "acc_norm_stderr": 0.04158632762097828 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.24509803921568626, "acc_stderr": 0.042801058373643966, "acc_norm": 0.24509803921568626, "acc_norm_stderr": 0.042801058373643966 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.4495798319327731, "acc_stderr": 0.03231293497137707, "acc_norm": 0.4495798319327731, "acc_norm_stderr": 0.03231293497137707 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.44871794871794873, "acc_stderr": 0.025217315184846482, "acc_norm": 0.44871794871794873, "acc_norm_stderr": 0.025217315184846482 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.55, "acc_stderr": 0.049999999999999996, "acc_norm": 0.55, "acc_norm_stderr": 0.049999999999999996 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.32, "acc_stderr": 0.04688261722621504, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621504 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5462962962962963, "acc_stderr": 0.048129173245368216, "acc_norm": 0.5462962962962963, "acc_norm_stderr": 0.048129173245368216 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3694581280788177, "acc_stderr": 0.03395970381998574, "acc_norm": 0.3694581280788177, "acc_norm_stderr": 0.03395970381998574 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.432258064516129, "acc_stderr": 0.028181739720019416, "acc_norm": 0.432258064516129, "acc_norm_stderr": 0.028181739720019416 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6623931623931624, "acc_stderr": 0.030980296992618554, "acc_norm": 0.6623931623931624, "acc_norm_stderr": 0.030980296992618554 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.45660377358490567, "acc_stderr": 0.03065674869673943, "acc_norm": 0.45660377358490567, "acc_norm_stderr": 0.03065674869673943 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5, "acc_stderr": 0.04789131426105757, "acc_norm": 0.5, "acc_norm_stderr": 0.04789131426105757 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.2814814814814815, "acc_stderr": 0.027420019350945273, "acc_norm": 0.2814814814814815, "acc_norm_stderr": 0.027420019350945273 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2781456953642384, "acc_stderr": 0.03658603262763743, "acc_norm": 0.2781456953642384, "acc_norm_stderr": 0.03658603262763743 }, "harness|ko_mmlu_sociology|5": { "acc": 0.582089552238806, "acc_stderr": 0.034875586404620636, "acc_norm": 0.582089552238806, "acc_norm_stderr": 0.034875586404620636 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.3699421965317919, "acc_stderr": 0.03681229633394318, "acc_norm": 0.3699421965317919, "acc_norm_stderr": 0.03681229633394318 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.30158730158730157, "acc_stderr": 0.023636975996101813, "acc_norm": 0.30158730158730157, "acc_norm_stderr": 0.023636975996101813 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3888888888888889, "acc_stderr": 0.04076663253918567, "acc_norm": 0.3888888888888889, "acc_norm_stderr": 0.04076663253918567 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.59, "acc_stderr": 0.04943110704237101, "acc_norm": 0.59, "acc_norm_stderr": 0.04943110704237101 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.49421965317919075, "acc_stderr": 0.026917296179149123, "acc_norm": 0.49421965317919075, "acc_norm_stderr": 0.026917296179149123 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.44785276073619634, "acc_stderr": 0.03906947479456601, "acc_norm": 0.44785276073619634, "acc_norm_stderr": 0.03906947479456601 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.5092592592592593, "acc_stderr": 0.027815973433878014, "acc_norm": 0.5092592592592593, "acc_norm_stderr": 0.027815973433878014 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.26, "acc_stderr": 0.0440844002276808, "acc_norm": 0.26, "acc_norm_stderr": 0.0440844002276808 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.5284974093264249, "acc_stderr": 0.036025735712884414, "acc_norm": 0.5284974093264249, "acc_norm_stderr": 0.036025735712884414 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.34210526315789475, "acc_stderr": 0.04462917535336936, "acc_norm": 0.34210526315789475, "acc_norm_stderr": 0.04462917535336936 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.5486238532110091, "acc_stderr": 0.02133571471126879, "acc_norm": 0.5486238532110091, "acc_norm_stderr": 0.02133571471126879 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.23809523809523808, "acc_stderr": 0.038095238095238126, "acc_norm": 0.23809523809523808, "acc_norm_stderr": 0.038095238095238126 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.4215686274509804, "acc_stderr": 0.028275490156791434, "acc_norm": 0.4215686274509804, "acc_norm_stderr": 0.028275490156791434 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.49, "acc_stderr": 0.05024183937956912, "acc_norm": 0.49, "acc_norm_stderr": 0.05024183937956912 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6033057851239669, "acc_stderr": 0.044658697805310094, "acc_norm": 0.6033057851239669, "acc_norm_stderr": 0.044658697805310094 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.4407894736842105, "acc_stderr": 0.04040311062490436, "acc_norm": 0.4407894736842105, "acc_norm_stderr": 0.04040311062490436 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.4117647058823529, "acc_stderr": 0.019910377463105935, "acc_norm": 0.4117647058823529, "acc_norm_stderr": 0.019910377463105935 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.3475177304964539, "acc_stderr": 0.028406627809590947, "acc_norm": 0.3475177304964539, "acc_norm_stderr": 0.028406627809590947 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.30357142857142855, "acc_stderr": 0.04364226155841044, "acc_norm": 0.30357142857142855, "acc_norm_stderr": 0.04364226155841044 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.28703703703703703, "acc_stderr": 0.030851992993257013, "acc_norm": 0.28703703703703703, "acc_norm_stderr": 0.030851992993257013 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24581005586592178, "acc_stderr": 0.01440029642922561, "acc_norm": 0.24581005586592178, "acc_norm_stderr": 0.01440029642922561 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.45, "acc_stderr": 0.049999999999999996, "acc_norm": 0.45, "acc_norm_stderr": 0.049999999999999996 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.34558823529411764, "acc_stderr": 0.02888819310398864, "acc_norm": 0.34558823529411764, "acc_norm_stderr": 0.02888819310398864 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.32653061224489793, "acc_stderr": 0.030021056238440307, "acc_norm": 0.32653061224489793, "acc_norm_stderr": 0.030021056238440307 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.6371308016877637, "acc_stderr": 0.03129920825530213, "acc_norm": 0.6371308016877637, "acc_norm_stderr": 0.03129920825530213 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.31877444589308995, "acc_stderr": 0.011901895635786095, "acc_norm": 0.31877444589308995, "acc_norm_stderr": 0.011901895635786095 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.5, "acc_stderr": 0.03509312031717982, "acc_norm": 0.5, "acc_norm_stderr": 0.03509312031717982 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.5878787878787879, "acc_stderr": 0.03843566993588717, "acc_norm": 0.5878787878787879, "acc_norm_stderr": 0.03843566993588717 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2864137086903305, "mc1_stderr": 0.015826142439502332, "mc2": 0.4365731923271676, "mc2_stderr": 0.015091476648832229 }, "harness|ko_commongen_v2|2": { "acc": 0.5442739079102715, "acc_stderr": 0.017122829143292658, "acc_norm": 0.615112160566706, "acc_norm_stderr": 0.01672857970149864 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "AIFT/AIFT-ko-orca-plat-Yi-ko-6b-v1.6", "model_sha": "b0020e1098ac6f0562aea85ee0fc49064d3858ce", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }