{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.3447098976109215, "acc_stderr": 0.01388881628678211, "acc_norm": 0.38993174061433444, "acc_norm_stderr": 0.014252959848892896 }, "harness|ko_hellaswag|10": { "acc": 0.3745269866560446, "acc_stderr": 0.004830113797327052, "acc_norm": 0.47998406691894047, "acc_norm_stderr": 0.00498578162046701 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.47368421052631576, "acc_stderr": 0.038295098689947286, "acc_norm": 0.47368421052631576, "acc_norm_stderr": 0.038295098689947286 }, "harness|ko_mmlu_management|5": { "acc": 0.5339805825242718, "acc_stderr": 0.04939291447273482, "acc_norm": 0.5339805825242718, "acc_norm_stderr": 0.04939291447273482 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.46871008939974457, "acc_stderr": 0.017844918090468537, "acc_norm": 0.46871008939974457, "acc_norm_stderr": 0.017844918090468537 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.37037037037037035, "acc_stderr": 0.04171654161354543, "acc_norm": 0.37037037037037035, "acc_norm_stderr": 0.04171654161354543 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.28, "acc_stderr": 0.04512608598542128, "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542128 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.3829787234042553, "acc_stderr": 0.03177821250236922, "acc_norm": 0.3829787234042553, "acc_norm_stderr": 0.03177821250236922 }, "harness|ko_mmlu_virology|5": { "acc": 0.4036144578313253, "acc_stderr": 0.03819486140758397, "acc_norm": 0.4036144578313253, "acc_norm_stderr": 0.03819486140758397 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.4405144694533762, "acc_stderr": 0.028196400574197426, "acc_norm": 0.4405144694533762, "acc_norm_stderr": 0.028196400574197426 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.4618834080717489, "acc_stderr": 0.03346015011973228, "acc_norm": 0.4618834080717489, "acc_norm_stderr": 0.03346015011973228 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.44274809160305345, "acc_stderr": 0.043564472026650695, "acc_norm": 0.44274809160305345, "acc_norm_stderr": 0.043564472026650695 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5252525252525253, "acc_stderr": 0.03557806245087314, "acc_norm": 0.5252525252525253, "acc_norm_stderr": 0.03557806245087314 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.45517241379310347, "acc_stderr": 0.04149886942192117, "acc_norm": 0.45517241379310347, "acc_norm_stderr": 0.04149886942192117 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.20588235294117646, "acc_stderr": 0.04023382273617746, "acc_norm": 0.20588235294117646, "acc_norm_stderr": 0.04023382273617746 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.48739495798319327, "acc_stderr": 0.03246816765752174, "acc_norm": 0.48739495798319327, "acc_norm_stderr": 0.03246816765752174 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.4641025641025641, "acc_stderr": 0.025285585990017834, "acc_norm": 0.4641025641025641, "acc_norm_stderr": 0.025285585990017834 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.58, "acc_stderr": 0.049604496374885836, "acc_norm": 0.58, "acc_norm_stderr": 0.049604496374885836 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.26, "acc_stderr": 0.0440844002276808, "acc_norm": 0.26, "acc_norm_stderr": 0.0440844002276808 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5370370370370371, "acc_stderr": 0.04820403072760627, "acc_norm": 0.5370370370370371, "acc_norm_stderr": 0.04820403072760627 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.35467980295566504, "acc_stderr": 0.0336612448905145, "acc_norm": 0.35467980295566504, "acc_norm_stderr": 0.0336612448905145 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.4096774193548387, "acc_stderr": 0.027976054915347354, "acc_norm": 0.4096774193548387, "acc_norm_stderr": 0.027976054915347354 }, "harness|ko_mmlu_marketing|5": { "acc": 0.7606837606837606, "acc_stderr": 0.027951826808924336, "acc_norm": 0.7606837606837606, "acc_norm_stderr": 0.027951826808924336 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.4641509433962264, "acc_stderr": 0.030693675018458003, "acc_norm": 0.4641509433962264, "acc_norm_stderr": 0.030693675018458003 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.4636363636363636, "acc_stderr": 0.04776449162396197, "acc_norm": 0.4636363636363636, "acc_norm_stderr": 0.04776449162396197 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.34814814814814815, "acc_stderr": 0.029045600290616255, "acc_norm": 0.34814814814814815, "acc_norm_stderr": 0.029045600290616255 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.304635761589404, "acc_stderr": 0.03757949922943342, "acc_norm": 0.304635761589404, "acc_norm_stderr": 0.03757949922943342 }, "harness|ko_mmlu_sociology|5": { "acc": 0.6119402985074627, "acc_stderr": 0.0344578996436275, "acc_norm": 0.6119402985074627, "acc_norm_stderr": 0.0344578996436275 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.41040462427745666, "acc_stderr": 0.03750757044895537, "acc_norm": 0.41040462427745666, "acc_norm_stderr": 0.03750757044895537 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.335978835978836, "acc_stderr": 0.024326310529149138, "acc_norm": 0.335978835978836, "acc_norm_stderr": 0.024326310529149138 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3888888888888889, "acc_stderr": 0.04076663253918567, "acc_norm": 0.3888888888888889, "acc_norm_stderr": 0.04076663253918567 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.35, "acc_stderr": 0.047937248544110196, "acc_norm": 0.35, "acc_norm_stderr": 0.047937248544110196 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.52, "acc_stderr": 0.050211673156867795, "acc_norm": 0.52, "acc_norm_stderr": 0.050211673156867795 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.45375722543352603, "acc_stderr": 0.02680372058320619, "acc_norm": 0.45375722543352603, "acc_norm_stderr": 0.02680372058320619 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.4539877300613497, "acc_stderr": 0.0391170190467718, "acc_norm": 0.4539877300613497, "acc_norm_stderr": 0.0391170190467718 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.4783950617283951, "acc_stderr": 0.027794760105008746, "acc_norm": 0.4783950617283951, "acc_norm_stderr": 0.027794760105008746 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.5129533678756477, "acc_stderr": 0.0360722806104775, "acc_norm": 0.5129533678756477, "acc_norm_stderr": 0.0360722806104775 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.24561403508771928, "acc_stderr": 0.0404933929774814, "acc_norm": 0.24561403508771928, "acc_norm_stderr": 0.0404933929774814 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.4917431192660551, "acc_stderr": 0.021434399918214334, "acc_norm": 0.4917431192660551, "acc_norm_stderr": 0.021434399918214334 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.4365079365079365, "acc_stderr": 0.04435932892851466, "acc_norm": 0.4365079365079365, "acc_norm_stderr": 0.04435932892851466 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.4477124183006536, "acc_stderr": 0.028472938478033526, "acc_norm": 0.4477124183006536, "acc_norm_stderr": 0.028472938478033526 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.44, "acc_stderr": 0.049888765156985884, "acc_norm": 0.44, "acc_norm_stderr": 0.049888765156985884 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6446280991735537, "acc_stderr": 0.0436923632657398, "acc_norm": 0.6446280991735537, "acc_norm_stderr": 0.0436923632657398 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.40131578947368424, "acc_stderr": 0.03988903703336285, "acc_norm": 0.40131578947368424, "acc_norm_stderr": 0.03988903703336285 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.36437908496732024, "acc_stderr": 0.019469518221573695, "acc_norm": 0.36437908496732024, "acc_norm_stderr": 0.019469518221573695 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.35106382978723405, "acc_stderr": 0.02847350127296376, "acc_norm": 0.35106382978723405, "acc_norm_stderr": 0.02847350127296376 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.44642857142857145, "acc_stderr": 0.047184714852195886, "acc_norm": 0.44642857142857145, "acc_norm_stderr": 0.047184714852195886 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.3148148148148148, "acc_stderr": 0.03167468706828979, "acc_norm": 0.3148148148148148, "acc_norm_stderr": 0.03167468706828979 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.27932960893854747, "acc_stderr": 0.015005762446786171, "acc_norm": 0.27932960893854747, "acc_norm_stderr": 0.015005762446786171 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.38, "acc_stderr": 0.04878317312145632, "acc_norm": 0.38, "acc_norm_stderr": 0.04878317312145632 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.61, "acc_stderr": 0.04902071300001975, "acc_norm": 0.61, "acc_norm_stderr": 0.04902071300001975 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.3602941176470588, "acc_stderr": 0.029163128570670733, "acc_norm": 0.3602941176470588, "acc_norm_stderr": 0.029163128570670733 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.49387755102040815, "acc_stderr": 0.03200682020163909, "acc_norm": 0.49387755102040815, "acc_norm_stderr": 0.03200682020163909 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.5949367088607594, "acc_stderr": 0.031955147413706725, "acc_norm": 0.5949367088607594, "acc_norm_stderr": 0.031955147413706725 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3344198174706649, "acc_stderr": 0.012049668983214938, "acc_norm": 0.3344198174706649, "acc_norm_stderr": 0.012049668983214938 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.4068627450980392, "acc_stderr": 0.03447891136353382, "acc_norm": 0.4068627450980392, "acc_norm_stderr": 0.03447891136353382 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.3696969696969697, "acc_stderr": 0.037694303145125674, "acc_norm": 0.3696969696969697, "acc_norm_stderr": 0.037694303145125674 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.3708690330477356, "mc1_stderr": 0.016909693580248807, "mc2": 0.5496429922579386, "mc2_stderr": 0.016045283495853307 }, "harness|ko_commongen_v2|2": { "acc": 0.4167650531286895, "acc_stderr": 0.016950489146108826, "acc_norm": 0.4380165289256198, "acc_norm_stderr": 0.017057753702160283 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Intel/neural-chat-7b-v3-1", "model_sha": "c70aa428800d151f4eae2b6d4b6a08c773868987", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }