{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.2901023890784983, "acc_stderr": 0.013261573677520773, "acc_norm": 0.3412969283276451, "acc_norm_stderr": 0.01385583128749772 }, "harness|ko_hellaswag|10": { "acc": 0.33827922724556864, "acc_stderr": 0.004721571443354458, "acc_norm": 0.4235212109141605, "acc_norm_stderr": 0.0049310654341736815 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.39766081871345027, "acc_stderr": 0.0375363895576169, "acc_norm": 0.39766081871345027, "acc_norm_stderr": 0.0375363895576169 }, "harness|ko_mmlu_management|5": { "acc": 0.5048543689320388, "acc_stderr": 0.04950504382128921, "acc_norm": 0.5048543689320388, "acc_norm_stderr": 0.04950504382128921 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.41890166028097064, "acc_stderr": 0.017643205052377167, "acc_norm": 0.41890166028097064, "acc_norm_stderr": 0.017643205052377167 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.3333333333333333, "acc_stderr": 0.04072314811876837, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.04072314811876837 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.28, "acc_stderr": 0.045126085985421276, "acc_norm": 0.28, "acc_norm_stderr": 0.045126085985421276 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.42127659574468085, "acc_stderr": 0.03227834510146267, "acc_norm": 0.42127659574468085, "acc_norm_stderr": 0.03227834510146267 }, "harness|ko_mmlu_virology|5": { "acc": 0.3795180722891566, "acc_stderr": 0.037777988227480165, "acc_norm": 0.3795180722891566, "acc_norm_stderr": 0.037777988227480165 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.38263665594855306, "acc_stderr": 0.02760468902858198, "acc_norm": 0.38263665594855306, "acc_norm_stderr": 0.02760468902858198 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.3901345291479821, "acc_stderr": 0.03273766725459157, "acc_norm": 0.3901345291479821, "acc_norm_stderr": 0.03273766725459157 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.40458015267175573, "acc_stderr": 0.043046937953806645, "acc_norm": 0.40458015267175573, "acc_norm_stderr": 0.043046937953806645 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.37, "acc_stderr": 0.04852365870939099, "acc_norm": 0.37, "acc_norm_stderr": 0.04852365870939099 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.4444444444444444, "acc_stderr": 0.035402943770953675, "acc_norm": 0.4444444444444444, "acc_norm_stderr": 0.035402943770953675 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.45517241379310347, "acc_stderr": 0.04149886942192117, "acc_norm": 0.45517241379310347, "acc_norm_stderr": 0.04149886942192117 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.24509803921568626, "acc_stderr": 0.042801058373643966, "acc_norm": 0.24509803921568626, "acc_norm_stderr": 0.042801058373643966 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.44537815126050423, "acc_stderr": 0.032284106267163895, "acc_norm": 0.44537815126050423, "acc_norm_stderr": 0.032284106267163895 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.36153846153846153, "acc_stderr": 0.02435958146539697, "acc_norm": 0.36153846153846153, "acc_norm_stderr": 0.02435958146539697 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.53, "acc_stderr": 0.05016135580465919, "acc_norm": 0.53, "acc_norm_stderr": 0.05016135580465919 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.32, "acc_stderr": 0.04688261722621505, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621505 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.48148148148148145, "acc_stderr": 0.04830366024635331, "acc_norm": 0.48148148148148145, "acc_norm_stderr": 0.04830366024635331 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.33004926108374383, "acc_stderr": 0.03308530426228258, "acc_norm": 0.33004926108374383, "acc_norm_stderr": 0.03308530426228258 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.36451612903225805, "acc_stderr": 0.027379871229943252, "acc_norm": 0.36451612903225805, "acc_norm_stderr": 0.027379871229943252 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6196581196581197, "acc_stderr": 0.031804252043840985, "acc_norm": 0.6196581196581197, "acc_norm_stderr": 0.031804252043840985 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.37358490566037733, "acc_stderr": 0.02977308271331988, "acc_norm": 0.37358490566037733, "acc_norm_stderr": 0.02977308271331988 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.45454545454545453, "acc_stderr": 0.04769300568972743, "acc_norm": 0.45454545454545453, "acc_norm_stderr": 0.04769300568972743 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.3074074074074074, "acc_stderr": 0.028133252578815646, "acc_norm": 0.3074074074074074, "acc_norm_stderr": 0.028133252578815646 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.33112582781456956, "acc_stderr": 0.038425817186598696, "acc_norm": 0.33112582781456956, "acc_norm_stderr": 0.038425817186598696 }, "harness|ko_mmlu_sociology|5": { "acc": 0.5522388059701493, "acc_stderr": 0.03516184772952167, "acc_norm": 0.5522388059701493, "acc_norm_stderr": 0.03516184772952167 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.32947976878612717, "acc_stderr": 0.03583901754736413, "acc_norm": 0.32947976878612717, "acc_norm_stderr": 0.03583901754736413 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.3333333333333333, "acc_stderr": 0.024278568024307695, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.024278568024307695 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2708333333333333, "acc_stderr": 0.03716177437566018, "acc_norm": 0.2708333333333333, "acc_norm_stderr": 0.03716177437566018 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.29, "acc_stderr": 0.045604802157206845, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206845 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.62, "acc_stderr": 0.04878317312145633, "acc_norm": 0.62, "acc_norm_stderr": 0.04878317312145633 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.43352601156069365, "acc_stderr": 0.026680134761679214, "acc_norm": 0.43352601156069365, "acc_norm_stderr": 0.026680134761679214 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.4294478527607362, "acc_stderr": 0.038890666191127216, "acc_norm": 0.4294478527607362, "acc_norm_stderr": 0.038890666191127216 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.41975308641975306, "acc_stderr": 0.027460099557005135, "acc_norm": 0.41975308641975306, "acc_norm_stderr": 0.027460099557005135 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.43005181347150256, "acc_stderr": 0.03572954333144808, "acc_norm": 0.43005181347150256, "acc_norm_stderr": 0.03572954333144808 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2631578947368421, "acc_stderr": 0.04142439719489362, "acc_norm": 0.2631578947368421, "acc_norm_stderr": 0.04142439719489362 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.4018348623853211, "acc_stderr": 0.02102010617299701, "acc_norm": 0.4018348623853211, "acc_norm_stderr": 0.02102010617299701 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.3333333333333333, "acc_stderr": 0.04216370213557835, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.04216370213557835 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.40522875816993464, "acc_stderr": 0.028110928492809075, "acc_norm": 0.40522875816993464, "acc_norm_stderr": 0.028110928492809075 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.43, "acc_stderr": 0.04975698519562428, "acc_norm": 0.43, "acc_norm_stderr": 0.04975698519562428 }, "harness|ko_mmlu_international_law|5": { "acc": 0.5454545454545454, "acc_stderr": 0.045454545454545456, "acc_norm": 0.5454545454545454, "acc_norm_stderr": 0.045454545454545456 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.32894736842105265, "acc_stderr": 0.03823428969926606, "acc_norm": 0.32894736842105265, "acc_norm_stderr": 0.03823428969926606 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.32189542483660133, "acc_stderr": 0.018901015322093095, "acc_norm": 0.32189542483660133, "acc_norm_stderr": 0.018901015322093095 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.36524822695035464, "acc_stderr": 0.02872386385328128, "acc_norm": 0.36524822695035464, "acc_norm_stderr": 0.02872386385328128 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.36607142857142855, "acc_stderr": 0.0457237235873743, "acc_norm": 0.36607142857142855, "acc_norm_stderr": 0.0457237235873743 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.3287037037037037, "acc_stderr": 0.03203614084670058, "acc_norm": 0.3287037037037037, "acc_norm_stderr": 0.03203614084670058 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.25921787709497207, "acc_stderr": 0.014655780837497722, "acc_norm": 0.25921787709497207, "acc_norm_stderr": 0.014655780837497722 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.33, "acc_stderr": 0.04725815626252605, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252605 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.52, "acc_stderr": 0.05021167315686779, "acc_norm": 0.52, "acc_norm_stderr": 0.05021167315686779 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.35661764705882354, "acc_stderr": 0.02909720956841196, "acc_norm": 0.35661764705882354, "acc_norm_stderr": 0.02909720956841196 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.46530612244897956, "acc_stderr": 0.03193207024425314, "acc_norm": 0.46530612244897956, "acc_norm_stderr": 0.03193207024425314 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.5063291139240507, "acc_stderr": 0.0325446201076786, "acc_norm": 0.5063291139240507, "acc_norm_stderr": 0.0325446201076786 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.2940026075619296, "acc_stderr": 0.011636062953698602, "acc_norm": 0.2940026075619296, "acc_norm_stderr": 0.011636062953698602 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.23039215686274508, "acc_stderr": 0.02955429260569506, "acc_norm": 0.23039215686274508, "acc_norm_stderr": 0.02955429260569506 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.2606060606060606, "acc_stderr": 0.03427743175816524, "acc_norm": 0.2606060606060606, "acc_norm_stderr": 0.03427743175816524 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2876376988984088, "mc1_stderr": 0.01584631510139481, "mc2": 0.45464957639603176, "mc2_stderr": 0.015510382231630185 }, "harness|ko_commongen_v2|2": { "acc": 0.31995277449822906, "acc_stderr": 0.01603715384028053, "acc_norm": 0.3837072018890201, "acc_norm_stderr": 0.01671892463723183 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Edentns/DataVortexM-7B-Instruct-v0.1", "model_sha": "8a43dc63c63ee8e03297327e2fc3e040c3997c0c", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }