{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.20819112627986347, "acc_stderr": 0.011864866118448069, "acc_norm": 0.2508532423208191, "acc_norm_stderr": 0.012668198621315433 }, "harness|ko_hellaswag|10": { "acc": 0.2610037841067516, "acc_stderr": 0.004382844128643425, "acc_norm": 0.26767576180043817, "acc_norm_stderr": 0.004418427613296678 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.21052631578947367, "acc_stderr": 0.0312678171466318, "acc_norm": 0.21052631578947367, "acc_norm_stderr": 0.0312678171466318 }, "harness|ko_mmlu_management|5": { "acc": 0.2524271844660194, "acc_stderr": 0.04301250399690877, "acc_norm": 0.2524271844660194, "acc_norm_stderr": 0.04301250399690877 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.2707535121328225, "acc_stderr": 0.015889888362560486, "acc_norm": 0.2707535121328225, "acc_norm_stderr": 0.015889888362560486 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.23703703703703705, "acc_stderr": 0.03673731683969506, "acc_norm": 0.23703703703703705, "acc_norm_stderr": 0.03673731683969506 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.28, "acc_stderr": 0.04512608598542127, "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542127 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.28085106382978725, "acc_stderr": 0.02937917046412482, "acc_norm": 0.28085106382978725, "acc_norm_stderr": 0.02937917046412482 }, "harness|ko_mmlu_virology|5": { "acc": 0.23493975903614459, "acc_stderr": 0.03300533186128922, "acc_norm": 0.23493975903614459, "acc_norm_stderr": 0.03300533186128922 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.2282958199356913, "acc_stderr": 0.023839303311398195, "acc_norm": 0.2282958199356913, "acc_norm_stderr": 0.023839303311398195 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.25112107623318386, "acc_stderr": 0.02910522083322462, "acc_norm": 0.25112107623318386, "acc_norm_stderr": 0.02910522083322462 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.1984732824427481, "acc_stderr": 0.034981493854624714, "acc_norm": 0.1984732824427481, "acc_norm_stderr": 0.034981493854624714 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.29, "acc_stderr": 0.04560480215720684, "acc_norm": 0.29, "acc_norm_stderr": 0.04560480215720684 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.25252525252525254, "acc_stderr": 0.030954055470365907, "acc_norm": 0.25252525252525254, "acc_norm_stderr": 0.030954055470365907 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.2620689655172414, "acc_stderr": 0.036646663372252565, "acc_norm": 0.2620689655172414, "acc_norm_stderr": 0.036646663372252565 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.18627450980392157, "acc_stderr": 0.03873958714149351, "acc_norm": 0.18627450980392157, "acc_norm_stderr": 0.03873958714149351 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.2815126050420168, "acc_stderr": 0.02921354941437216, "acc_norm": 0.2815126050420168, "acc_norm_stderr": 0.02921354941437216 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.26666666666666666, "acc_stderr": 0.022421273612923714, "acc_norm": 0.26666666666666666, "acc_norm_stderr": 0.022421273612923714 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.23, "acc_stderr": 0.04229525846816506, "acc_norm": 0.23, "acc_norm_stderr": 0.04229525846816506 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.19, "acc_stderr": 0.03942772444036623, "acc_norm": 0.19, "acc_norm_stderr": 0.03942772444036623 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.2222222222222222, "acc_stderr": 0.040191074725573483, "acc_norm": 0.2222222222222222, "acc_norm_stderr": 0.040191074725573483 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3103448275862069, "acc_stderr": 0.03255086769970103, "acc_norm": 0.3103448275862069, "acc_norm_stderr": 0.03255086769970103 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.31290322580645163, "acc_stderr": 0.026377567028645858, "acc_norm": 0.31290322580645163, "acc_norm_stderr": 0.026377567028645858 }, "harness|ko_mmlu_marketing|5": { "acc": 0.19230769230769232, "acc_stderr": 0.025819233256483727, "acc_norm": 0.19230769230769232, "acc_norm_stderr": 0.025819233256483727 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.23773584905660378, "acc_stderr": 0.026199808807561918, "acc_norm": 0.23773584905660378, "acc_norm_stderr": 0.026199808807561918 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.2909090909090909, "acc_stderr": 0.04350271442923243, "acc_norm": 0.2909090909090909, "acc_norm_stderr": 0.04350271442923243 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.26296296296296295, "acc_stderr": 0.026842057873833706, "acc_norm": 0.26296296296296295, "acc_norm_stderr": 0.026842057873833706 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.32450331125827814, "acc_stderr": 0.03822746937658754, "acc_norm": 0.32450331125827814, "acc_norm_stderr": 0.03822746937658754 }, "harness|ko_mmlu_sociology|5": { "acc": 0.208955223880597, "acc_stderr": 0.028748298931728655, "acc_norm": 0.208955223880597, "acc_norm_stderr": 0.028748298931728655 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.21965317919075145, "acc_stderr": 0.031568093627031744, "acc_norm": 0.21965317919075145, "acc_norm_stderr": 0.031568093627031744 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.25132275132275134, "acc_stderr": 0.022340482339643898, "acc_norm": 0.25132275132275134, "acc_norm_stderr": 0.022340482339643898 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2361111111111111, "acc_stderr": 0.03551446610810826, "acc_norm": 0.2361111111111111, "acc_norm_stderr": 0.03551446610810826 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.26, "acc_stderr": 0.044084400227680794, "acc_norm": 0.26, "acc_norm_stderr": 0.044084400227680794 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.24566473988439305, "acc_stderr": 0.02317629820399201, "acc_norm": 0.24566473988439305, "acc_norm_stderr": 0.02317629820399201 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.294478527607362, "acc_stderr": 0.03581165790474082, "acc_norm": 0.294478527607362, "acc_norm_stderr": 0.03581165790474082 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.2623456790123457, "acc_stderr": 0.024477222856135114, "acc_norm": 0.2623456790123457, "acc_norm_stderr": 0.024477222856135114 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.3316062176165803, "acc_stderr": 0.03397636541089116, "acc_norm": 0.3316062176165803, "acc_norm_stderr": 0.03397636541089116 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.22807017543859648, "acc_stderr": 0.03947152782669415, "acc_norm": 0.22807017543859648, "acc_norm_stderr": 0.03947152782669415 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.24036697247706423, "acc_stderr": 0.01832060732096407, "acc_norm": 0.24036697247706423, "acc_norm_stderr": 0.01832060732096407 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.15873015873015872, "acc_stderr": 0.03268454013011744, "acc_norm": 0.15873015873015872, "acc_norm_stderr": 0.03268454013011744 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.23529411764705882, "acc_stderr": 0.024288619466046105, "acc_norm": 0.23529411764705882, "acc_norm_stderr": 0.024288619466046105 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.28, "acc_stderr": 0.04512608598542127, "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542127 }, "harness|ko_mmlu_international_law|5": { "acc": 0.32231404958677684, "acc_stderr": 0.04266416363352167, "acc_norm": 0.32231404958677684, "acc_norm_stderr": 0.04266416363352167 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.17763157894736842, "acc_stderr": 0.031103182383123363, "acc_norm": 0.17763157894736842, "acc_norm_stderr": 0.031103182383123363 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.21568627450980393, "acc_stderr": 0.016639319350313264, "acc_norm": 0.21568627450980393, "acc_norm_stderr": 0.016639319350313264 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.2375886524822695, "acc_stderr": 0.025389512552729903, "acc_norm": 0.2375886524822695, "acc_norm_stderr": 0.025389512552729903 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.20535714285714285, "acc_stderr": 0.038342410214190714, "acc_norm": 0.20535714285714285, "acc_norm_stderr": 0.038342410214190714 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.47685185185185186, "acc_stderr": 0.03406315360711507, "acc_norm": 0.47685185185185186, "acc_norm_stderr": 0.03406315360711507 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.24581005586592178, "acc_stderr": 0.014400296429225605, "acc_norm": 0.24581005586592178, "acc_norm_stderr": 0.014400296429225605 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.33, "acc_stderr": 0.04725815626252604, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252604 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.27, "acc_stderr": 0.0446196043338474, "acc_norm": 0.27, "acc_norm_stderr": 0.0446196043338474 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.4485294117647059, "acc_stderr": 0.030211479609121593, "acc_norm": 0.4485294117647059, "acc_norm_stderr": 0.030211479609121593 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.2653061224489796, "acc_stderr": 0.028263889943784603, "acc_norm": 0.2653061224489796, "acc_norm_stderr": 0.028263889943784603 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.25316455696202533, "acc_stderr": 0.02830465794303529, "acc_norm": 0.25316455696202533, "acc_norm_stderr": 0.02830465794303529 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.2379400260756193, "acc_stderr": 0.010875700787694238, "acc_norm": 0.2379400260756193, "acc_norm_stderr": 0.010875700787694238 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.2647058823529412, "acc_stderr": 0.03096451792692341, "acc_norm": 0.2647058823529412, "acc_norm_stderr": 0.03096451792692341 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.24242424242424243, "acc_stderr": 0.03346409881055953, "acc_norm": 0.24242424242424243, "acc_norm_stderr": 0.03346409881055953 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2766217870257038, "mc1_stderr": 0.015659605755326923, "mc2": 0.4976972719721252, "mc2_stderr": 0.01632084070201214 }, "harness|ko_commongen_v2|2": { "acc": 0.21009389671361503, "acc_stderr": 0.013964637699696686, "acc_norm": 0.4694835680751174, "acc_norm_stderr": 0.017107826542381244 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "amphora/xllama-7b-0.3", "model_sha": "a2c409d42ade00bb1e4dbd1815d94618a6fa23ed", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }