|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.386518771331058, |
|
"acc_stderr": 0.014230084761910471, |
|
"acc_norm": 0.44880546075085326, |
|
"acc_norm_stderr": 0.014534599585097667 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.4251145190201155, |
|
"acc_stderr": 0.0049335002616835944, |
|
"acc_norm": 0.5650268870742879, |
|
"acc_norm_stderr": 0.004947402907996247 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.5029239766081871, |
|
"acc_stderr": 0.03834759370936839, |
|
"acc_norm": 0.5029239766081871, |
|
"acc_norm_stderr": 0.03834759370936839 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5825242718446602, |
|
"acc_stderr": 0.048828405482122375, |
|
"acc_norm": 0.5825242718446602, |
|
"acc_norm_stderr": 0.048828405482122375 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5606641123882503, |
|
"acc_stderr": 0.0177478742456836, |
|
"acc_norm": 0.5606641123882503, |
|
"acc_norm_stderr": 0.0177478742456836 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.45925925925925926, |
|
"acc_stderr": 0.04304979692464243, |
|
"acc_norm": 0.45925925925925926, |
|
"acc_norm_stderr": 0.04304979692464243 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.044619604333847415, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.044619604333847415 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.39148936170212767, |
|
"acc_stderr": 0.03190701242326812, |
|
"acc_norm": 0.39148936170212767, |
|
"acc_norm_stderr": 0.03190701242326812 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.42771084337349397, |
|
"acc_stderr": 0.03851597683718533, |
|
"acc_norm": 0.42771084337349397, |
|
"acc_norm_stderr": 0.03851597683718533 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.5209003215434084, |
|
"acc_stderr": 0.028373270961069414, |
|
"acc_norm": 0.5209003215434084, |
|
"acc_norm_stderr": 0.028373270961069414 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5022421524663677, |
|
"acc_stderr": 0.033557465352232634, |
|
"acc_norm": 0.5022421524663677, |
|
"acc_norm_stderr": 0.033557465352232634 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.5114503816793893, |
|
"acc_stderr": 0.043841400240780176, |
|
"acc_norm": 0.5114503816793893, |
|
"acc_norm_stderr": 0.043841400240780176 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5757575757575758, |
|
"acc_stderr": 0.03521224908841586, |
|
"acc_norm": 0.5757575757575758, |
|
"acc_norm_stderr": 0.03521224908841586 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.43448275862068964, |
|
"acc_stderr": 0.04130740879555497, |
|
"acc_norm": 0.43448275862068964, |
|
"acc_norm_stderr": 0.04130740879555497 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.2549019607843137, |
|
"acc_stderr": 0.043364327079931785, |
|
"acc_norm": 0.2549019607843137, |
|
"acc_norm_stderr": 0.043364327079931785 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.5126050420168067, |
|
"acc_stderr": 0.032468167657521745, |
|
"acc_norm": 0.5126050420168067, |
|
"acc_norm_stderr": 0.032468167657521745 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4717948717948718, |
|
"acc_stderr": 0.025310639254933914, |
|
"acc_norm": 0.4717948717948718, |
|
"acc_norm_stderr": 0.025310639254933914 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.51, |
|
"acc_stderr": 0.05024183937956912, |
|
"acc_norm": 0.51, |
|
"acc_norm_stderr": 0.05024183937956912 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.42, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.42, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.04833682445228318, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.04833682445228318 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.4482758620689655, |
|
"acc_stderr": 0.034991131376767445, |
|
"acc_norm": 0.4482758620689655, |
|
"acc_norm_stderr": 0.034991131376767445 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.4806451612903226, |
|
"acc_stderr": 0.0284226874043121, |
|
"acc_norm": 0.4806451612903226, |
|
"acc_norm_stderr": 0.0284226874043121 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6239316239316239, |
|
"acc_stderr": 0.03173393632969481, |
|
"acc_norm": 0.6239316239316239, |
|
"acc_norm_stderr": 0.03173393632969481 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.45660377358490567, |
|
"acc_stderr": 0.03065674869673943, |
|
"acc_norm": 0.45660377358490567, |
|
"acc_norm_stderr": 0.03065674869673943 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5454545454545454, |
|
"acc_stderr": 0.04769300568972744, |
|
"acc_norm": 0.5454545454545454, |
|
"acc_norm_stderr": 0.04769300568972744 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.2851851851851852, |
|
"acc_stderr": 0.027528599210340492, |
|
"acc_norm": 0.2851851851851852, |
|
"acc_norm_stderr": 0.027528599210340492 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.304635761589404, |
|
"acc_stderr": 0.03757949922943343, |
|
"acc_norm": 0.304635761589404, |
|
"acc_norm_stderr": 0.03757949922943343 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.5771144278606966, |
|
"acc_stderr": 0.034932317774212816, |
|
"acc_norm": 0.5771144278606966, |
|
"acc_norm_stderr": 0.034932317774212816 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.4393063583815029, |
|
"acc_stderr": 0.037842719328874674, |
|
"acc_norm": 0.4393063583815029, |
|
"acc_norm_stderr": 0.037842719328874674 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.3253968253968254, |
|
"acc_stderr": 0.024130158299762613, |
|
"acc_norm": 0.3253968253968254, |
|
"acc_norm_stderr": 0.024130158299762613 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4652777777777778, |
|
"acc_stderr": 0.04171115858181618, |
|
"acc_norm": 0.4652777777777778, |
|
"acc_norm_stderr": 0.04171115858181618 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.34, |
|
"acc_stderr": 0.04760952285695236, |
|
"acc_norm": 0.34, |
|
"acc_norm_stderr": 0.04760952285695236 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.64, |
|
"acc_stderr": 0.048241815132442176, |
|
"acc_norm": 0.64, |
|
"acc_norm_stderr": 0.048241815132442176 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5057803468208093, |
|
"acc_stderr": 0.026917296179149116, |
|
"acc_norm": 0.5057803468208093, |
|
"acc_norm_stderr": 0.026917296179149116 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.50920245398773, |
|
"acc_stderr": 0.03927705600787443, |
|
"acc_norm": 0.50920245398773, |
|
"acc_norm_stderr": 0.03927705600787443 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.49691358024691357, |
|
"acc_stderr": 0.027820214158594377, |
|
"acc_norm": 0.49691358024691357, |
|
"acc_norm_stderr": 0.027820214158594377 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.34, |
|
"acc_stderr": 0.04760952285695235, |
|
"acc_norm": 0.34, |
|
"acc_norm_stderr": 0.04760952285695235 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.5854922279792746, |
|
"acc_stderr": 0.035553003195576686, |
|
"acc_norm": 0.5854922279792746, |
|
"acc_norm_stderr": 0.035553003195576686 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.22807017543859648, |
|
"acc_stderr": 0.03947152782669415, |
|
"acc_norm": 0.22807017543859648, |
|
"acc_norm_stderr": 0.03947152782669415 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.6018348623853211, |
|
"acc_stderr": 0.02098798942265426, |
|
"acc_norm": 0.6018348623853211, |
|
"acc_norm_stderr": 0.02098798942265426 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.3492063492063492, |
|
"acc_stderr": 0.04263906892795132, |
|
"acc_norm": 0.3492063492063492, |
|
"acc_norm_stderr": 0.04263906892795132 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.45098039215686275, |
|
"acc_stderr": 0.028491993586171566, |
|
"acc_norm": 0.45098039215686275, |
|
"acc_norm_stderr": 0.028491993586171566 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.04923659639173309, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.04923659639173309 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6198347107438017, |
|
"acc_stderr": 0.04431324501968431, |
|
"acc_norm": 0.6198347107438017, |
|
"acc_norm_stderr": 0.04431324501968431 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.4276315789473684, |
|
"acc_stderr": 0.04026097083296559, |
|
"acc_norm": 0.4276315789473684, |
|
"acc_norm_stderr": 0.04026097083296559 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.38562091503267976, |
|
"acc_stderr": 0.019691459052354164, |
|
"acc_norm": 0.38562091503267976, |
|
"acc_norm_stderr": 0.019691459052354164 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3475177304964539, |
|
"acc_stderr": 0.02840662780959095, |
|
"acc_norm": 0.3475177304964539, |
|
"acc_norm_stderr": 0.02840662780959095 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.22321428571428573, |
|
"acc_stderr": 0.039523019677025116, |
|
"acc_norm": 0.22321428571428573, |
|
"acc_norm_stderr": 0.039523019677025116 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.4351851851851852, |
|
"acc_stderr": 0.033812000056435254, |
|
"acc_norm": 0.4351851851851852, |
|
"acc_norm_stderr": 0.033812000056435254 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2424581005586592, |
|
"acc_stderr": 0.01433352205921789, |
|
"acc_norm": 0.2424581005586592, |
|
"acc_norm_stderr": 0.01433352205921789 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.04943110704237102, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.04943110704237102 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.44485294117647056, |
|
"acc_stderr": 0.03018753206032938, |
|
"acc_norm": 0.44485294117647056, |
|
"acc_norm_stderr": 0.03018753206032938 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.5183673469387755, |
|
"acc_stderr": 0.03198761546763127, |
|
"acc_norm": 0.5183673469387755, |
|
"acc_norm_stderr": 0.03198761546763127 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.6244725738396625, |
|
"acc_stderr": 0.03152256243091156, |
|
"acc_norm": 0.6244725738396625, |
|
"acc_norm_stderr": 0.03152256243091156 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.35723598435462844, |
|
"acc_stderr": 0.012238615750316503, |
|
"acc_norm": 0.35723598435462844, |
|
"acc_norm_stderr": 0.012238615750316503 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.5049019607843137, |
|
"acc_stderr": 0.035091433756067866, |
|
"acc_norm": 0.5049019607843137, |
|
"acc_norm_stderr": 0.035091433756067866 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.5212121212121212, |
|
"acc_stderr": 0.03900828913737301, |
|
"acc_norm": 0.5212121212121212, |
|
"acc_norm_stderr": 0.03900828913737301 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2974296205630355, |
|
"mc1_stderr": 0.01600265148736101, |
|
"mc2": 0.45982516329816536, |
|
"mc2_stderr": 0.014857750246900359 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.46871310507674147, |
|
"acc_stderr": 0.017156666859785463, |
|
"acc_norm": 0.5442739079102715, |
|
"acc_norm_stderr": 0.017122829143292658 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "LDCC/LDCC-Instruct-Llama-2-ko-13B-v1.0", |
|
"model_sha": "9d704abe2ee490446c4bea6a94692841bdb92ddb", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |