|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.7056313993174061, |
|
"acc_stderr": 0.013318528460539422, |
|
"acc_norm": 0.7517064846416383, |
|
"acc_norm_stderr": 0.012624912868089762 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.5329615614419438, |
|
"acc_stderr": 0.00497892716479289, |
|
"acc_norm": 0.6955785700059749, |
|
"acc_norm_stderr": 0.004592215118295274 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.7309941520467836, |
|
"acc_stderr": 0.03401052620104089, |
|
"acc_norm": 0.7309941520467836, |
|
"acc_norm_stderr": 0.03401052620104089 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.7669902912621359, |
|
"acc_stderr": 0.04185832598928315, |
|
"acc_norm": 0.7669902912621359, |
|
"acc_norm_stderr": 0.04185832598928315 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.7369093231162197, |
|
"acc_stderr": 0.015745497169049046, |
|
"acc_norm": 0.7369093231162197, |
|
"acc_norm_stderr": 0.015745497169049046 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.45185185185185184, |
|
"acc_stderr": 0.04299268905480863, |
|
"acc_norm": 0.45185185185185184, |
|
"acc_norm_stderr": 0.04299268905480863 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.048523658709391, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.048523658709391 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.5404255319148936, |
|
"acc_stderr": 0.03257901482099834, |
|
"acc_norm": 0.5404255319148936, |
|
"acc_norm_stderr": 0.03257901482099834 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.5120481927710844, |
|
"acc_stderr": 0.03891364495835817, |
|
"acc_norm": 0.5120481927710844, |
|
"acc_norm_stderr": 0.03891364495835817 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.6591639871382636, |
|
"acc_stderr": 0.026920841260776155, |
|
"acc_norm": 0.6591639871382636, |
|
"acc_norm_stderr": 0.026920841260776155 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.672645739910314, |
|
"acc_stderr": 0.03149384670994131, |
|
"acc_norm": 0.672645739910314, |
|
"acc_norm_stderr": 0.03149384670994131 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.6106870229007634, |
|
"acc_stderr": 0.042764865428145914, |
|
"acc_norm": 0.6106870229007634, |
|
"acc_norm_stderr": 0.042764865428145914 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.51, |
|
"acc_stderr": 0.050241839379569095, |
|
"acc_norm": 0.51, |
|
"acc_norm_stderr": 0.050241839379569095 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.7626262626262627, |
|
"acc_stderr": 0.030313710538198906, |
|
"acc_norm": 0.7626262626262627, |
|
"acc_norm_stderr": 0.030313710538198906 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.5655172413793104, |
|
"acc_stderr": 0.04130740879555497, |
|
"acc_norm": 0.5655172413793104, |
|
"acc_norm_stderr": 0.04130740879555497 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.35294117647058826, |
|
"acc_stderr": 0.047551296160629475, |
|
"acc_norm": 0.35294117647058826, |
|
"acc_norm_stderr": 0.047551296160629475 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.6554621848739496, |
|
"acc_stderr": 0.030868682604121622, |
|
"acc_norm": 0.6554621848739496, |
|
"acc_norm_stderr": 0.030868682604121622 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.6487179487179487, |
|
"acc_stderr": 0.024203665177902806, |
|
"acc_norm": 0.6487179487179487, |
|
"acc_norm_stderr": 0.024203665177902806 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.75, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.75, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.34, |
|
"acc_stderr": 0.04760952285695235, |
|
"acc_norm": 0.34, |
|
"acc_norm_stderr": 0.04760952285695235 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6851851851851852, |
|
"acc_stderr": 0.04489931073591312, |
|
"acc_norm": 0.6851851851851852, |
|
"acc_norm_stderr": 0.04489931073591312 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.46798029556650245, |
|
"acc_stderr": 0.03510766597959217, |
|
"acc_norm": 0.46798029556650245, |
|
"acc_norm_stderr": 0.03510766597959217 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.635483870967742, |
|
"acc_stderr": 0.02737987122994324, |
|
"acc_norm": 0.635483870967742, |
|
"acc_norm_stderr": 0.02737987122994324 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.8589743589743589, |
|
"acc_stderr": 0.02280138253459753, |
|
"acc_norm": 0.8589743589743589, |
|
"acc_norm_stderr": 0.02280138253459753 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.6075471698113207, |
|
"acc_stderr": 0.03005258057955785, |
|
"acc_norm": 0.6075471698113207, |
|
"acc_norm_stderr": 0.03005258057955785 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.6363636363636364, |
|
"acc_stderr": 0.04607582090719976, |
|
"acc_norm": 0.6363636363636364, |
|
"acc_norm_stderr": 0.04607582090719976 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3925925925925926, |
|
"acc_stderr": 0.029773847012532967, |
|
"acc_norm": 0.3925925925925926, |
|
"acc_norm_stderr": 0.029773847012532967 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3708609271523179, |
|
"acc_stderr": 0.03943966699183629, |
|
"acc_norm": 0.3708609271523179, |
|
"acc_norm_stderr": 0.03943966699183629 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.7562189054726368, |
|
"acc_stderr": 0.03036049015401465, |
|
"acc_norm": 0.7562189054726368, |
|
"acc_norm_stderr": 0.03036049015401465 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.5664739884393064, |
|
"acc_stderr": 0.03778621079092055, |
|
"acc_norm": 0.5664739884393064, |
|
"acc_norm_stderr": 0.03778621079092055 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.4576719576719577, |
|
"acc_stderr": 0.02565886886205832, |
|
"acc_norm": 0.4576719576719577, |
|
"acc_norm_stderr": 0.02565886886205832 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.6180555555555556, |
|
"acc_stderr": 0.040629907841466674, |
|
"acc_norm": 0.6180555555555556, |
|
"acc_norm_stderr": 0.040629907841466674 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.48, |
|
"acc_stderr": 0.050211673156867795, |
|
"acc_norm": 0.48, |
|
"acc_norm_stderr": 0.050211673156867795 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.78, |
|
"acc_stderr": 0.04163331998932264, |
|
"acc_norm": 0.78, |
|
"acc_norm_stderr": 0.04163331998932264 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.630057803468208, |
|
"acc_stderr": 0.025992472029306393, |
|
"acc_norm": 0.630057803468208, |
|
"acc_norm_stderr": 0.025992472029306393 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.656441717791411, |
|
"acc_stderr": 0.03731133519673893, |
|
"acc_norm": 0.656441717791411, |
|
"acc_norm_stderr": 0.03731133519673893 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.6944444444444444, |
|
"acc_stderr": 0.02563082497562135, |
|
"acc_norm": 0.6944444444444444, |
|
"acc_norm_stderr": 0.02563082497562135 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.7512953367875648, |
|
"acc_stderr": 0.031195840877700304, |
|
"acc_norm": 0.7512953367875648, |
|
"acc_norm_stderr": 0.031195840877700304 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.047036043419179864, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.047036043419179864 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.7834862385321101, |
|
"acc_stderr": 0.01765871059444314, |
|
"acc_norm": 0.7834862385321101, |
|
"acc_norm_stderr": 0.01765871059444314 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.4365079365079365, |
|
"acc_stderr": 0.04435932892851466, |
|
"acc_norm": 0.4365079365079365, |
|
"acc_norm_stderr": 0.04435932892851466 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.6568627450980392, |
|
"acc_stderr": 0.027184498909941616, |
|
"acc_norm": 0.6568627450980392, |
|
"acc_norm_stderr": 0.027184498909941616 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.73, |
|
"acc_stderr": 0.044619604333847394, |
|
"acc_norm": 0.73, |
|
"acc_norm_stderr": 0.044619604333847394 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.7933884297520661, |
|
"acc_stderr": 0.03695980128098824, |
|
"acc_norm": 0.7933884297520661, |
|
"acc_norm_stderr": 0.03695980128098824 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.6578947368421053, |
|
"acc_stderr": 0.038607315993160904, |
|
"acc_norm": 0.6578947368421053, |
|
"acc_norm_stderr": 0.038607315993160904 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.6143790849673203, |
|
"acc_stderr": 0.01969145905235404, |
|
"acc_norm": 0.6143790849673203, |
|
"acc_norm_stderr": 0.01969145905235404 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.450354609929078, |
|
"acc_stderr": 0.02968010556502904, |
|
"acc_norm": 0.450354609929078, |
|
"acc_norm_stderr": 0.02968010556502904 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.45535714285714285, |
|
"acc_stderr": 0.04726835553719099, |
|
"acc_norm": 0.45535714285714285, |
|
"acc_norm_stderr": 0.04726835553719099 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.5231481481481481, |
|
"acc_stderr": 0.034063153607115065, |
|
"acc_norm": 0.5231481481481481, |
|
"acc_norm_stderr": 0.034063153607115065 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.45027932960893857, |
|
"acc_stderr": 0.016639615236845817, |
|
"acc_norm": 0.45027932960893857, |
|
"acc_norm_stderr": 0.016639615236845817 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.53, |
|
"acc_stderr": 0.050161355804659205, |
|
"acc_norm": 0.53, |
|
"acc_norm_stderr": 0.050161355804659205 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.7, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.7, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.5735294117647058, |
|
"acc_stderr": 0.030042615832714867, |
|
"acc_norm": 0.5735294117647058, |
|
"acc_norm_stderr": 0.030042615832714867 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.6857142857142857, |
|
"acc_stderr": 0.029719329422417454, |
|
"acc_norm": 0.6857142857142857, |
|
"acc_norm_stderr": 0.029719329422417454 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.759493670886076, |
|
"acc_stderr": 0.027820781981149678, |
|
"acc_norm": 0.759493670886076, |
|
"acc_norm_stderr": 0.027820781981149678 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.46284224250325945, |
|
"acc_stderr": 0.01273492357953206, |
|
"acc_norm": 0.46284224250325945, |
|
"acc_norm_stderr": 0.01273492357953206 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.7107843137254902, |
|
"acc_stderr": 0.031822318676475544, |
|
"acc_norm": 0.7107843137254902, |
|
"acc_norm_stderr": 0.031822318676475544 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.6727272727272727, |
|
"acc_stderr": 0.03663974994391242, |
|
"acc_norm": 0.6727272727272727, |
|
"acc_norm_stderr": 0.03663974994391242 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.6829865361077111, |
|
"mc1_stderr": 0.01628920337440338, |
|
"mc2": 0.7885502219855742, |
|
"mc2_stderr": 0.013364293397046381 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5525383707201889, |
|
"acc_stderr": 0.01709519030150058, |
|
"acc_norm": 0.5702479338842975, |
|
"acc_norm_stderr": 0.01701984753597221 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "JY623/KoSOLRA-10.7B-merge-v2.3", |
|
"model_sha": "7fca6e88fa262cf88251537d1deff1aab6b37a73", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |