|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3293515358361775, |
|
"acc_stderr": 0.013734057652635474, |
|
"acc_norm": 0.386518771331058, |
|
"acc_norm_stderr": 0.014230084761910474 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3835889265086636, |
|
"acc_stderr": 0.00485265887677539, |
|
"acc_norm": 0.5022903804023103, |
|
"acc_norm_stderr": 0.004989729059957435 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.391812865497076, |
|
"acc_stderr": 0.03743979825926401, |
|
"acc_norm": 0.391812865497076, |
|
"acc_norm_stderr": 0.03743979825926401 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.2815533980582524, |
|
"acc_stderr": 0.04453254836326466, |
|
"acc_norm": 0.2815533980582524, |
|
"acc_norm_stderr": 0.04453254836326466 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.367816091954023, |
|
"acc_stderr": 0.01724382889184626, |
|
"acc_norm": 0.367816091954023, |
|
"acc_norm_stderr": 0.01724382889184626 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.34074074074074073, |
|
"acc_stderr": 0.04094376269996795, |
|
"acc_norm": 0.34074074074074073, |
|
"acc_norm_stderr": 0.04094376269996795 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.2851063829787234, |
|
"acc_stderr": 0.02951319662553935, |
|
"acc_norm": 0.2851063829787234, |
|
"acc_norm_stderr": 0.02951319662553935 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.3132530120481928, |
|
"acc_stderr": 0.03610805018031024, |
|
"acc_norm": 0.3132530120481928, |
|
"acc_norm_stderr": 0.03610805018031024 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.3858520900321543, |
|
"acc_stderr": 0.027648149599751464, |
|
"acc_norm": 0.3858520900321543, |
|
"acc_norm_stderr": 0.027648149599751464 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.43946188340807174, |
|
"acc_stderr": 0.03331092511038179, |
|
"acc_norm": 0.43946188340807174, |
|
"acc_norm_stderr": 0.03331092511038179 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.4351145038167939, |
|
"acc_stderr": 0.04348208051644858, |
|
"acc_norm": 0.4351145038167939, |
|
"acc_norm_stderr": 0.04348208051644858 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.31313131313131315, |
|
"acc_stderr": 0.03304205087813653, |
|
"acc_norm": 0.31313131313131315, |
|
"acc_norm_stderr": 0.03304205087813653 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.3586206896551724, |
|
"acc_stderr": 0.039966295748767186, |
|
"acc_norm": 0.3586206896551724, |
|
"acc_norm_stderr": 0.039966295748767186 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.20588235294117646, |
|
"acc_stderr": 0.04023382273617747, |
|
"acc_norm": 0.20588235294117646, |
|
"acc_norm_stderr": 0.04023382273617747 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.3697478991596639, |
|
"acc_stderr": 0.031357095996135904, |
|
"acc_norm": 0.3697478991596639, |
|
"acc_norm_stderr": 0.031357095996135904 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.2794871794871795, |
|
"acc_stderr": 0.022752388839776823, |
|
"acc_norm": 0.2794871794871795, |
|
"acc_norm_stderr": 0.022752388839776823 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.049236596391733084, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.049236596391733084 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.04020151261036846, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.04020151261036846 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.37962962962962965, |
|
"acc_stderr": 0.04691521224077742, |
|
"acc_norm": 0.37962962962962965, |
|
"acc_norm_stderr": 0.04691521224077742 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.22660098522167488, |
|
"acc_stderr": 0.02945486383529298, |
|
"acc_norm": 0.22660098522167488, |
|
"acc_norm_stderr": 0.02945486383529298 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.3225806451612903, |
|
"acc_stderr": 0.026593084516572267, |
|
"acc_norm": 0.3225806451612903, |
|
"acc_norm_stderr": 0.026593084516572267 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.03275608910402091, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.03275608910402091 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.3471698113207547, |
|
"acc_stderr": 0.029300101705549652, |
|
"acc_norm": 0.3471698113207547, |
|
"acc_norm_stderr": 0.029300101705549652 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.36363636363636365, |
|
"acc_stderr": 0.04607582090719976, |
|
"acc_norm": 0.36363636363636365, |
|
"acc_norm_stderr": 0.04607582090719976 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.23703703703703705, |
|
"acc_stderr": 0.025928876132766118, |
|
"acc_norm": 0.23703703703703705, |
|
"acc_norm_stderr": 0.025928876132766118 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2119205298013245, |
|
"acc_stderr": 0.03336767086567977, |
|
"acc_norm": 0.2119205298013245, |
|
"acc_norm_stderr": 0.03336767086567977 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.39800995024875624, |
|
"acc_stderr": 0.034611994290400135, |
|
"acc_norm": 0.39800995024875624, |
|
"acc_norm_stderr": 0.034611994290400135 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.2774566473988439, |
|
"acc_stderr": 0.03414014007044036, |
|
"acc_norm": 0.2774566473988439, |
|
"acc_norm_stderr": 0.03414014007044036 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.24867724867724866, |
|
"acc_stderr": 0.022261817692400175, |
|
"acc_norm": 0.24867724867724866, |
|
"acc_norm_stderr": 0.022261817692400175 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.2916666666666667, |
|
"acc_stderr": 0.03800968060554859, |
|
"acc_norm": 0.2916666666666667, |
|
"acc_norm_stderr": 0.03800968060554859 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.21, |
|
"acc_stderr": 0.040936018074033256, |
|
"acc_norm": 0.21, |
|
"acc_norm_stderr": 0.040936018074033256 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.44, |
|
"acc_stderr": 0.04988876515698589, |
|
"acc_norm": 0.44, |
|
"acc_norm_stderr": 0.04988876515698589 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.38439306358381503, |
|
"acc_stderr": 0.026189666966272035, |
|
"acc_norm": 0.38439306358381503, |
|
"acc_norm_stderr": 0.026189666966272035 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.3496932515337423, |
|
"acc_stderr": 0.03746668325470022, |
|
"acc_norm": 0.3496932515337423, |
|
"acc_norm_stderr": 0.03746668325470022 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.3611111111111111, |
|
"acc_stderr": 0.026725868809100793, |
|
"acc_norm": 0.3611111111111111, |
|
"acc_norm_stderr": 0.026725868809100793 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.04512608598542128, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.04512608598542128 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.35751295336787564, |
|
"acc_stderr": 0.03458816042181005, |
|
"acc_norm": 0.35751295336787564, |
|
"acc_norm_stderr": 0.03458816042181005 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.23684210526315788, |
|
"acc_stderr": 0.03999423879281336, |
|
"acc_norm": 0.23684210526315788, |
|
"acc_norm_stderr": 0.03999423879281336 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.3394495412844037, |
|
"acc_stderr": 0.02030210934266235, |
|
"acc_norm": 0.3394495412844037, |
|
"acc_norm_stderr": 0.02030210934266235 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.2619047619047619, |
|
"acc_stderr": 0.039325376803928704, |
|
"acc_norm": 0.2619047619047619, |
|
"acc_norm_stderr": 0.039325376803928704 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.3660130718954248, |
|
"acc_stderr": 0.027582811415159607, |
|
"acc_norm": 0.3660130718954248, |
|
"acc_norm_stderr": 0.027582811415159607 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.04878317312145632, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.04878317312145632 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.49586776859504134, |
|
"acc_stderr": 0.04564198767432754, |
|
"acc_norm": 0.49586776859504134, |
|
"acc_norm_stderr": 0.04564198767432754 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.24342105263157895, |
|
"acc_stderr": 0.034923496688842384, |
|
"acc_norm": 0.24342105263157895, |
|
"acc_norm_stderr": 0.034923496688842384 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.3366013071895425, |
|
"acc_stderr": 0.019117213911495165, |
|
"acc_norm": 0.3366013071895425, |
|
"acc_norm_stderr": 0.019117213911495165 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.29432624113475175, |
|
"acc_stderr": 0.027187127011503796, |
|
"acc_norm": 0.29432624113475175, |
|
"acc_norm_stderr": 0.027187127011503796 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.29464285714285715, |
|
"acc_stderr": 0.04327040932578728, |
|
"acc_norm": 0.29464285714285715, |
|
"acc_norm_stderr": 0.04327040932578728 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.2824074074074074, |
|
"acc_stderr": 0.030701372111510927, |
|
"acc_norm": 0.2824074074074074, |
|
"acc_norm_stderr": 0.030701372111510927 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.25139664804469275, |
|
"acc_stderr": 0.014508979453553977, |
|
"acc_norm": 0.25139664804469275, |
|
"acc_norm_stderr": 0.014508979453553977 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621505, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621505 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.3786764705882353, |
|
"acc_stderr": 0.029465133639776125, |
|
"acc_norm": 0.3786764705882353, |
|
"acc_norm_stderr": 0.029465133639776125 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.3510204081632653, |
|
"acc_stderr": 0.03055531675557364, |
|
"acc_norm": 0.3510204081632653, |
|
"acc_norm_stderr": 0.03055531675557364 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.4767932489451477, |
|
"acc_stderr": 0.032512152011410174, |
|
"acc_norm": 0.4767932489451477, |
|
"acc_norm_stderr": 0.032512152011410174 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3089960886571056, |
|
"acc_stderr": 0.01180172977723925, |
|
"acc_norm": 0.3089960886571056, |
|
"acc_norm_stderr": 0.01180172977723925 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.31862745098039214, |
|
"acc_stderr": 0.032702871814820816, |
|
"acc_norm": 0.31862745098039214, |
|
"acc_norm_stderr": 0.032702871814820816 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.3212121212121212, |
|
"acc_stderr": 0.0364620496325381, |
|
"acc_norm": 0.3212121212121212, |
|
"acc_norm_stderr": 0.0364620496325381 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.25091799265605874, |
|
"mc1_stderr": 0.01517698502770769, |
|
"mc2": 0.38056097212603235, |
|
"mc2_stderr": 0.014936929596682727 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5704225352112676, |
|
"acc_stderr": 0.01696892392010678, |
|
"acc_norm": 0.6549295774647887, |
|
"acc_norm_stderr": 0.016296201644718785 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "42MARU/llama-2-ko-7b-instruction-v3", |
|
"model_sha": "c0fea9cb31d4ae90aa2ed048f774a9000341b538", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |