|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3660409556313993, |
|
"acc_stderr": 0.01407722310847014, |
|
"acc_norm": 0.41467576791808874, |
|
"acc_norm_stderr": 0.014397070564409175 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3996215893248357, |
|
"acc_stderr": 0.004888194985997395, |
|
"acc_norm": 0.5103565026887075, |
|
"acc_norm_stderr": 0.004988710917169331 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.543859649122807, |
|
"acc_stderr": 0.038200425866029654, |
|
"acc_norm": 0.543859649122807, |
|
"acc_norm_stderr": 0.038200425866029654 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5631067961165048, |
|
"acc_stderr": 0.04911147107365777, |
|
"acc_norm": 0.5631067961165048, |
|
"acc_norm_stderr": 0.04911147107365777 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5134099616858238, |
|
"acc_stderr": 0.017873531736510354, |
|
"acc_norm": 0.5134099616858238, |
|
"acc_norm_stderr": 0.017873531736510354 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4074074074074074, |
|
"acc_stderr": 0.04244633238353229, |
|
"acc_norm": 0.4074074074074074, |
|
"acc_norm_stderr": 0.04244633238353229 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.04512608598542128, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.04512608598542128 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.42127659574468085, |
|
"acc_stderr": 0.03227834510146268, |
|
"acc_norm": 0.42127659574468085, |
|
"acc_norm_stderr": 0.03227834510146268 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.42168674698795183, |
|
"acc_stderr": 0.03844453181770917, |
|
"acc_norm": 0.42168674698795183, |
|
"acc_norm_stderr": 0.03844453181770917 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.4887459807073955, |
|
"acc_stderr": 0.02839089739686352, |
|
"acc_norm": 0.4887459807073955, |
|
"acc_norm_stderr": 0.02839089739686352 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.452914798206278, |
|
"acc_stderr": 0.033408675019233246, |
|
"acc_norm": 0.452914798206278, |
|
"acc_norm_stderr": 0.033408675019233246 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.4122137404580153, |
|
"acc_stderr": 0.04317171194870255, |
|
"acc_norm": 0.4122137404580153, |
|
"acc_norm_stderr": 0.04317171194870255 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5808080808080808, |
|
"acc_stderr": 0.03515520728670417, |
|
"acc_norm": 0.5808080808080808, |
|
"acc_norm_stderr": 0.03515520728670417 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.4689655172413793, |
|
"acc_stderr": 0.04158632762097828, |
|
"acc_norm": 0.4689655172413793, |
|
"acc_norm_stderr": 0.04158632762097828 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.2549019607843137, |
|
"acc_stderr": 0.043364327079931785, |
|
"acc_norm": 0.2549019607843137, |
|
"acc_norm_stderr": 0.043364327079931785 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.48739495798319327, |
|
"acc_stderr": 0.03246816765752174, |
|
"acc_norm": 0.48739495798319327, |
|
"acc_norm_stderr": 0.03246816765752174 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.45897435897435895, |
|
"acc_stderr": 0.025265525491284295, |
|
"acc_norm": 0.45897435897435895, |
|
"acc_norm_stderr": 0.025265525491284295 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.56, |
|
"acc_stderr": 0.04988876515698589, |
|
"acc_norm": 0.56, |
|
"acc_norm_stderr": 0.04988876515698589 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252604, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252604 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5092592592592593, |
|
"acc_stderr": 0.04832853553437055, |
|
"acc_norm": 0.5092592592592593, |
|
"acc_norm_stderr": 0.04832853553437055 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.4187192118226601, |
|
"acc_stderr": 0.03471192860518468, |
|
"acc_norm": 0.4187192118226601, |
|
"acc_norm_stderr": 0.03471192860518468 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.47419354838709676, |
|
"acc_stderr": 0.028406095057653315, |
|
"acc_norm": 0.47419354838709676, |
|
"acc_norm_stderr": 0.028406095057653315 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7094017094017094, |
|
"acc_stderr": 0.029745048572674078, |
|
"acc_norm": 0.7094017094017094, |
|
"acc_norm_stderr": 0.029745048572674078 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.4528301886792453, |
|
"acc_stderr": 0.03063562795796182, |
|
"acc_norm": 0.4528301886792453, |
|
"acc_norm_stderr": 0.03063562795796182 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.4818181818181818, |
|
"acc_stderr": 0.04785964010794916, |
|
"acc_norm": 0.4818181818181818, |
|
"acc_norm_stderr": 0.04785964010794916 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.34074074074074073, |
|
"acc_stderr": 0.02889774874113115, |
|
"acc_norm": 0.34074074074074073, |
|
"acc_norm_stderr": 0.02889774874113115 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2582781456953642, |
|
"acc_stderr": 0.035737053147634576, |
|
"acc_norm": 0.2582781456953642, |
|
"acc_norm_stderr": 0.035737053147634576 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.5970149253731343, |
|
"acc_stderr": 0.034683432951111266, |
|
"acc_norm": 0.5970149253731343, |
|
"acc_norm_stderr": 0.034683432951111266 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.3988439306358382, |
|
"acc_stderr": 0.03733626655383509, |
|
"acc_norm": 0.3988439306358382, |
|
"acc_norm_stderr": 0.03733626655383509 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.0242785680243077, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.0242785680243077 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.375, |
|
"acc_stderr": 0.04048439222695598, |
|
"acc_norm": 0.375, |
|
"acc_norm_stderr": 0.04048439222695598 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.047258156262526045, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.63, |
|
"acc_stderr": 0.048523658709391, |
|
"acc_norm": 0.63, |
|
"acc_norm_stderr": 0.048523658709391 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5086705202312138, |
|
"acc_stderr": 0.026915047355369804, |
|
"acc_norm": 0.5086705202312138, |
|
"acc_norm_stderr": 0.026915047355369804 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.4662576687116564, |
|
"acc_stderr": 0.03919415545048409, |
|
"acc_norm": 0.4662576687116564, |
|
"acc_norm_stderr": 0.03919415545048409 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.46296296296296297, |
|
"acc_stderr": 0.02774431344337654, |
|
"acc_norm": 0.46296296296296297, |
|
"acc_norm_stderr": 0.02774431344337654 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.5492227979274611, |
|
"acc_stderr": 0.035909109522355244, |
|
"acc_norm": 0.5492227979274611, |
|
"acc_norm_stderr": 0.035909109522355244 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2894736842105263, |
|
"acc_stderr": 0.04266339443159394, |
|
"acc_norm": 0.2894736842105263, |
|
"acc_norm_stderr": 0.04266339443159394 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.4972477064220184, |
|
"acc_stderr": 0.02143699835976532, |
|
"acc_norm": 0.4972477064220184, |
|
"acc_norm_stderr": 0.02143699835976532 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.3412698412698413, |
|
"acc_stderr": 0.04240799327574924, |
|
"acc_norm": 0.3412698412698413, |
|
"acc_norm_stderr": 0.04240799327574924 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5065359477124183, |
|
"acc_stderr": 0.028627470550556047, |
|
"acc_norm": 0.5065359477124183, |
|
"acc_norm_stderr": 0.028627470550556047 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.47, |
|
"acc_stderr": 0.05016135580465919, |
|
"acc_norm": 0.47, |
|
"acc_norm_stderr": 0.05016135580465919 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.7107438016528925, |
|
"acc_stderr": 0.04139112727635464, |
|
"acc_norm": 0.7107438016528925, |
|
"acc_norm_stderr": 0.04139112727635464 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.40789473684210525, |
|
"acc_stderr": 0.03999309712777471, |
|
"acc_norm": 0.40789473684210525, |
|
"acc_norm_stderr": 0.03999309712777471 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.40522875816993464, |
|
"acc_stderr": 0.019861155193829153, |
|
"acc_norm": 0.40522875816993464, |
|
"acc_norm_stderr": 0.019861155193829153 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3404255319148936, |
|
"acc_stderr": 0.028267657482650137, |
|
"acc_norm": 0.3404255319148936, |
|
"acc_norm_stderr": 0.028267657482650137 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.36607142857142855, |
|
"acc_stderr": 0.0457237235873743, |
|
"acc_norm": 0.36607142857142855, |
|
"acc_norm_stderr": 0.0457237235873743 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.03214952147802749, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.03214952147802749 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.29608938547486036, |
|
"acc_stderr": 0.015268677317602298, |
|
"acc_norm": 0.29608938547486036, |
|
"acc_norm_stderr": 0.015268677317602298 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.04824181513244218, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.04824181513244218 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.58, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.58, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.41911764705882354, |
|
"acc_stderr": 0.029972807170464626, |
|
"acc_norm": 0.41911764705882354, |
|
"acc_norm_stderr": 0.029972807170464626 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.5265306122448979, |
|
"acc_stderr": 0.03196412734523272, |
|
"acc_norm": 0.5265306122448979, |
|
"acc_norm_stderr": 0.03196412734523272 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.6160337552742616, |
|
"acc_stderr": 0.031658678064106674, |
|
"acc_norm": 0.6160337552742616, |
|
"acc_norm_stderr": 0.031658678064106674 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3435462842242503, |
|
"acc_stderr": 0.012128961174190156, |
|
"acc_norm": 0.3435462842242503, |
|
"acc_norm_stderr": 0.012128961174190156 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.47549019607843135, |
|
"acc_stderr": 0.035050931943487976, |
|
"acc_norm": 0.47549019607843135, |
|
"acc_norm_stderr": 0.035050931943487976 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.46060606060606063, |
|
"acc_stderr": 0.03892207016552013, |
|
"acc_norm": 0.46060606060606063, |
|
"acc_norm_stderr": 0.03892207016552013 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2717258261933905, |
|
"mc1_stderr": 0.015572840452875835, |
|
"mc2": 0.4265196372341518, |
|
"mc2_stderr": 0.015607268833983732 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.3707201889020071, |
|
"acc_stderr": 0.0166058012892126, |
|
"acc_norm": 0.3955135773317591, |
|
"acc_norm_stderr": 0.01681081590220604 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "BM-K/mistral-7b-it-v1.7.2", |
|
"model_sha": "dbbe8fc186f5371e9ca608b9650e387b20de7ae3", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |