|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3677474402730375, |
|
"acc_stderr": 0.014090995618168489, |
|
"acc_norm": 0.4283276450511945, |
|
"acc_norm_stderr": 0.014460496367599026 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3997211710814579, |
|
"acc_stderr": 0.004888398535520493, |
|
"acc_norm": 0.5373431587333201, |
|
"acc_norm_stderr": 0.004975845335086618 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.5087719298245614, |
|
"acc_stderr": 0.038342347441649924, |
|
"acc_norm": 0.5087719298245614, |
|
"acc_norm_stderr": 0.038342347441649924 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5825242718446602, |
|
"acc_stderr": 0.048828405482122375, |
|
"acc_norm": 0.5825242718446602, |
|
"acc_norm_stderr": 0.048828405482122375 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5734355044699873, |
|
"acc_stderr": 0.01768606697567565, |
|
"acc_norm": 0.5734355044699873, |
|
"acc_norm_stderr": 0.01768606697567565 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4888888888888889, |
|
"acc_stderr": 0.04318275491977978, |
|
"acc_norm": 0.4888888888888889, |
|
"acc_norm_stderr": 0.04318275491977978 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.39574468085106385, |
|
"acc_stderr": 0.03196758697835363, |
|
"acc_norm": 0.39574468085106385, |
|
"acc_norm_stderr": 0.03196758697835363 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.3674698795180723, |
|
"acc_stderr": 0.03753267402120574, |
|
"acc_norm": 0.3674698795180723, |
|
"acc_norm_stderr": 0.03753267402120574 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.4919614147909968, |
|
"acc_stderr": 0.02839442137098453, |
|
"acc_norm": 0.4919614147909968, |
|
"acc_norm_stderr": 0.02839442137098453 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5112107623318386, |
|
"acc_stderr": 0.033549366530984746, |
|
"acc_norm": 0.5112107623318386, |
|
"acc_norm_stderr": 0.033549366530984746 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.5038167938931297, |
|
"acc_stderr": 0.04385162325601553, |
|
"acc_norm": 0.5038167938931297, |
|
"acc_norm_stderr": 0.04385162325601553 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.39, |
|
"acc_stderr": 0.04902071300001975, |
|
"acc_norm": 0.39, |
|
"acc_norm_stderr": 0.04902071300001975 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.601010101010101, |
|
"acc_stderr": 0.0348890161685273, |
|
"acc_norm": 0.601010101010101, |
|
"acc_norm_stderr": 0.0348890161685273 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.4827586206896552, |
|
"acc_stderr": 0.041641887201693754, |
|
"acc_norm": 0.4827586206896552, |
|
"acc_norm_stderr": 0.041641887201693754 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.22549019607843138, |
|
"acc_stderr": 0.04158307533083286, |
|
"acc_norm": 0.22549019607843138, |
|
"acc_norm_stderr": 0.04158307533083286 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.4789915966386555, |
|
"acc_stderr": 0.03244980849990029, |
|
"acc_norm": 0.4789915966386555, |
|
"acc_norm_stderr": 0.03244980849990029 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4282051282051282, |
|
"acc_stderr": 0.025088301454694834, |
|
"acc_norm": 0.4282051282051282, |
|
"acc_norm_stderr": 0.025088301454694834 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.63, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.63, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.34, |
|
"acc_stderr": 0.04760952285695235, |
|
"acc_norm": 0.34, |
|
"acc_norm_stderr": 0.04760952285695235 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.48148148148148145, |
|
"acc_stderr": 0.04830366024635331, |
|
"acc_norm": 0.48148148148148145, |
|
"acc_norm_stderr": 0.04830366024635331 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.37438423645320196, |
|
"acc_stderr": 0.03405155380561952, |
|
"acc_norm": 0.37438423645320196, |
|
"acc_norm_stderr": 0.03405155380561952 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.43870967741935485, |
|
"acc_stderr": 0.028229497320317216, |
|
"acc_norm": 0.43870967741935485, |
|
"acc_norm_stderr": 0.028229497320317216 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7136752136752137, |
|
"acc_stderr": 0.029614323690456648, |
|
"acc_norm": 0.7136752136752137, |
|
"acc_norm_stderr": 0.029614323690456648 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.47547169811320755, |
|
"acc_stderr": 0.030735822206205608, |
|
"acc_norm": 0.47547169811320755, |
|
"acc_norm_stderr": 0.030735822206205608 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5454545454545454, |
|
"acc_stderr": 0.04769300568972744, |
|
"acc_norm": 0.5454545454545454, |
|
"acc_norm_stderr": 0.04769300568972744 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.32222222222222224, |
|
"acc_stderr": 0.028493465091028597, |
|
"acc_norm": 0.32222222222222224, |
|
"acc_norm_stderr": 0.028493465091028597 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.304635761589404, |
|
"acc_stderr": 0.037579499229433426, |
|
"acc_norm": 0.304635761589404, |
|
"acc_norm_stderr": 0.037579499229433426 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.5522388059701493, |
|
"acc_stderr": 0.03516184772952167, |
|
"acc_norm": 0.5522388059701493, |
|
"acc_norm_stderr": 0.03516184772952167 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.3583815028901734, |
|
"acc_stderr": 0.036563436533531585, |
|
"acc_norm": 0.3583815028901734, |
|
"acc_norm_stderr": 0.036563436533531585 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.3148148148148148, |
|
"acc_stderr": 0.023919984164047732, |
|
"acc_norm": 0.3148148148148148, |
|
"acc_norm_stderr": 0.023919984164047732 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4097222222222222, |
|
"acc_stderr": 0.04112490974670787, |
|
"acc_norm": 0.4097222222222222, |
|
"acc_norm_stderr": 0.04112490974670787 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.6, |
|
"acc_stderr": 0.04923659639173309, |
|
"acc_norm": 0.6, |
|
"acc_norm_stderr": 0.04923659639173309 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.49710982658959535, |
|
"acc_stderr": 0.026918645383239004, |
|
"acc_norm": 0.49710982658959535, |
|
"acc_norm_stderr": 0.026918645383239004 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.4785276073619632, |
|
"acc_stderr": 0.03924746876751129, |
|
"acc_norm": 0.4785276073619632, |
|
"acc_norm_stderr": 0.03924746876751129 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.45987654320987653, |
|
"acc_stderr": 0.02773102275353928, |
|
"acc_norm": 0.45987654320987653, |
|
"acc_norm_stderr": 0.02773102275353928 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.5699481865284974, |
|
"acc_stderr": 0.03572954333144808, |
|
"acc_norm": 0.5699481865284974, |
|
"acc_norm_stderr": 0.03572954333144808 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.3157894736842105, |
|
"acc_stderr": 0.043727482902780085, |
|
"acc_norm": 0.3157894736842105, |
|
"acc_norm_stderr": 0.043727482902780085 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.5706422018348624, |
|
"acc_stderr": 0.021222286397236504, |
|
"acc_norm": 0.5706422018348624, |
|
"acc_norm_stderr": 0.021222286397236504 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.2698412698412698, |
|
"acc_stderr": 0.03970158273235173, |
|
"acc_norm": 0.2698412698412698, |
|
"acc_norm_stderr": 0.03970158273235173 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.43790849673202614, |
|
"acc_stderr": 0.02840830202033269, |
|
"acc_norm": 0.43790849673202614, |
|
"acc_norm_stderr": 0.02840830202033269 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.56, |
|
"acc_stderr": 0.04988876515698589, |
|
"acc_norm": 0.56, |
|
"acc_norm_stderr": 0.04988876515698589 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6694214876033058, |
|
"acc_stderr": 0.04294340845212094, |
|
"acc_norm": 0.6694214876033058, |
|
"acc_norm_stderr": 0.04294340845212094 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.4473684210526316, |
|
"acc_stderr": 0.04046336883978252, |
|
"acc_norm": 0.4473684210526316, |
|
"acc_norm_stderr": 0.04046336883978252 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.41013071895424835, |
|
"acc_stderr": 0.01989841271763589, |
|
"acc_norm": 0.41013071895424835, |
|
"acc_norm_stderr": 0.01989841271763589 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3475177304964539, |
|
"acc_stderr": 0.02840662780959095, |
|
"acc_norm": 0.3475177304964539, |
|
"acc_norm_stderr": 0.02840662780959095 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.3392857142857143, |
|
"acc_stderr": 0.0449394906861354, |
|
"acc_norm": 0.3392857142857143, |
|
"acc_norm_stderr": 0.0449394906861354 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.27314814814814814, |
|
"acc_stderr": 0.030388051301678116, |
|
"acc_norm": 0.27314814814814814, |
|
"acc_norm_stderr": 0.030388051301678116 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.21564245810055865, |
|
"acc_stderr": 0.013754835975482337, |
|
"acc_norm": 0.21564245810055865, |
|
"acc_norm_stderr": 0.013754835975482337 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.048241815132442176, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.048241815132442176 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.52, |
|
"acc_stderr": 0.050211673156867795, |
|
"acc_norm": 0.52, |
|
"acc_norm_stderr": 0.050211673156867795 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.375, |
|
"acc_stderr": 0.029408372932278746, |
|
"acc_norm": 0.375, |
|
"acc_norm_stderr": 0.029408372932278746 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.3836734693877551, |
|
"acc_stderr": 0.031130880396235926, |
|
"acc_norm": 0.3836734693877551, |
|
"acc_norm_stderr": 0.031130880396235926 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.5949367088607594, |
|
"acc_stderr": 0.03195514741370672, |
|
"acc_norm": 0.5949367088607594, |
|
"acc_norm_stderr": 0.03195514741370672 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3305084745762712, |
|
"acc_stderr": 0.012014142101842977, |
|
"acc_norm": 0.3305084745762712, |
|
"acc_norm_stderr": 0.012014142101842977 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.5147058823529411, |
|
"acc_stderr": 0.035077938347913236, |
|
"acc_norm": 0.5147058823529411, |
|
"acc_norm_stderr": 0.035077938347913236 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.5393939393939394, |
|
"acc_stderr": 0.03892207016552012, |
|
"acc_norm": 0.5393939393939394, |
|
"acc_norm_stderr": 0.03892207016552012 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.26193390452876375, |
|
"mc1_stderr": 0.015392118805015006, |
|
"mc2": 0.41956796147607484, |
|
"mc2_stderr": 0.01500218367833407 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.4085005903187721, |
|
"acc_stderr": 0.01690006287942712, |
|
"acc_norm": 0.43683589138134593, |
|
"acc_norm_stderr": 0.017052633559856076 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "ITT-AF/ITT-Yi-Ko-6B-v2.0", |
|
"model_sha": "89b900e75c8f20c494081b6d498de6ab63f97a19", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |