results / 42MARU /polyglot-ko-12.8b-instruct /result_2023-09-27 21:10:18.json
open-ko-llm-bot's picture
Add results for 2023-09-27 21:10:18
ad48028
raw
history blame
17.9 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.3046075085324232,
"acc_stderr": 0.013449522109932492,
"acc_norm": 0.363481228668942,
"acc_norm_stderr": 0.014056207319068282
},
"harness|ko_hellaswag|10": {
"acc": 0.3981278629755029,
"acc_stderr": 0.0048851164655502755,
"acc_norm": 0.5159330810595499,
"acc_norm_stderr": 0.004987247325495624
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.22807017543859648,
"acc_stderr": 0.03218093795602357,
"acc_norm": 0.22807017543859648,
"acc_norm_stderr": 0.03218093795602357
},
"harness|ko_mmlu_management|5": {
"acc": 0.24271844660194175,
"acc_stderr": 0.04245022486384495,
"acc_norm": 0.24271844660194175,
"acc_norm_stderr": 0.04245022486384495
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.23754789272030652,
"acc_stderr": 0.015218733046150193,
"acc_norm": 0.23754789272030652,
"acc_norm_stderr": 0.015218733046150193
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.28888888888888886,
"acc_stderr": 0.039154506304142495,
"acc_norm": 0.28888888888888886,
"acc_norm_stderr": 0.039154506304142495
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.27,
"acc_stderr": 0.0446196043338474,
"acc_norm": 0.27,
"acc_norm_stderr": 0.0446196043338474
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.2297872340425532,
"acc_stderr": 0.02750175294441242,
"acc_norm": 0.2297872340425532,
"acc_norm_stderr": 0.02750175294441242
},
"harness|ko_mmlu_virology|5": {
"acc": 0.2469879518072289,
"acc_stderr": 0.03357351982064536,
"acc_norm": 0.2469879518072289,
"acc_norm_stderr": 0.03357351982064536
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.33440514469453375,
"acc_stderr": 0.026795422327893944,
"acc_norm": 0.33440514469453375,
"acc_norm_stderr": 0.026795422327893944
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.1031390134529148,
"acc_stderr": 0.020412564289839272,
"acc_norm": 0.1031390134529148,
"acc_norm_stderr": 0.020412564289839272
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.2748091603053435,
"acc_stderr": 0.039153454088478354,
"acc_norm": 0.2748091603053435,
"acc_norm_stderr": 0.039153454088478354
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.37,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.37,
"acc_norm_stderr": 0.048523658709391
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.40404040404040403,
"acc_stderr": 0.03496130972056128,
"acc_norm": 0.40404040404040403,
"acc_norm_stderr": 0.03496130972056128
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.3310344827586207,
"acc_stderr": 0.03921545312467122,
"acc_norm": 0.3310344827586207,
"acc_norm_stderr": 0.03921545312467122
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.2647058823529412,
"acc_stderr": 0.04389869956808777,
"acc_norm": 0.2647058823529412,
"acc_norm_stderr": 0.04389869956808777
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.18907563025210083,
"acc_stderr": 0.02543511943810536,
"acc_norm": 0.18907563025210083,
"acc_norm_stderr": 0.02543511943810536
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.21794871794871795,
"acc_stderr": 0.020932445774463175,
"acc_norm": 0.21794871794871795,
"acc_norm_stderr": 0.020932445774463175
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.18,
"acc_stderr": 0.03861229196653694,
"acc_norm": 0.18,
"acc_norm_stderr": 0.03861229196653694
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.16,
"acc_stderr": 0.03684529491774708,
"acc_norm": 0.16,
"acc_norm_stderr": 0.03684529491774708
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.26851851851851855,
"acc_stderr": 0.04284467968052191,
"acc_norm": 0.26851851851851855,
"acc_norm_stderr": 0.04284467968052191
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.1921182266009852,
"acc_stderr": 0.027719315709614778,
"acc_norm": 0.1921182266009852,
"acc_norm_stderr": 0.027719315709614778
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.25483870967741934,
"acc_stderr": 0.024790118459332204,
"acc_norm": 0.25483870967741934,
"acc_norm_stderr": 0.024790118459332204
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.2905982905982906,
"acc_stderr": 0.029745048572674057,
"acc_norm": 0.2905982905982906,
"acc_norm_stderr": 0.029745048572674057
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.29056603773584905,
"acc_stderr": 0.027943219989337156,
"acc_norm": 0.29056603773584905,
"acc_norm_stderr": 0.027943219989337156
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.2636363636363636,
"acc_stderr": 0.04220224692971987,
"acc_norm": 0.2636363636363636,
"acc_norm_stderr": 0.04220224692971987
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.25925925925925924,
"acc_stderr": 0.026719240783712166,
"acc_norm": 0.25925925925925924,
"acc_norm_stderr": 0.026719240783712166
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.2913907284768212,
"acc_stderr": 0.03710185726119995,
"acc_norm": 0.2913907284768212,
"acc_norm_stderr": 0.03710185726119995
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.19402985074626866,
"acc_stderr": 0.027962677604768914,
"acc_norm": 0.19402985074626866,
"acc_norm_stderr": 0.027962677604768914
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.2254335260115607,
"acc_stderr": 0.03186209851641143,
"acc_norm": 0.2254335260115607,
"acc_norm_stderr": 0.03186209851641143
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.24867724867724866,
"acc_stderr": 0.02226181769240018,
"acc_norm": 0.24867724867724866,
"acc_norm_stderr": 0.02226181769240018
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.3125,
"acc_stderr": 0.038760854559127644,
"acc_norm": 0.3125,
"acc_norm_stderr": 0.038760854559127644
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.21,
"acc_stderr": 0.04093601807403326,
"acc_norm": 0.21,
"acc_norm_stderr": 0.04093601807403326
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.27,
"acc_stderr": 0.044619604333847394,
"acc_norm": 0.27,
"acc_norm_stderr": 0.044619604333847394
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.2774566473988439,
"acc_stderr": 0.024105712607754307,
"acc_norm": 0.2774566473988439,
"acc_norm_stderr": 0.024105712607754307
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.3067484662576687,
"acc_stderr": 0.036230899157241474,
"acc_norm": 0.3067484662576687,
"acc_norm_stderr": 0.036230899157241474
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.24074074074074073,
"acc_stderr": 0.023788583551658537,
"acc_norm": 0.24074074074074073,
"acc_norm_stderr": 0.023788583551658537
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.26,
"acc_stderr": 0.04408440022768077,
"acc_norm": 0.26,
"acc_norm_stderr": 0.04408440022768077
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.22797927461139897,
"acc_stderr": 0.030276909945178253,
"acc_norm": 0.22797927461139897,
"acc_norm_stderr": 0.030276909945178253
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.30701754385964913,
"acc_stderr": 0.043391383225798594,
"acc_norm": 0.30701754385964913,
"acc_norm_stderr": 0.043391383225798594
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.3376146788990826,
"acc_stderr": 0.02027526598663891,
"acc_norm": 0.3376146788990826,
"acc_norm_stderr": 0.02027526598663891
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.23809523809523808,
"acc_stderr": 0.038095238095238126,
"acc_norm": 0.23809523809523808,
"acc_norm_stderr": 0.038095238095238126
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.27450980392156865,
"acc_stderr": 0.025553169991826524,
"acc_norm": 0.27450980392156865,
"acc_norm_stderr": 0.025553169991826524
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.32,
"acc_stderr": 0.04688261722621505,
"acc_norm": 0.32,
"acc_norm_stderr": 0.04688261722621505
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.24793388429752067,
"acc_stderr": 0.03941897526516302,
"acc_norm": 0.24793388429752067,
"acc_norm_stderr": 0.03941897526516302
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.23684210526315788,
"acc_stderr": 0.03459777606810537,
"acc_norm": 0.23684210526315788,
"acc_norm_stderr": 0.03459777606810537
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.25980392156862747,
"acc_stderr": 0.01774089950917779,
"acc_norm": 0.25980392156862747,
"acc_norm_stderr": 0.01774089950917779
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.22340425531914893,
"acc_stderr": 0.024847921358063962,
"acc_norm": 0.22340425531914893,
"acc_norm_stderr": 0.024847921358063962
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.25892857142857145,
"acc_stderr": 0.04157751539865629,
"acc_norm": 0.25892857142857145,
"acc_norm_stderr": 0.04157751539865629
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.24537037037037038,
"acc_stderr": 0.029346665094372937,
"acc_norm": 0.24537037037037038,
"acc_norm_stderr": 0.029346665094372937
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.27262569832402234,
"acc_stderr": 0.014893391735249608,
"acc_norm": 0.27262569832402234,
"acc_norm_stderr": 0.014893391735249608
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.32,
"acc_stderr": 0.046882617226215034,
"acc_norm": 0.32,
"acc_norm_stderr": 0.046882617226215034
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.3786764705882353,
"acc_stderr": 0.02946513363977613,
"acc_norm": 0.3786764705882353,
"acc_norm_stderr": 0.02946513363977613
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.35918367346938773,
"acc_stderr": 0.03071356045510849,
"acc_norm": 0.35918367346938773,
"acc_norm_stderr": 0.03071356045510849
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.24472573839662448,
"acc_stderr": 0.027985699387036416,
"acc_norm": 0.24472573839662448,
"acc_norm_stderr": 0.027985699387036416
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.2588005215123859,
"acc_stderr": 0.011186109046564608,
"acc_norm": 0.2588005215123859,
"acc_norm_stderr": 0.011186109046564608
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.28431372549019607,
"acc_stderr": 0.031660096793998116,
"acc_norm": 0.28431372549019607,
"acc_norm_stderr": 0.031660096793998116
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.2,
"acc_stderr": 0.03123475237772118,
"acc_norm": 0.2,
"acc_norm_stderr": 0.03123475237772118
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.27906976744186046,
"mc1_stderr": 0.015702107090627887,
"mc2": 0.4515720476496737,
"mc2_stderr": 0.015493161984611252
},
"harness|ko_commongen_v2|2": {
"acc": 0.5528169014084507,
"acc_stderr": 0.017043883876215398,
"acc_norm": 0.5997652582159625,
"acc_norm_stderr": 0.016795125938543782
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "42MARU/polyglot-ko-12.8b-instruct",
"model_sha": "a8354bcedc167e8e1f7dac8a347bf4b61d9c9bf0",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}