results / KT-AI /midm-bitext-S-7B-inst-v2 /result_2024-01-15 07:08:32Z.json
open-ko-llm-bot's picture
Add results for 2024-01-15 07:08:32Z
84b7d5d verified
raw history blame
No virus
17.8 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.378839590443686,
"acc_stderr": 0.014175915490000322,
"acc_norm": 0.4513651877133106,
"acc_norm_stderr": 0.014542104569955264
},
"harness|ko_hellaswag|10": {
"acc": 0.433877713602868,
"acc_stderr": 0.004945956744943811,
"acc_norm": 0.5880302728540131,
"acc_norm_stderr": 0.004911837730582209
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.6608187134502924,
"acc_stderr": 0.03631053496488904,
"acc_norm": 0.6608187134502924,
"acc_norm_stderr": 0.03631053496488904
},
"harness|ko_mmlu_management|5": {
"acc": 0.6310679611650486,
"acc_stderr": 0.0477761518115674,
"acc_norm": 0.6310679611650486,
"acc_norm_stderr": 0.0477761518115674
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.6526181353767561,
"acc_stderr": 0.01702667174865574,
"acc_norm": 0.6526181353767561,
"acc_norm_stderr": 0.01702667174865574
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.4148148148148148,
"acc_stderr": 0.04256193767901407,
"acc_norm": 0.4148148148148148,
"acc_norm_stderr": 0.04256193767901407
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.27,
"acc_stderr": 0.04461960433384741,
"acc_norm": 0.27,
"acc_norm_stderr": 0.04461960433384741
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.41702127659574467,
"acc_stderr": 0.03223276266711712,
"acc_norm": 0.41702127659574467,
"acc_norm_stderr": 0.03223276266711712
},
"harness|ko_mmlu_virology|5": {
"acc": 0.40963855421686746,
"acc_stderr": 0.03828401115079022,
"acc_norm": 0.40963855421686746,
"acc_norm_stderr": 0.03828401115079022
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.4790996784565916,
"acc_stderr": 0.028373270961069414,
"acc_norm": 0.4790996784565916,
"acc_norm_stderr": 0.028373270961069414
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.5426008968609866,
"acc_stderr": 0.033435777055830646,
"acc_norm": 0.5426008968609866,
"acc_norm_stderr": 0.033435777055830646
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.5343511450381679,
"acc_stderr": 0.04374928560599738,
"acc_norm": 0.5343511450381679,
"acc_norm_stderr": 0.04374928560599738
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.55,
"acc_stderr": 0.05,
"acc_norm": 0.55,
"acc_norm_stderr": 0.05
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.5656565656565656,
"acc_stderr": 0.03531505879359182,
"acc_norm": 0.5656565656565656,
"acc_norm_stderr": 0.03531505879359182
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.47586206896551725,
"acc_stderr": 0.041618085035015295,
"acc_norm": 0.47586206896551725,
"acc_norm_stderr": 0.041618085035015295
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.21568627450980393,
"acc_stderr": 0.04092563958237655,
"acc_norm": 0.21568627450980393,
"acc_norm_stderr": 0.04092563958237655
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.46218487394957986,
"acc_stderr": 0.032385469487589795,
"acc_norm": 0.46218487394957986,
"acc_norm_stderr": 0.032385469487589795
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.4641025641025641,
"acc_stderr": 0.025285585990017834,
"acc_norm": 0.4641025641025641,
"acc_norm_stderr": 0.025285585990017834
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.57,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.57,
"acc_norm_stderr": 0.049756985195624284
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.26,
"acc_stderr": 0.0440844002276808,
"acc_norm": 0.26,
"acc_norm_stderr": 0.0440844002276808
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.5370370370370371,
"acc_stderr": 0.04820403072760627,
"acc_norm": 0.5370370370370371,
"acc_norm_stderr": 0.04820403072760627
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.33497536945812806,
"acc_stderr": 0.0332085274234831,
"acc_norm": 0.33497536945812806,
"acc_norm_stderr": 0.0332085274234831
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.5387096774193548,
"acc_stderr": 0.028358634859836935,
"acc_norm": 0.5387096774193548,
"acc_norm_stderr": 0.028358634859836935
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.7222222222222222,
"acc_stderr": 0.029343114798094462,
"acc_norm": 0.7222222222222222,
"acc_norm_stderr": 0.029343114798094462
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4867924528301887,
"acc_stderr": 0.030762134874500476,
"acc_norm": 0.4867924528301887,
"acc_norm_stderr": 0.030762134874500476
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.5272727272727272,
"acc_stderr": 0.04782001791380061,
"acc_norm": 0.5272727272727272,
"acc_norm_stderr": 0.04782001791380061
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.2851851851851852,
"acc_stderr": 0.027528599210340492,
"acc_norm": 0.2851851851851852,
"acc_norm_stderr": 0.027528599210340492
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.26490066225165565,
"acc_stderr": 0.03603038545360384,
"acc_norm": 0.26490066225165565,
"acc_norm_stderr": 0.03603038545360384
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.6318407960199005,
"acc_stderr": 0.03410410565495302,
"acc_norm": 0.6318407960199005,
"acc_norm_stderr": 0.03410410565495302
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.47398843930635837,
"acc_stderr": 0.03807301726504513,
"acc_norm": 0.47398843930635837,
"acc_norm_stderr": 0.03807301726504513
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.328042328042328,
"acc_stderr": 0.02418049716437691,
"acc_norm": 0.328042328042328,
"acc_norm_stderr": 0.02418049716437691
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.4097222222222222,
"acc_stderr": 0.04112490974670788,
"acc_norm": 0.4097222222222222,
"acc_norm_stderr": 0.04112490974670788
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695235,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695235
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.7,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.7,
"acc_norm_stderr": 0.046056618647183814
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.4797687861271676,
"acc_stderr": 0.026897049996382868,
"acc_norm": 0.4797687861271676,
"acc_norm_stderr": 0.026897049996382868
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.5276073619631901,
"acc_stderr": 0.0392237829061099,
"acc_norm": 0.5276073619631901,
"acc_norm_stderr": 0.0392237829061099
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.5,
"acc_stderr": 0.02782074420373286,
"acc_norm": 0.5,
"acc_norm_stderr": 0.02782074420373286
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001974,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001974
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.6217616580310881,
"acc_stderr": 0.03499807276193339,
"acc_norm": 0.6217616580310881,
"acc_norm_stderr": 0.03499807276193339
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.2719298245614035,
"acc_stderr": 0.04185774424022057,
"acc_norm": 0.2719298245614035,
"acc_norm_stderr": 0.04185774424022057
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.6238532110091743,
"acc_stderr": 0.02076923196820508,
"acc_norm": 0.6238532110091743,
"acc_norm_stderr": 0.02076923196820508
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.3492063492063492,
"acc_stderr": 0.04263906892795133,
"acc_norm": 0.3492063492063492,
"acc_norm_stderr": 0.04263906892795133
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.4869281045751634,
"acc_stderr": 0.028620130800700246,
"acc_norm": 0.4869281045751634,
"acc_norm_stderr": 0.028620130800700246
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.51,
"acc_stderr": 0.05024183937956912,
"acc_norm": 0.51,
"acc_norm_stderr": 0.05024183937956912
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6528925619834711,
"acc_stderr": 0.043457245702925335,
"acc_norm": 0.6528925619834711,
"acc_norm_stderr": 0.043457245702925335
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.5,
"acc_stderr": 0.04068942293855797,
"acc_norm": 0.5,
"acc_norm_stderr": 0.04068942293855797
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.4166666666666667,
"acc_stderr": 0.019944914136873583,
"acc_norm": 0.4166666666666667,
"acc_norm_stderr": 0.019944914136873583
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.02812163604063988,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.02812163604063988
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.3482142857142857,
"acc_stderr": 0.04521829902833586,
"acc_norm": 0.3482142857142857,
"acc_norm_stderr": 0.04521829902833586
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.2962962962962963,
"acc_stderr": 0.03114144782353604,
"acc_norm": 0.2962962962962963,
"acc_norm_stderr": 0.03114144782353604
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.33519553072625696,
"acc_stderr": 0.01578800719018589,
"acc_norm": 0.33519553072625696,
"acc_norm_stderr": 0.01578800719018589
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.4,
"acc_stderr": 0.049236596391733084,
"acc_norm": 0.4,
"acc_norm_stderr": 0.049236596391733084
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.45,
"acc_stderr": 0.05,
"acc_norm": 0.45,
"acc_norm_stderr": 0.05
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.4411764705882353,
"acc_stderr": 0.030161911930767105,
"acc_norm": 0.4411764705882353,
"acc_norm_stderr": 0.030161911930767105
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.46122448979591835,
"acc_stderr": 0.03191282052669277,
"acc_norm": 0.46122448979591835,
"acc_norm_stderr": 0.03191282052669277
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.6624472573839663,
"acc_stderr": 0.030781549102026223,
"acc_norm": 0.6624472573839663,
"acc_norm_stderr": 0.030781549102026223
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.31681877444589307,
"acc_stderr": 0.011882349954722997,
"acc_norm": 0.31681877444589307,
"acc_norm_stderr": 0.011882349954722997
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.24019607843137256,
"acc_stderr": 0.02998373305591361,
"acc_norm": 0.24019607843137256,
"acc_norm_stderr": 0.02998373305591361
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.2545454545454545,
"acc_stderr": 0.03401506715249039,
"acc_norm": 0.2545454545454545,
"acc_norm_stderr": 0.03401506715249039
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.2839657282741738,
"mc1_stderr": 0.01578537085839672,
"mc2": 0.4359830362647055,
"mc2_stderr": 0.015231569782976262
},
"harness|ko_commongen_v2|2": {
"acc": 0.5785123966942148,
"acc_stderr": 0.016977101932601518,
"acc_norm": 0.5855962219598583,
"acc_norm_stderr": 0.016936583383943615
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "KT-AI/midm-bitext-S-7B-inst-v2",
"model_sha": "a1d976b7852265bdac94f05a9ce34a48add84236",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}