results / BM-K /mistral-7b-it-v1.7.0 /result_2023-11-20 09:22:14.json
choco9966
add backup results
70a679f
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.3447098976109215,
"acc_stderr": 0.01388881628678211,
"acc_norm": 0.386518771331058,
"acc_norm_stderr": 0.014230084761910474
},
"harness|ko_hellaswag|10": {
"acc": 0.3873730332603067,
"acc_stderr": 0.004861544478451848,
"acc_norm": 0.5089623580959968,
"acc_norm_stderr": 0.004988979750014428
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.49122807017543857,
"acc_stderr": 0.038342347441649924,
"acc_norm": 0.49122807017543857,
"acc_norm_stderr": 0.038342347441649924
},
"harness|ko_mmlu_management|5": {
"acc": 0.6019417475728155,
"acc_stderr": 0.04846748253977238,
"acc_norm": 0.6019417475728155,
"acc_norm_stderr": 0.04846748253977238
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.4827586206896552,
"acc_stderr": 0.017869330154003705,
"acc_norm": 0.4827586206896552,
"acc_norm_stderr": 0.017869330154003705
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.362962962962963,
"acc_stderr": 0.041539484047424,
"acc_norm": 0.362962962962963,
"acc_norm_stderr": 0.041539484047424
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.26,
"acc_stderr": 0.0440844002276808,
"acc_norm": 0.26,
"acc_norm_stderr": 0.0440844002276808
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.41702127659574467,
"acc_stderr": 0.03223276266711712,
"acc_norm": 0.41702127659574467,
"acc_norm_stderr": 0.03223276266711712
},
"harness|ko_mmlu_virology|5": {
"acc": 0.39156626506024095,
"acc_stderr": 0.03799857454479636,
"acc_norm": 0.39156626506024095,
"acc_norm_stderr": 0.03799857454479636
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.4662379421221865,
"acc_stderr": 0.02833327710956278,
"acc_norm": 0.4662379421221865,
"acc_norm_stderr": 0.02833327710956278
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.43946188340807174,
"acc_stderr": 0.03331092511038179,
"acc_norm": 0.43946188340807174,
"acc_norm_stderr": 0.03331092511038179
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.42748091603053434,
"acc_stderr": 0.043389203057924,
"acc_norm": 0.42748091603053434,
"acc_norm_stderr": 0.043389203057924
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695235,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695235
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.5353535353535354,
"acc_stderr": 0.03553436368828063,
"acc_norm": 0.5353535353535354,
"acc_norm_stderr": 0.03553436368828063
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.4896551724137931,
"acc_stderr": 0.041657747757287644,
"acc_norm": 0.4896551724137931,
"acc_norm_stderr": 0.041657747757287644
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.29411764705882354,
"acc_stderr": 0.04533838195929778,
"acc_norm": 0.29411764705882354,
"acc_norm_stderr": 0.04533838195929778
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.5546218487394958,
"acc_stderr": 0.0322841062671639,
"acc_norm": 0.5546218487394958,
"acc_norm_stderr": 0.0322841062671639
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.45897435897435895,
"acc_stderr": 0.025265525491284295,
"acc_norm": 0.45897435897435895,
"acc_norm_stderr": 0.025265525491284295
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.56,
"acc_stderr": 0.0498887651569859,
"acc_norm": 0.56,
"acc_norm_stderr": 0.0498887651569859
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695235,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695235
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.49074074074074076,
"acc_stderr": 0.04832853553437055,
"acc_norm": 0.49074074074074076,
"acc_norm_stderr": 0.04832853553437055
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.4088669950738916,
"acc_stderr": 0.0345905881588323,
"acc_norm": 0.4088669950738916,
"acc_norm_stderr": 0.0345905881588323
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.4774193548387097,
"acc_stderr": 0.028414985019707868,
"acc_norm": 0.4774193548387097,
"acc_norm_stderr": 0.028414985019707868
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.7307692307692307,
"acc_stderr": 0.029058588303748842,
"acc_norm": 0.7307692307692307,
"acc_norm_stderr": 0.029058588303748842
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.41509433962264153,
"acc_stderr": 0.030325945789286105,
"acc_norm": 0.41509433962264153,
"acc_norm_stderr": 0.030325945789286105
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.509090909090909,
"acc_stderr": 0.0478833976870286,
"acc_norm": 0.509090909090909,
"acc_norm_stderr": 0.0478833976870286
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.028742040903948482,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.028742040903948482
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.26490066225165565,
"acc_stderr": 0.03603038545360384,
"acc_norm": 0.26490066225165565,
"acc_norm_stderr": 0.03603038545360384
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.6119402985074627,
"acc_stderr": 0.03445789964362749,
"acc_norm": 0.6119402985074627,
"acc_norm_stderr": 0.03445789964362749
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.3930635838150289,
"acc_stderr": 0.03724249595817729,
"acc_norm": 0.3930635838150289,
"acc_norm_stderr": 0.03724249595817729
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.35714285714285715,
"acc_stderr": 0.024677862841332783,
"acc_norm": 0.35714285714285715,
"acc_norm_stderr": 0.024677862841332783
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.3888888888888889,
"acc_stderr": 0.04076663253918567,
"acc_norm": 0.3888888888888889,
"acc_norm_stderr": 0.04076663253918567
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.38,
"acc_stderr": 0.04878317312145632,
"acc_norm": 0.38,
"acc_norm_stderr": 0.04878317312145632
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.61,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.61,
"acc_norm_stderr": 0.04902071300001975
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.47398843930635837,
"acc_stderr": 0.026882643434022885,
"acc_norm": 0.47398843930635837,
"acc_norm_stderr": 0.026882643434022885
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.49693251533742333,
"acc_stderr": 0.03928297078179662,
"acc_norm": 0.49693251533742333,
"acc_norm_stderr": 0.03928297078179662
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.4537037037037037,
"acc_stderr": 0.0277012284685426,
"acc_norm": 0.4537037037037037,
"acc_norm_stderr": 0.0277012284685426
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.5129533678756477,
"acc_stderr": 0.03607228061047749,
"acc_norm": 0.5129533678756477,
"acc_norm_stderr": 0.03607228061047749
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.2719298245614035,
"acc_stderr": 0.041857744240220575,
"acc_norm": 0.2719298245614035,
"acc_norm_stderr": 0.041857744240220575
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.48990825688073397,
"acc_stderr": 0.021432956203453316,
"acc_norm": 0.48990825688073397,
"acc_norm_stderr": 0.021432956203453316
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.4603174603174603,
"acc_stderr": 0.04458029125470973,
"acc_norm": 0.4603174603174603,
"acc_norm_stderr": 0.04458029125470973
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.5032679738562091,
"acc_stderr": 0.02862930519400355,
"acc_norm": 0.5032679738562091,
"acc_norm_stderr": 0.02862930519400355
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.47,
"acc_stderr": 0.05016135580465919,
"acc_norm": 0.47,
"acc_norm_stderr": 0.05016135580465919
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6859504132231405,
"acc_stderr": 0.04236964753041018,
"acc_norm": 0.6859504132231405,
"acc_norm_stderr": 0.04236964753041018
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.42105263157894735,
"acc_stderr": 0.04017901275981748,
"acc_norm": 0.42105263157894735,
"acc_norm_stderr": 0.04017901275981748
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.39869281045751637,
"acc_stderr": 0.01980828131744984,
"acc_norm": 0.39869281045751637,
"acc_norm_stderr": 0.01980828131744984
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.31560283687943264,
"acc_stderr": 0.027724989449509314,
"acc_norm": 0.31560283687943264,
"acc_norm_stderr": 0.027724989449509314
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.3482142857142857,
"acc_stderr": 0.045218299028335865,
"acc_norm": 0.3482142857142857,
"acc_norm_stderr": 0.045218299028335865
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.4398148148148148,
"acc_stderr": 0.03385177976044811,
"acc_norm": 0.4398148148148148,
"acc_norm_stderr": 0.03385177976044811
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2748603351955307,
"acc_stderr": 0.014931316703220517,
"acc_norm": 0.2748603351955307,
"acc_norm_stderr": 0.014931316703220517
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.61,
"acc_stderr": 0.04902071300001974,
"acc_norm": 0.61,
"acc_norm_stderr": 0.04902071300001974
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.3492647058823529,
"acc_stderr": 0.028959755196824862,
"acc_norm": 0.3492647058823529,
"acc_norm_stderr": 0.028959755196824862
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.5346938775510204,
"acc_stderr": 0.03193207024425314,
"acc_norm": 0.5346938775510204,
"acc_norm_stderr": 0.03193207024425314
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.5738396624472574,
"acc_stderr": 0.03219035703131774,
"acc_norm": 0.5738396624472574,
"acc_norm_stderr": 0.03219035703131774
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.33116036505867014,
"acc_stderr": 0.012020128195985774,
"acc_norm": 0.33116036505867014,
"acc_norm_stderr": 0.012020128195985774
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.4411764705882353,
"acc_stderr": 0.034849415144292316,
"acc_norm": 0.4411764705882353,
"acc_norm_stderr": 0.034849415144292316
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.4666666666666667,
"acc_stderr": 0.03895658065271846,
"acc_norm": 0.4666666666666667,
"acc_norm_stderr": 0.03895658065271846
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.3023255813953488,
"mc1_stderr": 0.016077509266133033,
"mc2": 0.4714881280704747,
"mc2_stderr": 0.01554412183162796
},
"harness|ko_commongen_v2|2": {
"acc": 0.5230224321133412,
"acc_stderr": 0.017172121546727634,
"acc_norm": 0.577331759149941,
"acc_norm_stderr": 0.016983506079577604
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "BM-K/mistral-7b-it-v1.7.0",
"model_sha": "f62174ae285bf46cc453305f1e0b76899a8bcf82",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}