results / ITT-AF /ITT-Yi-Ko-6B-v6.0 /result_2024-03-07 03:47:37.json
open-ko-llm-bot's picture
Add results for 2024-03-07 03:47:37
3f1a266 verified
raw
history blame
No virus
17.8 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.378839590443686,
"acc_stderr": 0.014175915490000328,
"acc_norm": 0.4351535836177474,
"acc_norm_stderr": 0.014487986197186045
},
"harness|ko_hellaswag|10": {
"acc": 0.4155546703843856,
"acc_stderr": 0.004918102168717933,
"acc_norm": 0.5569607647878908,
"acc_norm_stderr": 0.004957296691391572
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.5029239766081871,
"acc_stderr": 0.03834759370936839,
"acc_norm": 0.5029239766081871,
"acc_norm_stderr": 0.03834759370936839
},
"harness|ko_mmlu_management|5": {
"acc": 0.5631067961165048,
"acc_stderr": 0.049111471073657764,
"acc_norm": 0.5631067961165048,
"acc_norm_stderr": 0.049111471073657764
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.565772669220945,
"acc_stderr": 0.01772458938967779,
"acc_norm": 0.565772669220945,
"acc_norm_stderr": 0.01772458938967779
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.3851851851851852,
"acc_stderr": 0.042039210401562783,
"acc_norm": 0.3851851851851852,
"acc_norm_stderr": 0.042039210401562783
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.37,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.37,
"acc_norm_stderr": 0.048523658709391
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.425531914893617,
"acc_stderr": 0.03232146916224469,
"acc_norm": 0.425531914893617,
"acc_norm_stderr": 0.03232146916224469
},
"harness|ko_mmlu_virology|5": {
"acc": 0.3855421686746988,
"acc_stderr": 0.03789134424611548,
"acc_norm": 0.3855421686746988,
"acc_norm_stderr": 0.03789134424611548
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.49517684887459806,
"acc_stderr": 0.02839677044411129,
"acc_norm": 0.49517684887459806,
"acc_norm_stderr": 0.02839677044411129
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.4618834080717489,
"acc_stderr": 0.033460150119732274,
"acc_norm": 0.4618834080717489,
"acc_norm_stderr": 0.033460150119732274
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.5114503816793893,
"acc_stderr": 0.043841400240780176,
"acc_norm": 0.5114503816793893,
"acc_norm_stderr": 0.043841400240780176
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.45,
"acc_stderr": 0.05,
"acc_norm": 0.45,
"acc_norm_stderr": 0.05
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.6464646464646465,
"acc_stderr": 0.03406086723547155,
"acc_norm": 0.6464646464646465,
"acc_norm_stderr": 0.03406086723547155
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.5379310344827586,
"acc_stderr": 0.04154659671707548,
"acc_norm": 0.5379310344827586,
"acc_norm_stderr": 0.04154659671707548
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.19607843137254902,
"acc_stderr": 0.039505818611799616,
"acc_norm": 0.19607843137254902,
"acc_norm_stderr": 0.039505818611799616
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.5084033613445378,
"acc_stderr": 0.03247390276569669,
"acc_norm": 0.5084033613445378,
"acc_norm_stderr": 0.03247390276569669
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.4641025641025641,
"acc_stderr": 0.025285585990017834,
"acc_norm": 0.4641025641025641,
"acc_norm_stderr": 0.025285585990017834
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.62,
"acc_stderr": 0.048783173121456316,
"acc_norm": 0.62,
"acc_norm_stderr": 0.048783173121456316
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.32,
"acc_stderr": 0.04688261722621504,
"acc_norm": 0.32,
"acc_norm_stderr": 0.04688261722621504
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.48148148148148145,
"acc_stderr": 0.04830366024635331,
"acc_norm": 0.48148148148148145,
"acc_norm_stderr": 0.04830366024635331
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.43349753694581283,
"acc_stderr": 0.03486731727419872,
"acc_norm": 0.43349753694581283,
"acc_norm_stderr": 0.03486731727419872
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.4806451612903226,
"acc_stderr": 0.0284226874043121,
"acc_norm": 0.4806451612903226,
"acc_norm_stderr": 0.0284226874043121
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.6965811965811965,
"acc_stderr": 0.030118210106942656,
"acc_norm": 0.6965811965811965,
"acc_norm_stderr": 0.030118210106942656
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4679245283018868,
"acc_stderr": 0.03070948699255655,
"acc_norm": 0.4679245283018868,
"acc_norm_stderr": 0.03070948699255655
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.509090909090909,
"acc_stderr": 0.04788339768702861,
"acc_norm": 0.509090909090909,
"acc_norm_stderr": 0.04788339768702861
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.2962962962962963,
"acc_stderr": 0.027840811495871927,
"acc_norm": 0.2962962962962963,
"acc_norm_stderr": 0.027840811495871927
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.32450331125827814,
"acc_stderr": 0.03822746937658752,
"acc_norm": 0.32450331125827814,
"acc_norm_stderr": 0.03822746937658752
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.5920398009950248,
"acc_stderr": 0.03475116365194092,
"acc_norm": 0.5920398009950248,
"acc_norm_stderr": 0.03475116365194092
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.41040462427745666,
"acc_stderr": 0.037507570448955384,
"acc_norm": 0.41040462427745666,
"acc_norm_stderr": 0.037507570448955384
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.32275132275132273,
"acc_stderr": 0.024078943243597016,
"acc_norm": 0.32275132275132273,
"acc_norm_stderr": 0.024078943243597016
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.4652777777777778,
"acc_stderr": 0.041711158581816184,
"acc_norm": 0.4652777777777778,
"acc_norm_stderr": 0.041711158581816184
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.26,
"acc_stderr": 0.044084400227680794,
"acc_norm": 0.26,
"acc_norm_stderr": 0.044084400227680794
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.61,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.61,
"acc_norm_stderr": 0.04902071300001975
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.476878612716763,
"acc_stderr": 0.026890297881303118,
"acc_norm": 0.476878612716763,
"acc_norm_stderr": 0.026890297881303118
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.4723926380368098,
"acc_stderr": 0.03922378290610988,
"acc_norm": 0.4723926380368098,
"acc_norm_stderr": 0.03922378290610988
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.49382716049382713,
"acc_stderr": 0.027818623962583302,
"acc_norm": 0.49382716049382713,
"acc_norm_stderr": 0.027818623962583302
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.5751295336787565,
"acc_stderr": 0.0356747133521254,
"acc_norm": 0.5751295336787565,
"acc_norm_stderr": 0.0356747133521254
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.34210526315789475,
"acc_stderr": 0.04462917535336937,
"acc_norm": 0.34210526315789475,
"acc_norm_stderr": 0.04462917535336937
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.6165137614678899,
"acc_stderr": 0.02084715664191598,
"acc_norm": 0.6165137614678899,
"acc_norm_stderr": 0.02084715664191598
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.29365079365079366,
"acc_stderr": 0.04073524322147124,
"acc_norm": 0.29365079365079366,
"acc_norm_stderr": 0.04073524322147124
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.4803921568627451,
"acc_stderr": 0.028607893699576063,
"acc_norm": 0.4803921568627451,
"acc_norm_stderr": 0.028607893699576063
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.56,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.56,
"acc_norm_stderr": 0.04988876515698589
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6033057851239669,
"acc_stderr": 0.044658697805310094,
"acc_norm": 0.6033057851239669,
"acc_norm_stderr": 0.044658697805310094
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.46710526315789475,
"acc_stderr": 0.040601270352363966,
"acc_norm": 0.46710526315789475,
"acc_norm_stderr": 0.040601270352363966
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.4297385620915033,
"acc_stderr": 0.020027122784928554,
"acc_norm": 0.4297385620915033,
"acc_norm_stderr": 0.020027122784928554
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.3333333333333333,
"acc_stderr": 0.02812163604063988,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.02812163604063988
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.3125,
"acc_stderr": 0.043994650575715215,
"acc_norm": 0.3125,
"acc_norm_stderr": 0.043994650575715215
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.36574074074074076,
"acc_stderr": 0.03284738857647206,
"acc_norm": 0.36574074074074076,
"acc_norm_stderr": 0.03284738857647206
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2245810055865922,
"acc_stderr": 0.013956803666544641,
"acc_norm": 0.2245810055865922,
"acc_norm_stderr": 0.013956803666544641
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695235,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695235
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.5,
"acc_stderr": 0.050251890762960605,
"acc_norm": 0.5,
"acc_norm_stderr": 0.050251890762960605
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.40441176470588236,
"acc_stderr": 0.02981263070156974,
"acc_norm": 0.40441176470588236,
"acc_norm_stderr": 0.02981263070156974
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.49795918367346936,
"acc_stderr": 0.0320089533497105,
"acc_norm": 0.49795918367346936,
"acc_norm_stderr": 0.0320089533497105
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.5949367088607594,
"acc_stderr": 0.03195514741370672,
"acc_norm": 0.5949367088607594,
"acc_norm_stderr": 0.03195514741370672
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3389830508474576,
"acc_stderr": 0.01208994185758447,
"acc_norm": 0.3389830508474576,
"acc_norm_stderr": 0.01208994185758447
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.47058823529411764,
"acc_stderr": 0.03503235296367992,
"acc_norm": 0.47058823529411764,
"acc_norm_stderr": 0.03503235296367992
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.6,
"acc_stderr": 0.03825460278380026,
"acc_norm": 0.6,
"acc_norm_stderr": 0.03825460278380026
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.401468788249694,
"mc1_stderr": 0.01716027390169365,
"mc2": 0.5377364375024334,
"mc2_stderr": 0.015575357735285155
},
"harness|ko_commongen_v2|2": {
"acc": 0.5714285714285714,
"acc_stderr": 0.01701403811929749,
"acc_norm": 0.577331759149941,
"acc_norm_stderr": 0.016983506079577607
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "ITT-AF/ITT-Yi-Ko-6B-v6.0",
"model_sha": "24e9e905be8917263a9ea2a4e6ff193b5635800c",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}