results / 4yo1 /llama3-eng-ko-8b-sl4 /result_2024-07-10 12:28:24.json
choco9966's picture
Add results for 2024-07-10 12:28:24
95d8245 verified
raw
history blame
17.9 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.310580204778157,
"acc_stderr": 0.013522292098053054,
"acc_norm": 0.363481228668942,
"acc_norm_stderr": 0.01405620731906828
},
"harness|ko_hellaswag|10": {
"acc": 0.3595897231627166,
"acc_stderr": 0.004788994060654275,
"acc_norm": 0.460565624377614,
"acc_norm_stderr": 0.004974238284524824
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.47953216374269003,
"acc_stderr": 0.038316105328219316,
"acc_norm": 0.47953216374269003,
"acc_norm_stderr": 0.038316105328219316
},
"harness|ko_mmlu_management|5": {
"acc": 0.5436893203883495,
"acc_stderr": 0.049318019942204146,
"acc_norm": 0.5436893203883495,
"acc_norm_stderr": 0.049318019942204146
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.4112388250319285,
"acc_stderr": 0.017595971908056576,
"acc_norm": 0.4112388250319285,
"acc_norm_stderr": 0.017595971908056576
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.34074074074074073,
"acc_stderr": 0.04094376269996794,
"acc_norm": 0.34074074074074073,
"acc_norm_stderr": 0.04094376269996794
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.22,
"acc_stderr": 0.041633319989322695,
"acc_norm": 0.22,
"acc_norm_stderr": 0.041633319989322695
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.4,
"acc_stderr": 0.03202563076101736,
"acc_norm": 0.4,
"acc_norm_stderr": 0.03202563076101736
},
"harness|ko_mmlu_virology|5": {
"acc": 0.3795180722891566,
"acc_stderr": 0.037777988227480165,
"acc_norm": 0.3795180722891566,
"acc_norm_stderr": 0.037777988227480165
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.47266881028938906,
"acc_stderr": 0.02835563356832818,
"acc_norm": 0.47266881028938906,
"acc_norm_stderr": 0.02835563356832818
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.4304932735426009,
"acc_stderr": 0.033231973029429394,
"acc_norm": 0.4304932735426009,
"acc_norm_stderr": 0.033231973029429394
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.4732824427480916,
"acc_stderr": 0.04379024936553893,
"acc_norm": 0.4732824427480916,
"acc_norm_stderr": 0.04379024936553893
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.4,
"acc_stderr": 0.049236596391733084,
"acc_norm": 0.4,
"acc_norm_stderr": 0.049236596391733084
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.45454545454545453,
"acc_stderr": 0.03547601494006937,
"acc_norm": 0.45454545454545453,
"acc_norm_stderr": 0.03547601494006937
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.4896551724137931,
"acc_stderr": 0.041657747757287644,
"acc_norm": 0.4896551724137931,
"acc_norm_stderr": 0.041657747757287644
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.39215686274509803,
"acc_stderr": 0.04858083574266347,
"acc_norm": 0.39215686274509803,
"acc_norm_stderr": 0.04858083574266347
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.47058823529411764,
"acc_stderr": 0.03242225027115007,
"acc_norm": 0.47058823529411764,
"acc_norm_stderr": 0.03242225027115007
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.4666666666666667,
"acc_stderr": 0.025294608023986483,
"acc_norm": 0.4666666666666667,
"acc_norm_stderr": 0.025294608023986483
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.56,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.56,
"acc_norm_stderr": 0.04988876515698589
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.36,
"acc_stderr": 0.04824181513244218,
"acc_norm": 0.36,
"acc_norm_stderr": 0.04824181513244218
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.49074074074074076,
"acc_stderr": 0.04832853553437055,
"acc_norm": 0.49074074074074076,
"acc_norm_stderr": 0.04832853553437055
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.41379310344827586,
"acc_stderr": 0.03465304488406795,
"acc_norm": 0.41379310344827586,
"acc_norm_stderr": 0.03465304488406795
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.45806451612903226,
"acc_stderr": 0.02834378725054064,
"acc_norm": 0.45806451612903226,
"acc_norm_stderr": 0.02834378725054064
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.6153846153846154,
"acc_stderr": 0.03187195347942466,
"acc_norm": 0.6153846153846154,
"acc_norm_stderr": 0.03187195347942466
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4867924528301887,
"acc_stderr": 0.03076213487450049,
"acc_norm": 0.4867924528301887,
"acc_norm_stderr": 0.03076213487450049
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.42727272727272725,
"acc_stderr": 0.04738198703545483,
"acc_norm": 0.42727272727272725,
"acc_norm_stderr": 0.04738198703545483
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.3296296296296296,
"acc_stderr": 0.028661201116524593,
"acc_norm": 0.3296296296296296,
"acc_norm_stderr": 0.028661201116524593
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.3576158940397351,
"acc_stderr": 0.03913453431177258,
"acc_norm": 0.3576158940397351,
"acc_norm_stderr": 0.03913453431177258
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.5522388059701493,
"acc_stderr": 0.03516184772952168,
"acc_norm": 0.5522388059701493,
"acc_norm_stderr": 0.03516184772952168
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.4682080924855491,
"acc_stderr": 0.03804749744364764,
"acc_norm": 0.4682080924855491,
"acc_norm_stderr": 0.03804749744364764
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.36243386243386244,
"acc_stderr": 0.024757473902752045,
"acc_norm": 0.36243386243386244,
"acc_norm_stderr": 0.024757473902752045
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.3819444444444444,
"acc_stderr": 0.040629907841466674,
"acc_norm": 0.3819444444444444,
"acc_norm_stderr": 0.040629907841466674
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.35,
"acc_stderr": 0.047937248544110196,
"acc_norm": 0.35,
"acc_norm_stderr": 0.047937248544110196
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.59,
"acc_stderr": 0.04943110704237101,
"acc_norm": 0.59,
"acc_norm_stderr": 0.04943110704237101
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.43352601156069365,
"acc_stderr": 0.026680134761679217,
"acc_norm": 0.43352601156069365,
"acc_norm_stderr": 0.026680134761679217
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.39263803680981596,
"acc_stderr": 0.03836740907831028,
"acc_norm": 0.39263803680981596,
"acc_norm_stderr": 0.03836740907831028
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.48148148148148145,
"acc_stderr": 0.027801656212323667,
"acc_norm": 0.48148148148148145,
"acc_norm_stderr": 0.027801656212323667
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.32,
"acc_stderr": 0.046882617226215034,
"acc_norm": 0.32,
"acc_norm_stderr": 0.046882617226215034
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.44559585492227977,
"acc_stderr": 0.03587014986075659,
"acc_norm": 0.44559585492227977,
"acc_norm_stderr": 0.03587014986075659
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.30701754385964913,
"acc_stderr": 0.0433913832257986,
"acc_norm": 0.30701754385964913,
"acc_norm_stderr": 0.0433913832257986
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.5192660550458715,
"acc_stderr": 0.02142140298254889,
"acc_norm": 0.5192660550458715,
"acc_norm_stderr": 0.02142140298254889
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.3888888888888889,
"acc_stderr": 0.04360314860077459,
"acc_norm": 0.3888888888888889,
"acc_norm_stderr": 0.04360314860077459
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.5196078431372549,
"acc_stderr": 0.028607893699576066,
"acc_norm": 0.5196078431372549,
"acc_norm_stderr": 0.028607893699576066
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.56,
"acc_stderr": 0.0498887651569859,
"acc_norm": 0.56,
"acc_norm_stderr": 0.0498887651569859
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6115702479338843,
"acc_stderr": 0.044492703500683815,
"acc_norm": 0.6115702479338843,
"acc_norm_stderr": 0.044492703500683815
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.48026315789473684,
"acc_stderr": 0.040657710025626036,
"acc_norm": 0.48026315789473684,
"acc_norm_stderr": 0.040657710025626036
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.3937908496732026,
"acc_stderr": 0.019766211991073063,
"acc_norm": 0.3937908496732026,
"acc_norm_stderr": 0.019766211991073063
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.3723404255319149,
"acc_stderr": 0.02883892147125145,
"acc_norm": 0.3723404255319149,
"acc_norm_stderr": 0.02883892147125145
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.3482142857142857,
"acc_stderr": 0.04521829902833586,
"acc_norm": 0.3482142857142857,
"acc_norm_stderr": 0.04521829902833586
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.39351851851851855,
"acc_stderr": 0.03331747876370312,
"acc_norm": 0.39351851851851855,
"acc_norm_stderr": 0.03331747876370312
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2770949720670391,
"acc_stderr": 0.014968772435812143,
"acc_norm": 0.2770949720670391,
"acc_norm_stderr": 0.014968772435812143
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001975
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.55,
"acc_stderr": 0.04999999999999999,
"acc_norm": 0.55,
"acc_norm_stderr": 0.04999999999999999
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.33088235294117646,
"acc_stderr": 0.028582709753898445,
"acc_norm": 0.33088235294117646,
"acc_norm_stderr": 0.028582709753898445
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.5469387755102041,
"acc_stderr": 0.031867859300041275,
"acc_norm": 0.5469387755102041,
"acc_norm_stderr": 0.031867859300041275
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.5485232067510548,
"acc_stderr": 0.032393600173974704,
"acc_norm": 0.5485232067510548,
"acc_norm_stderr": 0.032393600173974704
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3005215123859192,
"acc_stderr": 0.011709918883039117,
"acc_norm": 0.3005215123859192,
"acc_norm_stderr": 0.011709918883039117
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.46568627450980393,
"acc_stderr": 0.03501038327635897,
"acc_norm": 0.46568627450980393,
"acc_norm_stderr": 0.03501038327635897
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.5393939393939394,
"acc_stderr": 0.03892207016552012,
"acc_norm": 0.5393939393939394,
"acc_norm_stderr": 0.03892207016552012
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.2827417380660955,
"mc1_stderr": 0.01576477083677731,
"mc2": 0.4661586592281064,
"mc2_stderr": 0.016330593604368164
},
"harness|ko_commongen_v2|2": {
"acc": 0.40613931523022434,
"acc_stderr": 0.016884749503191392,
"acc_norm": 0.43919716646989376,
"acc_norm_stderr": 0.017062775744780705
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "4yo1/llama3-eng-ko-8b-sl4",
"model_sha": "25a488e868ecf3d10fef14a534257043b895df7e",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}