results / maywell /Synatra_TbST11B_EP01 /result_2023-10-18 07:35:40.json
open-ko-llm-bot's picture
Add results for 2023-10-18 07:35:40
5647dbc
raw
history blame
No virus
17.8 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.34812286689419797,
"acc_stderr": 0.013921008595179344,
"acc_norm": 0.40784982935153585,
"acc_norm_stderr": 0.014361097288449691
},
"harness|ko_hellaswag|10": {
"acc": 0.36875124477195775,
"acc_stderr": 0.0048148030984368085,
"acc_norm": 0.4722166899024099,
"acc_norm_stderr": 0.004982072108448084
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.45614035087719296,
"acc_stderr": 0.03820042586602966,
"acc_norm": 0.45614035087719296,
"acc_norm_stderr": 0.03820042586602966
},
"harness|ko_mmlu_management|5": {
"acc": 0.5145631067961165,
"acc_stderr": 0.04948637324026637,
"acc_norm": 0.5145631067961165,
"acc_norm_stderr": 0.04948637324026637
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.4840357598978289,
"acc_stderr": 0.01787084750608173,
"acc_norm": 0.4840357598978289,
"acc_norm_stderr": 0.01787084750608173
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.35555555555555557,
"acc_stderr": 0.04135176749720386,
"acc_norm": 0.35555555555555557,
"acc_norm_stderr": 0.04135176749720386
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.37,
"acc_stderr": 0.048523658709391,
"acc_norm": 0.37,
"acc_norm_stderr": 0.048523658709391
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.3574468085106383,
"acc_stderr": 0.03132941789476425,
"acc_norm": 0.3574468085106383,
"acc_norm_stderr": 0.03132941789476425
},
"harness|ko_mmlu_virology|5": {
"acc": 0.3795180722891566,
"acc_stderr": 0.03777798822748016,
"acc_norm": 0.3795180722891566,
"acc_norm_stderr": 0.03777798822748016
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.5016077170418006,
"acc_stderr": 0.02839794490780661,
"acc_norm": 0.5016077170418006,
"acc_norm_stderr": 0.02839794490780661
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.38565022421524664,
"acc_stderr": 0.03266842214289201,
"acc_norm": 0.38565022421524664,
"acc_norm_stderr": 0.03266842214289201
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.4351145038167939,
"acc_stderr": 0.04348208051644858,
"acc_norm": 0.4351145038167939,
"acc_norm_stderr": 0.04348208051644858
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.5808080808080808,
"acc_stderr": 0.035155207286704175,
"acc_norm": 0.5808080808080808,
"acc_norm_stderr": 0.035155207286704175
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.4827586206896552,
"acc_stderr": 0.04164188720169377,
"acc_norm": 0.4827586206896552,
"acc_norm_stderr": 0.04164188720169377
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.29411764705882354,
"acc_stderr": 0.045338381959297736,
"acc_norm": 0.29411764705882354,
"acc_norm_stderr": 0.045338381959297736
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.47478991596638653,
"acc_stderr": 0.0324371805513741,
"acc_norm": 0.47478991596638653,
"acc_norm_stderr": 0.0324371805513741
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.4461538461538462,
"acc_stderr": 0.025203571773028333,
"acc_norm": 0.4461538461538462,
"acc_norm_stderr": 0.025203571773028333
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.5,
"acc_stderr": 0.050251890762960605,
"acc_norm": 0.5,
"acc_norm_stderr": 0.050251890762960605
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.32,
"acc_stderr": 0.04688261722621504,
"acc_norm": 0.32,
"acc_norm_stderr": 0.04688261722621504
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.49074074074074076,
"acc_stderr": 0.04832853553437055,
"acc_norm": 0.49074074074074076,
"acc_norm_stderr": 0.04832853553437055
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.37438423645320196,
"acc_stderr": 0.03405155380561952,
"acc_norm": 0.37438423645320196,
"acc_norm_stderr": 0.03405155380561952
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.432258064516129,
"acc_stderr": 0.02818173972001941,
"acc_norm": 0.432258064516129,
"acc_norm_stderr": 0.02818173972001941
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.6666666666666666,
"acc_stderr": 0.030882736974138653,
"acc_norm": 0.6666666666666666,
"acc_norm_stderr": 0.030882736974138653
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.5094339622641509,
"acc_stderr": 0.030767394707808107,
"acc_norm": 0.5094339622641509,
"acc_norm_stderr": 0.030767394707808107
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.4090909090909091,
"acc_stderr": 0.04709306978661896,
"acc_norm": 0.4090909090909091,
"acc_norm_stderr": 0.04709306978661896
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.3074074074074074,
"acc_stderr": 0.028133252578815642,
"acc_norm": 0.3074074074074074,
"acc_norm_stderr": 0.028133252578815642
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.3443708609271523,
"acc_stderr": 0.03879687024073327,
"acc_norm": 0.3443708609271523,
"acc_norm_stderr": 0.03879687024073327
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.6069651741293532,
"acc_stderr": 0.0345368246603156,
"acc_norm": 0.6069651741293532,
"acc_norm_stderr": 0.0345368246603156
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.4393063583815029,
"acc_stderr": 0.037842719328874674,
"acc_norm": 0.4393063583815029,
"acc_norm_stderr": 0.037842719328874674
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.3544973544973545,
"acc_stderr": 0.024636830602841997,
"acc_norm": 0.3544973544973545,
"acc_norm_stderr": 0.024636830602841997
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.3472222222222222,
"acc_stderr": 0.039812405437178615,
"acc_norm": 0.3472222222222222,
"acc_norm_stderr": 0.039812405437178615
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.35,
"acc_stderr": 0.0479372485441102,
"acc_norm": 0.35,
"acc_norm_stderr": 0.0479372485441102
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.55,
"acc_stderr": 0.05,
"acc_norm": 0.55,
"acc_norm_stderr": 0.05
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.4884393063583815,
"acc_stderr": 0.026911898686377913,
"acc_norm": 0.4884393063583815,
"acc_norm_stderr": 0.026911898686377913
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.44171779141104295,
"acc_stderr": 0.039015918258361836,
"acc_norm": 0.44171779141104295,
"acc_norm_stderr": 0.039015918258361836
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.43209876543209874,
"acc_stderr": 0.02756301097160668,
"acc_norm": 0.43209876543209874,
"acc_norm_stderr": 0.02756301097160668
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.38,
"acc_stderr": 0.048783173121456316,
"acc_norm": 0.38,
"acc_norm_stderr": 0.048783173121456316
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.47668393782383417,
"acc_stderr": 0.03604513672442205,
"acc_norm": 0.47668393782383417,
"acc_norm_stderr": 0.03604513672442205
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.2982456140350877,
"acc_stderr": 0.04303684033537317,
"acc_norm": 0.2982456140350877,
"acc_norm_stderr": 0.04303684033537317
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.48807339449541287,
"acc_stderr": 0.021431223617362223,
"acc_norm": 0.48807339449541287,
"acc_norm_stderr": 0.021431223617362223
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.29365079365079366,
"acc_stderr": 0.04073524322147124,
"acc_norm": 0.29365079365079366,
"acc_norm_stderr": 0.04073524322147124
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.4673202614379085,
"acc_stderr": 0.02856869975222588,
"acc_norm": 0.4673202614379085,
"acc_norm_stderr": 0.02856869975222588
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.628099173553719,
"acc_stderr": 0.04412015806624504,
"acc_norm": 0.628099173553719,
"acc_norm_stderr": 0.04412015806624504
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.4342105263157895,
"acc_stderr": 0.040335656678483184,
"acc_norm": 0.4342105263157895,
"acc_norm_stderr": 0.040335656678483184
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.3660130718954248,
"acc_stderr": 0.019488025745529658,
"acc_norm": 0.3660130718954248,
"acc_norm_stderr": 0.019488025745529658
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.32269503546099293,
"acc_stderr": 0.027889139300534792,
"acc_norm": 0.32269503546099293,
"acc_norm_stderr": 0.027889139300534792
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.3125,
"acc_stderr": 0.043994650575715215,
"acc_norm": 0.3125,
"acc_norm_stderr": 0.043994650575715215
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.3611111111111111,
"acc_stderr": 0.032757734861009996,
"acc_norm": 0.3611111111111111,
"acc_norm_stderr": 0.032757734861009996
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.23016759776536314,
"acc_stderr": 0.014078339253425807,
"acc_norm": 0.23016759776536314,
"acc_norm_stderr": 0.014078339253425807
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.43,
"acc_stderr": 0.049756985195624284,
"acc_norm": 0.43,
"acc_norm_stderr": 0.049756985195624284
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.45,
"acc_stderr": 0.05,
"acc_norm": 0.45,
"acc_norm_stderr": 0.05
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.3860294117647059,
"acc_stderr": 0.029573269134411124,
"acc_norm": 0.3860294117647059,
"acc_norm_stderr": 0.029573269134411124
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.4775510204081633,
"acc_stderr": 0.03197694118713673,
"acc_norm": 0.4775510204081633,
"acc_norm_stderr": 0.03197694118713673
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.5864978902953587,
"acc_stderr": 0.03205649904851858,
"acc_norm": 0.5864978902953587,
"acc_norm_stderr": 0.03205649904851858
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3155149934810952,
"acc_stderr": 0.011869184843058643,
"acc_norm": 0.3155149934810952,
"acc_norm_stderr": 0.011869184843058643
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.49019607843137253,
"acc_stderr": 0.03508637358630572,
"acc_norm": 0.49019607843137253,
"acc_norm_stderr": 0.03508637358630572
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.5151515151515151,
"acc_stderr": 0.03902551007374448,
"acc_norm": 0.5151515151515151,
"acc_norm_stderr": 0.03902551007374448
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.2766217870257038,
"mc1_stderr": 0.015659605755326912,
"mc2": 0.4484601943910918,
"mc2_stderr": 0.015458891626438749
},
"harness|ko_commongen_v2|2": {
"acc": 0.3482880755608028,
"acc_stderr": 0.01637992673914804,
"acc_norm": 0.39787485242030696,
"acc_norm_stderr": 0.016827959054733388
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "maywell/Synatra_TbST11B_EP01",
"model_sha": "76b40393481aa567733eff5107dd4b6944862fdd",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}