results / Alphacode-AI /AlphaMist7B-slr-v2 /result_2024-04-08 15:32:51.json
open-ko-llm-bot's picture
Add results for 2024-04-08 15:32:51
b2978df verified
raw
history blame
No virus
17.8 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.47440273037542663,
"acc_stderr": 0.014592230885298959,
"acc_norm": 0.515358361774744,
"acc_norm_stderr": 0.014604496129394916
},
"harness|ko_hellaswag|10": {
"acc": 0.5727942640908186,
"acc_stderr": 0.00493661642892264,
"acc_norm": 0.69398526190002,
"acc_norm_stderr": 0.004598940722374079
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.49707602339181284,
"acc_stderr": 0.03834759370936839,
"acc_norm": 0.49707602339181284,
"acc_norm_stderr": 0.03834759370936839
},
"harness|ko_mmlu_management|5": {
"acc": 0.5728155339805825,
"acc_stderr": 0.04897957737781168,
"acc_norm": 0.5728155339805825,
"acc_norm_stderr": 0.04897957737781168
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.5236270753512133,
"acc_stderr": 0.017859989765176453,
"acc_norm": 0.5236270753512133,
"acc_norm_stderr": 0.017859989765176453
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.35555555555555557,
"acc_stderr": 0.04135176749720386,
"acc_norm": 0.35555555555555557,
"acc_norm_stderr": 0.04135176749720386
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.24,
"acc_stderr": 0.04292346959909284,
"acc_norm": 0.24,
"acc_norm_stderr": 0.04292346959909284
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.34893617021276596,
"acc_stderr": 0.031158522131357766,
"acc_norm": 0.34893617021276596,
"acc_norm_stderr": 0.031158522131357766
},
"harness|ko_mmlu_virology|5": {
"acc": 0.4457831325301205,
"acc_stderr": 0.03869543323472101,
"acc_norm": 0.4457831325301205,
"acc_norm_stderr": 0.03869543323472101
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.5112540192926045,
"acc_stderr": 0.028390897396863537,
"acc_norm": 0.5112540192926045,
"acc_norm_stderr": 0.028390897396863537
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.45739910313901344,
"acc_stderr": 0.03343577705583064,
"acc_norm": 0.45739910313901344,
"acc_norm_stderr": 0.03343577705583064
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.4580152671755725,
"acc_stderr": 0.04369802690578756,
"acc_norm": 0.4580152671755725,
"acc_norm_stderr": 0.04369802690578756
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.54,
"acc_stderr": 0.05009082659620333,
"acc_norm": 0.54,
"acc_norm_stderr": 0.05009082659620333
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.6161616161616161,
"acc_stderr": 0.03464881675016338,
"acc_norm": 0.6161616161616161,
"acc_norm_stderr": 0.03464881675016338
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.45517241379310347,
"acc_stderr": 0.04149886942192117,
"acc_norm": 0.45517241379310347,
"acc_norm_stderr": 0.04149886942192117
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.3137254901960784,
"acc_stderr": 0.04617034827006716,
"acc_norm": 0.3137254901960784,
"acc_norm_stderr": 0.04617034827006716
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.5882352941176471,
"acc_stderr": 0.031968769891957786,
"acc_norm": 0.5882352941176471,
"acc_norm_stderr": 0.031968769891957786
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.5358974358974359,
"acc_stderr": 0.02528558599001786,
"acc_norm": 0.5358974358974359,
"acc_norm_stderr": 0.02528558599001786
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.47,
"acc_stderr": 0.050161355804659205,
"acc_norm": 0.47,
"acc_norm_stderr": 0.050161355804659205
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.31,
"acc_stderr": 0.04648231987117316,
"acc_norm": 0.31,
"acc_norm_stderr": 0.04648231987117316
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.5462962962962963,
"acc_stderr": 0.04812917324536823,
"acc_norm": 0.5462962962962963,
"acc_norm_stderr": 0.04812917324536823
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.4433497536945813,
"acc_stderr": 0.03495334582162933,
"acc_norm": 0.4433497536945813,
"acc_norm_stderr": 0.03495334582162933
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.4806451612903226,
"acc_stderr": 0.0284226874043121,
"acc_norm": 0.4806451612903226,
"acc_norm_stderr": 0.0284226874043121
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.688034188034188,
"acc_stderr": 0.03035152732334494,
"acc_norm": 0.688034188034188,
"acc_norm_stderr": 0.03035152732334494
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4226415094339623,
"acc_stderr": 0.03040233144576954,
"acc_norm": 0.4226415094339623,
"acc_norm_stderr": 0.03040233144576954
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.5272727272727272,
"acc_stderr": 0.0478200179138006,
"acc_norm": 0.5272727272727272,
"acc_norm_stderr": 0.0478200179138006
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.3,
"acc_stderr": 0.027940457136228405,
"acc_norm": 0.3,
"acc_norm_stderr": 0.027940457136228405
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.31125827814569534,
"acc_stderr": 0.03780445850526733,
"acc_norm": 0.31125827814569534,
"acc_norm_stderr": 0.03780445850526733
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.5870646766169154,
"acc_stderr": 0.03481520803367348,
"acc_norm": 0.5870646766169154,
"acc_norm_stderr": 0.03481520803367348
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.4046242774566474,
"acc_stderr": 0.03742461193887249,
"acc_norm": 0.4046242774566474,
"acc_norm_stderr": 0.03742461193887249
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.35185185185185186,
"acc_stderr": 0.024594975128920938,
"acc_norm": 0.35185185185185186,
"acc_norm_stderr": 0.024594975128920938
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.3888888888888889,
"acc_stderr": 0.04076663253918567,
"acc_norm": 0.3888888888888889,
"acc_norm_stderr": 0.04076663253918567
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.44,
"acc_stderr": 0.04988876515698589,
"acc_norm": 0.44,
"acc_norm_stderr": 0.04988876515698589
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.57,
"acc_stderr": 0.04975698519562427,
"acc_norm": 0.57,
"acc_norm_stderr": 0.04975698519562427
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.5,
"acc_stderr": 0.026919095102908273,
"acc_norm": 0.5,
"acc_norm_stderr": 0.026919095102908273
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.5153374233128835,
"acc_stderr": 0.03926522378708843,
"acc_norm": 0.5153374233128835,
"acc_norm_stderr": 0.03926522378708843
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.49074074074074076,
"acc_stderr": 0.027815973433878014,
"acc_norm": 0.49074074074074076,
"acc_norm_stderr": 0.027815973433878014
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.29,
"acc_stderr": 0.045604802157206845,
"acc_norm": 0.29,
"acc_norm_stderr": 0.045604802157206845
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.5595854922279793,
"acc_stderr": 0.035827245300360945,
"acc_norm": 0.5595854922279793,
"acc_norm_stderr": 0.035827245300360945
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.3157894736842105,
"acc_stderr": 0.04372748290278008,
"acc_norm": 0.3157894736842105,
"acc_norm_stderr": 0.04372748290278008
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.5339449541284403,
"acc_stderr": 0.021387863350353992,
"acc_norm": 0.5339449541284403,
"acc_norm_stderr": 0.021387863350353992
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.4126984126984127,
"acc_stderr": 0.04403438954768177,
"acc_norm": 0.4126984126984127,
"acc_norm_stderr": 0.04403438954768177
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.5098039215686274,
"acc_stderr": 0.02862441255016795,
"acc_norm": 0.5098039215686274,
"acc_norm_stderr": 0.02862441255016795
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.51,
"acc_stderr": 0.05024183937956913,
"acc_norm": 0.51,
"acc_norm_stderr": 0.05024183937956913
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6776859504132231,
"acc_stderr": 0.042664163633521685,
"acc_norm": 0.6776859504132231,
"acc_norm_stderr": 0.042664163633521685
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.42105263157894735,
"acc_stderr": 0.04017901275981748,
"acc_norm": 0.42105263157894735,
"acc_norm_stderr": 0.04017901275981748
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.41830065359477125,
"acc_stderr": 0.01995597514583555,
"acc_norm": 0.41830065359477125,
"acc_norm_stderr": 0.01995597514583555
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.2907801418439716,
"acc_stderr": 0.027090664368353178,
"acc_norm": 0.2907801418439716,
"acc_norm_stderr": 0.027090664368353178
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.3392857142857143,
"acc_stderr": 0.04493949068613539,
"acc_norm": 0.3392857142857143,
"acc_norm_stderr": 0.04493949068613539
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.4166666666666667,
"acc_stderr": 0.033622774366080424,
"acc_norm": 0.4166666666666667,
"acc_norm_stderr": 0.033622774366080424
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2558659217877095,
"acc_stderr": 0.014593620923210758,
"acc_norm": 0.2558659217877095,
"acc_norm_stderr": 0.014593620923210758
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.43,
"acc_stderr": 0.04975698519562428,
"acc_norm": 0.43,
"acc_norm_stderr": 0.04975698519562428
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.59,
"acc_stderr": 0.04943110704237102,
"acc_norm": 0.59,
"acc_norm_stderr": 0.04943110704237102
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.45588235294117646,
"acc_stderr": 0.03025437257397669,
"acc_norm": 0.45588235294117646,
"acc_norm_stderr": 0.03025437257397669
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.5510204081632653,
"acc_stderr": 0.03184213866687578,
"acc_norm": 0.5510204081632653,
"acc_norm_stderr": 0.03184213866687578
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.5907172995780591,
"acc_stderr": 0.03200704183359591,
"acc_norm": 0.5907172995780591,
"acc_norm_stderr": 0.03200704183359591
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3520208604954368,
"acc_stderr": 0.012198140605353604,
"acc_norm": 0.3520208604954368,
"acc_norm_stderr": 0.012198140605353604
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.4803921568627451,
"acc_stderr": 0.03506612560524866,
"acc_norm": 0.4803921568627451,
"acc_norm_stderr": 0.03506612560524866
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.4666666666666667,
"acc_stderr": 0.038956580652718446,
"acc_norm": 0.4666666666666667,
"acc_norm_stderr": 0.038956580652718446
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.42962056303549573,
"mc1_stderr": 0.01732923458040911,
"mc2": 0.5783267067449626,
"mc2_stderr": 0.01641473247415343
},
"harness|ko_commongen_v2|2": {
"acc": 0.4970484061393152,
"acc_stderr": 0.017190054580194694,
"acc_norm": 0.525383707201889,
"acc_norm_stderr": 0.01716818720142925
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "Alphacode-AI/AlphaMist7B-slr-v2",
"model_sha": "286b2fa985e38a1b51e174e448c75b570a6ee620",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}