results / PracticeLLM /Custom-KoLLM-13B-v6 /result_2023-11-29 06:51:23.json
open-ko-llm-bot's picture
Add results for 2023-11-29 06:51:23
bbcaa60
raw
history blame
No virus
17.9 kB
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.39334470989761094,
"acc_stderr": 0.014275101465693024,
"acc_norm": 0.4522184300341297,
"acc_norm_stderr": 0.014544519880633832
},
"harness|ko_hellaswag|10": {
"acc": 0.4245170284803824,
"acc_stderr": 0.00493259334881362,
"acc_norm": 0.5660227046405099,
"acc_norm_stderr": 0.004946089230153022
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.47368421052631576,
"acc_stderr": 0.038295098689947286,
"acc_norm": 0.47368421052631576,
"acc_norm_stderr": 0.038295098689947286
},
"harness|ko_mmlu_management|5": {
"acc": 0.5339805825242718,
"acc_stderr": 0.0493929144727348,
"acc_norm": 0.5339805825242718,
"acc_norm_stderr": 0.0493929144727348
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.5376756066411239,
"acc_stderr": 0.017829131764287187,
"acc_norm": 0.5376756066411239,
"acc_norm_stderr": 0.017829131764287187
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.43703703703703706,
"acc_stderr": 0.04284958639753399,
"acc_norm": 0.43703703703703706,
"acc_norm_stderr": 0.04284958639753399
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.29,
"acc_stderr": 0.045604802157206824,
"acc_norm": 0.29,
"acc_norm_stderr": 0.045604802157206824
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.4127659574468085,
"acc_stderr": 0.03218471141400351,
"acc_norm": 0.4127659574468085,
"acc_norm_stderr": 0.03218471141400351
},
"harness|ko_mmlu_virology|5": {
"acc": 0.42168674698795183,
"acc_stderr": 0.038444531817709175,
"acc_norm": 0.42168674698795183,
"acc_norm_stderr": 0.038444531817709175
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.4662379421221865,
"acc_stderr": 0.02833327710956278,
"acc_norm": 0.4662379421221865,
"acc_norm_stderr": 0.02833327710956278
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.48878923766816146,
"acc_stderr": 0.033549366530984746,
"acc_norm": 0.48878923766816146,
"acc_norm_stderr": 0.033549366530984746
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.46564885496183206,
"acc_stderr": 0.043749285605997376,
"acc_norm": 0.46564885496183206,
"acc_norm_stderr": 0.043749285605997376
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001974,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001974
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.5656565656565656,
"acc_stderr": 0.03531505879359182,
"acc_norm": 0.5656565656565656,
"acc_norm_stderr": 0.03531505879359182
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.3724137931034483,
"acc_stderr": 0.0402873153294756,
"acc_norm": 0.3724137931034483,
"acc_norm_stderr": 0.0402873153294756
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.24509803921568626,
"acc_stderr": 0.042801058373643966,
"acc_norm": 0.24509803921568626,
"acc_norm_stderr": 0.042801058373643966
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.46638655462184875,
"acc_stderr": 0.03240501447690071,
"acc_norm": 0.46638655462184875,
"acc_norm_stderr": 0.03240501447690071
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.441025641025641,
"acc_stderr": 0.025174048384000777,
"acc_norm": 0.441025641025641,
"acc_norm_stderr": 0.025174048384000777
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.58,
"acc_stderr": 0.04960449637488583,
"acc_norm": 0.58,
"acc_norm_stderr": 0.04960449637488583
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.39,
"acc_stderr": 0.04902071300001975,
"acc_norm": 0.39,
"acc_norm_stderr": 0.04902071300001975
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.5092592592592593,
"acc_stderr": 0.04832853553437055,
"acc_norm": 0.5092592592592593,
"acc_norm_stderr": 0.04832853553437055
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.3891625615763547,
"acc_stderr": 0.034304624161038716,
"acc_norm": 0.3891625615763547,
"acc_norm_stderr": 0.034304624161038716
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.47419354838709676,
"acc_stderr": 0.02840609505765332,
"acc_norm": 0.47419354838709676,
"acc_norm_stderr": 0.02840609505765332
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.6410256410256411,
"acc_stderr": 0.03142616993791924,
"acc_norm": 0.6410256410256411,
"acc_norm_stderr": 0.03142616993791924
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.4490566037735849,
"acc_stderr": 0.030612730713641092,
"acc_norm": 0.4490566037735849,
"acc_norm_stderr": 0.030612730713641092
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.4727272727272727,
"acc_stderr": 0.04782001791380063,
"acc_norm": 0.4727272727272727,
"acc_norm_stderr": 0.04782001791380063
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.2777777777777778,
"acc_stderr": 0.027309140588230175,
"acc_norm": 0.2777777777777778,
"acc_norm_stderr": 0.027309140588230175
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.304635761589404,
"acc_stderr": 0.03757949922943342,
"acc_norm": 0.304635761589404,
"acc_norm_stderr": 0.03757949922943342
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.5472636815920398,
"acc_stderr": 0.03519702717576915,
"acc_norm": 0.5472636815920398,
"acc_norm_stderr": 0.03519702717576915
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.35260115606936415,
"acc_stderr": 0.036430371689585496,
"acc_norm": 0.35260115606936415,
"acc_norm_stderr": 0.036430371689585496
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.2698412698412698,
"acc_stderr": 0.022860838309232072,
"acc_norm": 0.2698412698412698,
"acc_norm_stderr": 0.022860838309232072
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.3472222222222222,
"acc_stderr": 0.039812405437178615,
"acc_norm": 0.3472222222222222,
"acc_norm_stderr": 0.039812405437178615
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695236,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695236
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.59,
"acc_stderr": 0.04943110704237101,
"acc_norm": 0.59,
"acc_norm_stderr": 0.04943110704237101
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.5,
"acc_stderr": 0.026919095102908273,
"acc_norm": 0.5,
"acc_norm_stderr": 0.026919095102908273
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.44785276073619634,
"acc_stderr": 0.03906947479456601,
"acc_norm": 0.44785276073619634,
"acc_norm_stderr": 0.03906947479456601
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.4567901234567901,
"acc_stderr": 0.02771666165019404,
"acc_norm": 0.4567901234567901,
"acc_norm_stderr": 0.02771666165019404
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.34,
"acc_stderr": 0.04760952285695235,
"acc_norm": 0.34,
"acc_norm_stderr": 0.04760952285695235
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.5233160621761658,
"acc_stderr": 0.036045136724422014,
"acc_norm": 0.5233160621761658,
"acc_norm_stderr": 0.036045136724422014
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.2894736842105263,
"acc_stderr": 0.04266339443159394,
"acc_norm": 0.2894736842105263,
"acc_norm_stderr": 0.04266339443159394
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.5376146788990825,
"acc_stderr": 0.021376575274397576,
"acc_norm": 0.5376146788990825,
"acc_norm_stderr": 0.021376575274397576
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.3253968253968254,
"acc_stderr": 0.04190596438871136,
"acc_norm": 0.3253968253968254,
"acc_norm_stderr": 0.04190596438871136
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.4150326797385621,
"acc_stderr": 0.028213504177824093,
"acc_norm": 0.4150326797385621,
"acc_norm_stderr": 0.028213504177824093
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.44,
"acc_stderr": 0.049888765156985884,
"acc_norm": 0.44,
"acc_norm_stderr": 0.049888765156985884
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.6033057851239669,
"acc_stderr": 0.044658697805310094,
"acc_norm": 0.6033057851239669,
"acc_norm_stderr": 0.044658697805310094
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.35526315789473684,
"acc_stderr": 0.03894734487013317,
"acc_norm": 0.35526315789473684,
"acc_norm_stderr": 0.03894734487013317
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.39052287581699346,
"acc_stderr": 0.019737008998094607,
"acc_norm": 0.39052287581699346,
"acc_norm_stderr": 0.019737008998094607
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.30141843971631205,
"acc_stderr": 0.02737412888263115,
"acc_norm": 0.30141843971631205,
"acc_norm_stderr": 0.02737412888263115
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.2767857142857143,
"acc_stderr": 0.04246624336697624,
"acc_norm": 0.2767857142857143,
"acc_norm_stderr": 0.04246624336697624
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.32407407407407407,
"acc_stderr": 0.03191923445686185,
"acc_norm": 0.32407407407407407,
"acc_norm_stderr": 0.03191923445686185
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.2424581005586592,
"acc_stderr": 0.01433352205921789,
"acc_norm": 0.2424581005586592,
"acc_norm_stderr": 0.01433352205921789
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.32,
"acc_stderr": 0.04688261722621504,
"acc_norm": 0.32,
"acc_norm_stderr": 0.04688261722621504
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.4,
"acc_stderr": 0.04923659639173309,
"acc_norm": 0.4,
"acc_norm_stderr": 0.04923659639173309
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.41911764705882354,
"acc_stderr": 0.029972807170464626,
"acc_norm": 0.41911764705882354,
"acc_norm_stderr": 0.029972807170464626
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.37551020408163266,
"acc_stderr": 0.03100120903989484,
"acc_norm": 0.37551020408163266,
"acc_norm_stderr": 0.03100120903989484
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.6118143459915611,
"acc_stderr": 0.0317229500433233,
"acc_norm": 0.6118143459915611,
"acc_norm_stderr": 0.0317229500433233
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.3122555410691004,
"acc_stderr": 0.011835798135683185,
"acc_norm": 0.3122555410691004,
"acc_norm_stderr": 0.011835798135683185
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.46568627450980393,
"acc_stderr": 0.03501038327635897,
"acc_norm": 0.46568627450980393,
"acc_norm_stderr": 0.03501038327635897
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.5272727272727272,
"acc_stderr": 0.03898531605579418,
"acc_norm": 0.5272727272727272,
"acc_norm_stderr": 0.03898531605579418
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.2766217870257038,
"mc1_stderr": 0.01565960575532691,
"mc2": 0.42310297040812733,
"mc2_stderr": 0.01480446318735434
},
"harness|ko_commongen_v2|2": {
"acc": 0.4498229043683589,
"acc_stderr": 0.01710357334382571,
"acc_norm": 0.5242030696576151,
"acc_norm_stderr": 0.017170202466520748
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "PracticeLLM/Custom-KoLLM-13B-v6",
"model_sha": "4e3b471cca7e769c11dbb9f30ca3dd4256911d66",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}