|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.41723549488054607, |
|
"acc_stderr": 0.01440982551840308, |
|
"acc_norm": 0.49402730375426623, |
|
"acc_norm_stderr": 0.014610348300255802 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.43537143995220073, |
|
"acc_stderr": 0.0049479226926888355, |
|
"acc_norm": 0.59699263095001, |
|
"acc_norm_stderr": 0.004894997736719047 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.6549707602339181, |
|
"acc_stderr": 0.036459813773888065, |
|
"acc_norm": 0.6549707602339181, |
|
"acc_norm_stderr": 0.036459813773888065 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6310679611650486, |
|
"acc_stderr": 0.0477761518115674, |
|
"acc_norm": 0.6310679611650486, |
|
"acc_norm_stderr": 0.0477761518115674 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.6781609195402298, |
|
"acc_stderr": 0.01670638141505791, |
|
"acc_norm": 0.6781609195402298, |
|
"acc_norm_stderr": 0.01670638141505791 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.45925925925925926, |
|
"acc_stderr": 0.04304979692464244, |
|
"acc_norm": 0.45925925925925926, |
|
"acc_norm_stderr": 0.04304979692464244 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.24, |
|
"acc_stderr": 0.04292346959909283, |
|
"acc_norm": 0.24, |
|
"acc_norm_stderr": 0.04292346959909283 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.5106382978723404, |
|
"acc_stderr": 0.03267862331014063, |
|
"acc_norm": 0.5106382978723404, |
|
"acc_norm_stderr": 0.03267862331014063 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.5060240963855421, |
|
"acc_stderr": 0.038922121953330446, |
|
"acc_norm": 0.5060240963855421, |
|
"acc_norm_stderr": 0.038922121953330446 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.6141479099678456, |
|
"acc_stderr": 0.027648149599751468, |
|
"acc_norm": 0.6141479099678456, |
|
"acc_norm_stderr": 0.027648149599751468 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5560538116591929, |
|
"acc_stderr": 0.03334625674242728, |
|
"acc_norm": 0.5560538116591929, |
|
"acc_norm_stderr": 0.03334625674242728 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.6641221374045801, |
|
"acc_stderr": 0.04142313771996665, |
|
"acc_norm": 0.6641221374045801, |
|
"acc_norm_stderr": 0.04142313771996665 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.49, |
|
"acc_stderr": 0.05024183937956914, |
|
"acc_norm": 0.49, |
|
"acc_norm_stderr": 0.05024183937956914 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.7272727272727273, |
|
"acc_stderr": 0.031730712390717244, |
|
"acc_norm": 0.7272727272727273, |
|
"acc_norm_stderr": 0.031730712390717244 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.5655172413793104, |
|
"acc_stderr": 0.04130740879555497, |
|
"acc_norm": 0.5655172413793104, |
|
"acc_norm_stderr": 0.04130740879555497 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.3137254901960784, |
|
"acc_stderr": 0.04617034827006718, |
|
"acc_norm": 0.3137254901960784, |
|
"acc_norm_stderr": 0.04617034827006718 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.6386554621848739, |
|
"acc_stderr": 0.031204691225150016, |
|
"acc_norm": 0.6386554621848739, |
|
"acc_norm_stderr": 0.031204691225150016 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.5307692307692308, |
|
"acc_stderr": 0.025302958890850154, |
|
"acc_norm": 0.5307692307692308, |
|
"acc_norm_stderr": 0.025302958890850154 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.59, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.59, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.047258156262526045, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.047258156262526045 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.6574074074074074, |
|
"acc_stderr": 0.045879047413018105, |
|
"acc_norm": 0.6574074074074074, |
|
"acc_norm_stderr": 0.045879047413018105 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.4039408866995074, |
|
"acc_stderr": 0.03452453903822039, |
|
"acc_norm": 0.4039408866995074, |
|
"acc_norm_stderr": 0.03452453903822039 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.6580645161290323, |
|
"acc_stderr": 0.026985289576552732, |
|
"acc_norm": 0.6580645161290323, |
|
"acc_norm_stderr": 0.026985289576552732 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7948717948717948, |
|
"acc_stderr": 0.02645350805404036, |
|
"acc_norm": 0.7948717948717948, |
|
"acc_norm_stderr": 0.02645350805404036 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.5584905660377358, |
|
"acc_stderr": 0.030561590426731837, |
|
"acc_norm": 0.5584905660377358, |
|
"acc_norm_stderr": 0.030561590426731837 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5909090909090909, |
|
"acc_stderr": 0.04709306978661895, |
|
"acc_norm": 0.5909090909090909, |
|
"acc_norm_stderr": 0.04709306978661895 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.028742040903948485, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.028742040903948485 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3443708609271523, |
|
"acc_stderr": 0.038796870240733264, |
|
"acc_norm": 0.3443708609271523, |
|
"acc_norm_stderr": 0.038796870240733264 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6666666666666666, |
|
"acc_stderr": 0.03333333333333333, |
|
"acc_norm": 0.6666666666666666, |
|
"acc_norm_stderr": 0.03333333333333333 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.5664739884393064, |
|
"acc_stderr": 0.03778621079092056, |
|
"acc_norm": 0.5664739884393064, |
|
"acc_norm_stderr": 0.03778621079092056 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.42063492063492064, |
|
"acc_stderr": 0.025424835086923996, |
|
"acc_norm": 0.42063492063492064, |
|
"acc_norm_stderr": 0.025424835086923996 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.5486111111111112, |
|
"acc_stderr": 0.041614023984032786, |
|
"acc_norm": 0.5486111111111112, |
|
"acc_norm_stderr": 0.041614023984032786 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.72, |
|
"acc_stderr": 0.04512608598542126, |
|
"acc_norm": 0.72, |
|
"acc_norm_stderr": 0.04512608598542126 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.6069364161849711, |
|
"acc_stderr": 0.026296227915613677, |
|
"acc_norm": 0.6069364161849711, |
|
"acc_norm_stderr": 0.026296227915613677 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.5276073619631901, |
|
"acc_stderr": 0.039223782906109894, |
|
"acc_norm": 0.5276073619631901, |
|
"acc_norm_stderr": 0.039223782906109894 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.6234567901234568, |
|
"acc_stderr": 0.026959344518747787, |
|
"acc_norm": 0.6234567901234568, |
|
"acc_norm_stderr": 0.026959344518747787 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.048783173121456316, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.048783173121456316 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.7046632124352331, |
|
"acc_stderr": 0.032922966391551414, |
|
"acc_norm": 0.7046632124352331, |
|
"acc_norm_stderr": 0.032922966391551414 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.47368421052631576, |
|
"acc_stderr": 0.046970851366478626, |
|
"acc_norm": 0.47368421052631576, |
|
"acc_norm_stderr": 0.046970851366478626 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.7064220183486238, |
|
"acc_stderr": 0.019525151122639667, |
|
"acc_norm": 0.7064220183486238, |
|
"acc_norm_stderr": 0.019525151122639667 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.4365079365079365, |
|
"acc_stderr": 0.04435932892851466, |
|
"acc_norm": 0.4365079365079365, |
|
"acc_norm_stderr": 0.04435932892851466 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.6111111111111112, |
|
"acc_stderr": 0.027914055510468008, |
|
"acc_norm": 0.6111111111111112, |
|
"acc_norm_stderr": 0.027914055510468008 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.55, |
|
"acc_stderr": 0.049999999999999996, |
|
"acc_norm": 0.55, |
|
"acc_norm_stderr": 0.049999999999999996 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.7272727272727273, |
|
"acc_stderr": 0.04065578140908705, |
|
"acc_norm": 0.7272727272727273, |
|
"acc_norm_stderr": 0.04065578140908705 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5921052631578947, |
|
"acc_stderr": 0.039993097127774734, |
|
"acc_norm": 0.5921052631578947, |
|
"acc_norm_stderr": 0.039993097127774734 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.5098039215686274, |
|
"acc_stderr": 0.02022394600507432, |
|
"acc_norm": 0.5098039215686274, |
|
"acc_norm_stderr": 0.02022394600507432 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.38652482269503546, |
|
"acc_stderr": 0.029049190342543454, |
|
"acc_norm": 0.38652482269503546, |
|
"acc_norm_stderr": 0.029049190342543454 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.4107142857142857, |
|
"acc_stderr": 0.046695106638751926, |
|
"acc_norm": 0.4107142857142857, |
|
"acc_norm_stderr": 0.046695106638751926 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.5462962962962963, |
|
"acc_stderr": 0.03395322726375797, |
|
"acc_norm": 0.5462962962962963, |
|
"acc_norm_stderr": 0.03395322726375797 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.21564245810055865, |
|
"acc_stderr": 0.013754835975482351, |
|
"acc_norm": 0.21564245810055865, |
|
"acc_norm_stderr": 0.013754835975482351 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.48, |
|
"acc_stderr": 0.050211673156867795, |
|
"acc_norm": 0.48, |
|
"acc_norm_stderr": 0.050211673156867795 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.72, |
|
"acc_stderr": 0.04512608598542128, |
|
"acc_norm": 0.72, |
|
"acc_norm_stderr": 0.04512608598542128 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.47794117647058826, |
|
"acc_stderr": 0.03034326422421352, |
|
"acc_norm": 0.47794117647058826, |
|
"acc_norm_stderr": 0.03034326422421352 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.6040816326530613, |
|
"acc_stderr": 0.03130802899065686, |
|
"acc_norm": 0.6040816326530613, |
|
"acc_norm_stderr": 0.03130802899065686 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.729957805907173, |
|
"acc_stderr": 0.028900721906293426, |
|
"acc_norm": 0.729957805907173, |
|
"acc_norm_stderr": 0.028900721906293426 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.4132985658409387, |
|
"acc_stderr": 0.012576779494860083, |
|
"acc_norm": 0.4132985658409387, |
|
"acc_norm_stderr": 0.012576779494860083 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.6470588235294118, |
|
"acc_stderr": 0.03354092437591519, |
|
"acc_norm": 0.6470588235294118, |
|
"acc_norm_stderr": 0.03354092437591519 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.6848484848484848, |
|
"acc_stderr": 0.0362773057502241, |
|
"acc_norm": 0.6848484848484848, |
|
"acc_norm_stderr": 0.0362773057502241 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.3182374541003672, |
|
"mc1_stderr": 0.016305988648920612, |
|
"mc2": 0.47503231724175854, |
|
"mc2_stderr": 0.015521283734648058 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5678866587957497, |
|
"acc_stderr": 0.017031170198851746, |
|
"acc_norm": 0.5950413223140496, |
|
"acc_norm_stderr": 0.016876941165045616 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "Edentns/DataVortexS-10.7B-v0.4", |
|
"model_sha": "463b40db961f2d4fa6901268e04bd323528391f9", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |