|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3626279863481229, |
|
"acc_stderr": 0.014049106564955003, |
|
"acc_norm": 0.41467576791808874, |
|
"acc_norm_stderr": 0.014397070564409174 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3784106751643099, |
|
"acc_stderr": 0.004839995745602312, |
|
"acc_norm": 0.49352718581955785, |
|
"acc_norm_stderr": 0.0049893632769551655 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.5146198830409356, |
|
"acc_stderr": 0.03833185275213023, |
|
"acc_norm": 0.5146198830409356, |
|
"acc_norm_stderr": 0.03833185275213023 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6310679611650486, |
|
"acc_stderr": 0.0477761518115674, |
|
"acc_norm": 0.6310679611650486, |
|
"acc_norm_stderr": 0.0477761518115674 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5044699872286079, |
|
"acc_stderr": 0.017879248970584356, |
|
"acc_norm": 0.5044699872286079, |
|
"acc_norm_stderr": 0.017879248970584356 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4148148148148148, |
|
"acc_stderr": 0.04256193767901407, |
|
"acc_norm": 0.4148148148148148, |
|
"acc_norm_stderr": 0.04256193767901407 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.39148936170212767, |
|
"acc_stderr": 0.03190701242326812, |
|
"acc_norm": 0.39148936170212767, |
|
"acc_norm_stderr": 0.03190701242326812 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.3433734939759036, |
|
"acc_stderr": 0.03696584317010601, |
|
"acc_norm": 0.3433734939759036, |
|
"acc_norm_stderr": 0.03696584317010601 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.4919614147909968, |
|
"acc_stderr": 0.028394421370984545, |
|
"acc_norm": 0.4919614147909968, |
|
"acc_norm_stderr": 0.028394421370984545 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.5067264573991032, |
|
"acc_stderr": 0.03355476596234355, |
|
"acc_norm": 0.5067264573991032, |
|
"acc_norm_stderr": 0.03355476596234355 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.4732824427480916, |
|
"acc_stderr": 0.04379024936553893, |
|
"acc_norm": 0.4732824427480916, |
|
"acc_norm_stderr": 0.04379024936553893 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.45, |
|
"acc_stderr": 0.04999999999999999, |
|
"acc_norm": 0.45, |
|
"acc_norm_stderr": 0.04999999999999999 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5757575757575758, |
|
"acc_stderr": 0.03521224908841585, |
|
"acc_norm": 0.5757575757575758, |
|
"acc_norm_stderr": 0.03521224908841585 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.45517241379310347, |
|
"acc_stderr": 0.04149886942192117, |
|
"acc_norm": 0.45517241379310347, |
|
"acc_norm_stderr": 0.04149886942192117 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.23529411764705882, |
|
"acc_stderr": 0.04220773659171453, |
|
"acc_norm": 0.23529411764705882, |
|
"acc_norm_stderr": 0.04220773659171453 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.47058823529411764, |
|
"acc_stderr": 0.03242225027115007, |
|
"acc_norm": 0.47058823529411764, |
|
"acc_norm_stderr": 0.03242225027115007 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.45384615384615384, |
|
"acc_stderr": 0.025242770987126167, |
|
"acc_norm": 0.45384615384615384, |
|
"acc_norm_stderr": 0.025242770987126167 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.58, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.58, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.04512608598542127, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.04512608598542127 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5185185185185185, |
|
"acc_stderr": 0.04830366024635331, |
|
"acc_norm": 0.5185185185185185, |
|
"acc_norm_stderr": 0.04830366024635331 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.3645320197044335, |
|
"acc_stderr": 0.0338640574606209, |
|
"acc_norm": 0.3645320197044335, |
|
"acc_norm_stderr": 0.0338640574606209 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.45806451612903226, |
|
"acc_stderr": 0.028343787250540636, |
|
"acc_norm": 0.45806451612903226, |
|
"acc_norm_stderr": 0.028343787250540636 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7222222222222222, |
|
"acc_stderr": 0.02934311479809445, |
|
"acc_norm": 0.7222222222222222, |
|
"acc_norm_stderr": 0.02934311479809445 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.47547169811320755, |
|
"acc_stderr": 0.030735822206205608, |
|
"acc_norm": 0.47547169811320755, |
|
"acc_norm_stderr": 0.030735822206205608 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.4636363636363636, |
|
"acc_stderr": 0.04776449162396197, |
|
"acc_norm": 0.4636363636363636, |
|
"acc_norm_stderr": 0.04776449162396197 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.028742040903948496, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.028742040903948496 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3576158940397351, |
|
"acc_stderr": 0.03913453431177258, |
|
"acc_norm": 0.3576158940397351, |
|
"acc_norm_stderr": 0.03913453431177258 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6268656716417911, |
|
"acc_stderr": 0.034198326081760065, |
|
"acc_norm": 0.6268656716417911, |
|
"acc_norm_stderr": 0.034198326081760065 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.4393063583815029, |
|
"acc_stderr": 0.037842719328874674, |
|
"acc_norm": 0.4393063583815029, |
|
"acc_norm_stderr": 0.037842719328874674 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.335978835978836, |
|
"acc_stderr": 0.02432631052914915, |
|
"acc_norm": 0.335978835978836, |
|
"acc_norm_stderr": 0.02432631052914915 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4097222222222222, |
|
"acc_stderr": 0.04112490974670787, |
|
"acc_norm": 0.4097222222222222, |
|
"acc_norm_stderr": 0.04112490974670787 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621504, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621504 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.61, |
|
"acc_stderr": 0.04902071300001975, |
|
"acc_norm": 0.61, |
|
"acc_norm_stderr": 0.04902071300001975 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.4913294797687861, |
|
"acc_stderr": 0.026915047355369804, |
|
"acc_norm": 0.4913294797687861, |
|
"acc_norm_stderr": 0.026915047355369804 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.4785276073619632, |
|
"acc_stderr": 0.03924746876751129, |
|
"acc_norm": 0.4785276073619632, |
|
"acc_norm_stderr": 0.03924746876751129 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.4660493827160494, |
|
"acc_stderr": 0.02775653525734767, |
|
"acc_norm": 0.4660493827160494, |
|
"acc_norm_stderr": 0.02775653525734767 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.044084400227680794, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.044084400227680794 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.538860103626943, |
|
"acc_stderr": 0.03597524411734578, |
|
"acc_norm": 0.538860103626943, |
|
"acc_norm_stderr": 0.03597524411734578 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.23684210526315788, |
|
"acc_stderr": 0.039994238792813365, |
|
"acc_norm": 0.23684210526315788, |
|
"acc_norm_stderr": 0.039994238792813365 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.5119266055045871, |
|
"acc_stderr": 0.021431223617362223, |
|
"acc_norm": 0.5119266055045871, |
|
"acc_norm_stderr": 0.021431223617362223 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.31746031746031744, |
|
"acc_stderr": 0.0416345303130286, |
|
"acc_norm": 0.31746031746031744, |
|
"acc_norm_stderr": 0.0416345303130286 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.46078431372549017, |
|
"acc_stderr": 0.028541722692618874, |
|
"acc_norm": 0.46078431372549017, |
|
"acc_norm_stderr": 0.028541722692618874 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.48, |
|
"acc_stderr": 0.050211673156867795, |
|
"acc_norm": 0.48, |
|
"acc_norm_stderr": 0.050211673156867795 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.628099173553719, |
|
"acc_stderr": 0.04412015806624504, |
|
"acc_norm": 0.628099173553719, |
|
"acc_norm_stderr": 0.04412015806624504 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.45394736842105265, |
|
"acc_stderr": 0.04051646342874142, |
|
"acc_norm": 0.45394736842105265, |
|
"acc_norm_stderr": 0.04051646342874142 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.39705882352941174, |
|
"acc_stderr": 0.01979448890002411, |
|
"acc_norm": 0.39705882352941174, |
|
"acc_norm_stderr": 0.01979448890002411 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.32978723404255317, |
|
"acc_stderr": 0.028045946942042398, |
|
"acc_norm": 0.32978723404255317, |
|
"acc_norm_stderr": 0.028045946942042398 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.32142857142857145, |
|
"acc_stderr": 0.0443280405529152, |
|
"acc_norm": 0.32142857142857145, |
|
"acc_norm_stderr": 0.0443280405529152 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.3611111111111111, |
|
"acc_stderr": 0.032757734861009996, |
|
"acc_norm": 0.3611111111111111, |
|
"acc_norm_stderr": 0.032757734861009996 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.27932960893854747, |
|
"acc_stderr": 0.015005762446786154, |
|
"acc_norm": 0.27932960893854747, |
|
"acc_norm_stderr": 0.015005762446786154 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.46, |
|
"acc_stderr": 0.05009082659620332, |
|
"acc_norm": 0.46, |
|
"acc_norm_stderr": 0.05009082659620332 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.050251890762960605, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.050251890762960605 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.39338235294117646, |
|
"acc_stderr": 0.02967428828131118, |
|
"acc_norm": 0.39338235294117646, |
|
"acc_norm_stderr": 0.02967428828131118 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.5673469387755102, |
|
"acc_stderr": 0.031717528240626645, |
|
"acc_norm": 0.5673469387755102, |
|
"acc_norm_stderr": 0.031717528240626645 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.6160337552742616, |
|
"acc_stderr": 0.031658678064106674, |
|
"acc_norm": 0.6160337552742616, |
|
"acc_norm_stderr": 0.031658678064106674 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.34028683181225555, |
|
"acc_stderr": 0.012101217610223805, |
|
"acc_norm": 0.34028683181225555, |
|
"acc_norm_stderr": 0.012101217610223805 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.46568627450980393, |
|
"acc_stderr": 0.03501038327635897, |
|
"acc_norm": 0.46568627450980393, |
|
"acc_norm_stderr": 0.03501038327635897 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.5212121212121212, |
|
"acc_stderr": 0.03900828913737302, |
|
"acc_norm": 0.5212121212121212, |
|
"acc_norm_stderr": 0.03900828913737302 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2827417380660955, |
|
"mc1_stderr": 0.01576477083677731, |
|
"mc2": 0.452859232158323, |
|
"mc2_stderr": 0.015516884053903536 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.4049586776859504, |
|
"acc_stderr": 0.01687694116504561, |
|
"acc_norm": 0.4946871310507674, |
|
"acc_norm_stderr": 0.017189383627229687 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "Changgil/K2S3-Mistral-7b-v1.2", |
|
"model_sha": "52d060cd9e93f176911c91ee232f582f253e7f8f", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |