{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.3293515358361775, "acc_stderr": 0.013734057652635474, "acc_norm": 0.3890784982935154, "acc_norm_stderr": 0.014247309976045605 }, "harness|ko_hellaswag|10": { "acc": 0.3681537542322247, "acc_stderr": 0.004813177057496272, "acc_norm": 0.4651463851822346, "acc_norm_stderr": 0.004977643730848598 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.4269005847953216, "acc_stderr": 0.03793620616529917, "acc_norm": 0.4269005847953216, "acc_norm_stderr": 0.03793620616529917 }, "harness|ko_mmlu_management|5": { "acc": 0.47572815533980584, "acc_stderr": 0.049449010929737795, "acc_norm": 0.47572815533980584, "acc_norm_stderr": 0.049449010929737795 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.4355044699872286, "acc_stderr": 0.01773058992792661, "acc_norm": 0.4355044699872286, "acc_norm_stderr": 0.01773058992792661 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.4, "acc_stderr": 0.042320736951515885, "acc_norm": 0.4, "acc_norm_stderr": 0.042320736951515885 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.26, "acc_stderr": 0.04408440022768079, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768079 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.35319148936170214, "acc_stderr": 0.031245325202761926, "acc_norm": 0.35319148936170214, "acc_norm_stderr": 0.031245325202761926 }, "harness|ko_mmlu_virology|5": { "acc": 0.35542168674698793, "acc_stderr": 0.03726214354322415, "acc_norm": 0.35542168674698793, "acc_norm_stderr": 0.03726214354322415 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.39228295819935693, "acc_stderr": 0.027731258647011994, "acc_norm": 0.39228295819935693, "acc_norm_stderr": 0.027731258647011994 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.37668161434977576, "acc_stderr": 0.03252113489929188, "acc_norm": 0.37668161434977576, "acc_norm_stderr": 0.03252113489929188 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.37404580152671757, "acc_stderr": 0.042438692422305246, "acc_norm": 0.37404580152671757, "acc_norm_stderr": 0.042438692422305246 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.5303030303030303, "acc_stderr": 0.03555804051763929, "acc_norm": 0.5303030303030303, "acc_norm_stderr": 0.03555804051763929 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.41379310344827586, "acc_stderr": 0.04104269211806232, "acc_norm": 0.41379310344827586, "acc_norm_stderr": 0.04104269211806232 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.17647058823529413, "acc_stderr": 0.0379328118530781, "acc_norm": 0.17647058823529413, "acc_norm_stderr": 0.0379328118530781 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.42857142857142855, "acc_stderr": 0.03214536859788639, "acc_norm": 0.42857142857142855, "acc_norm_stderr": 0.03214536859788639 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.37435897435897436, "acc_stderr": 0.024537591572830517, "acc_norm": 0.37435897435897436, "acc_norm_stderr": 0.024537591572830517 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.52, "acc_stderr": 0.05021167315686779, "acc_norm": 0.52, "acc_norm_stderr": 0.05021167315686779 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.36, "acc_stderr": 0.04824181513244218, "acc_norm": 0.36, "acc_norm_stderr": 0.04824181513244218 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.5, "acc_stderr": 0.04833682445228318, "acc_norm": 0.5, "acc_norm_stderr": 0.04833682445228318 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.3497536945812808, "acc_stderr": 0.03355400904969566, "acc_norm": 0.3497536945812808, "acc_norm_stderr": 0.03355400904969566 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.4064516129032258, "acc_stderr": 0.027941727346256308, "acc_norm": 0.4064516129032258, "acc_norm_stderr": 0.027941727346256308 }, "harness|ko_mmlu_marketing|5": { "acc": 0.6239316239316239, "acc_stderr": 0.03173393632969481, "acc_norm": 0.6239316239316239, "acc_norm_stderr": 0.03173393632969481 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.37735849056603776, "acc_stderr": 0.029832808114796005, "acc_norm": 0.37735849056603776, "acc_norm_stderr": 0.029832808114796005 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.43636363636363634, "acc_stderr": 0.04750185058907296, "acc_norm": 0.43636363636363634, "acc_norm_stderr": 0.04750185058907296 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.2777777777777778, "acc_stderr": 0.027309140588230186, "acc_norm": 0.2777777777777778, "acc_norm_stderr": 0.027309140588230186 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.25165562913907286, "acc_stderr": 0.03543304234389985, "acc_norm": 0.25165562913907286, "acc_norm_stderr": 0.03543304234389985 }, "harness|ko_mmlu_sociology|5": { "acc": 0.5174129353233831, "acc_stderr": 0.03533389234739245, "acc_norm": 0.5174129353233831, "acc_norm_stderr": 0.03533389234739245 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.3583815028901734, "acc_stderr": 0.03656343653353159, "acc_norm": 0.3583815028901734, "acc_norm_stderr": 0.03656343653353159 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.30423280423280424, "acc_stderr": 0.02369541500946308, "acc_norm": 0.30423280423280424, "acc_norm_stderr": 0.02369541500946308 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.3194444444444444, "acc_stderr": 0.03899073687357335, "acc_norm": 0.3194444444444444, "acc_norm_stderr": 0.03899073687357335 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.33, "acc_stderr": 0.04725815626252606, "acc_norm": 0.33, "acc_norm_stderr": 0.04725815626252606 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.54, "acc_stderr": 0.05009082659620333, "acc_norm": 0.54, "acc_norm_stderr": 0.05009082659620333 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.4277456647398844, "acc_stderr": 0.026636539741116082, "acc_norm": 0.4277456647398844, "acc_norm_stderr": 0.026636539741116082 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.38650306748466257, "acc_stderr": 0.038258255488486076, "acc_norm": 0.38650306748466257, "acc_norm_stderr": 0.038258255488486076 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.3734567901234568, "acc_stderr": 0.02691500301138015, "acc_norm": 0.3734567901234568, "acc_norm_stderr": 0.02691500301138015 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.29, "acc_stderr": 0.04560480215720684, "acc_norm": 0.29, "acc_norm_stderr": 0.04560480215720684 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.46113989637305697, "acc_stderr": 0.035975244117345775, "acc_norm": 0.46113989637305697, "acc_norm_stderr": 0.035975244117345775 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.2543859649122807, "acc_stderr": 0.0409698513984367, "acc_norm": 0.2543859649122807, "acc_norm_stderr": 0.0409698513984367 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.42752293577981654, "acc_stderr": 0.02121091020430043, "acc_norm": 0.42752293577981654, "acc_norm_stderr": 0.02121091020430043 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.29365079365079366, "acc_stderr": 0.04073524322147125, "acc_norm": 0.29365079365079366, "acc_norm_stderr": 0.04073524322147125 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.4117647058823529, "acc_stderr": 0.02818059632825929, "acc_norm": 0.4117647058823529, "acc_norm_stderr": 0.02818059632825929 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.41, "acc_stderr": 0.049431107042371025, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "harness|ko_mmlu_international_law|5": { "acc": 0.6115702479338843, "acc_stderr": 0.044492703500683815, "acc_norm": 0.6115702479338843, "acc_norm_stderr": 0.044492703500683815 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.375, "acc_stderr": 0.039397364351956274, "acc_norm": 0.375, "acc_norm_stderr": 0.039397364351956274 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.35784313725490197, "acc_stderr": 0.019393058402355442, "acc_norm": 0.35784313725490197, "acc_norm_stderr": 0.019393058402355442 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.31560283687943264, "acc_stderr": 0.027724989449509314, "acc_norm": 0.31560283687943264, "acc_norm_stderr": 0.027724989449509314 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.2857142857142857, "acc_stderr": 0.04287858751340455, "acc_norm": 0.2857142857142857, "acc_norm_stderr": 0.04287858751340455 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.37962962962962965, "acc_stderr": 0.03309682581119035, "acc_norm": 0.37962962962962965, "acc_norm_stderr": 0.03309682581119035 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.26033519553072626, "acc_stderr": 0.014676252009319475, "acc_norm": 0.26033519553072626, "acc_norm_stderr": 0.014676252009319475 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.35, "acc_stderr": 0.047937248544110196, "acc_norm": 0.35, "acc_norm_stderr": 0.047937248544110196 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.49, "acc_stderr": 0.05024183937956912, "acc_norm": 0.49, "acc_norm_stderr": 0.05024183937956912 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.34558823529411764, "acc_stderr": 0.028888193103988647, "acc_norm": 0.34558823529411764, "acc_norm_stderr": 0.028888193103988647 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.3795918367346939, "acc_stderr": 0.031067211262872485, "acc_norm": 0.3795918367346939, "acc_norm_stderr": 0.031067211262872485 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.48945147679324896, "acc_stderr": 0.032539983791662855, "acc_norm": 0.48945147679324896, "acc_norm_stderr": 0.032539983791662855 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.3076923076923077, "acc_stderr": 0.011787910251664589, "acc_norm": 0.3076923076923077, "acc_norm_stderr": 0.011787910251664589 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.3284313725490196, "acc_stderr": 0.032962451101722294, "acc_norm": 0.3284313725490196, "acc_norm_stderr": 0.032962451101722294 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.3333333333333333, "acc_stderr": 0.036810508691615486, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.036810508691615486 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.25703794369645044, "mc1_stderr": 0.015298077509485083, "mc2": 0.4338489222841401, "mc2_stderr": 0.01562611722529729 }, "harness|ko_commongen_v2|2": { "acc": 0.3659976387249115, "acc_stderr": 0.016561489664895696, "acc_norm": 0.4592680047225502, "acc_norm_stderr": 0.017133218276537673 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "LI-ST/Mistral-7B-ko-v0.006", "model_sha": "85711e2b40b627828aec4ad76955aa66062e9c97", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }