evaluation / gpt3_finnish_medium_bigbench.csv
Muennighoff's picture
Add
40c3147
task,metric,value,err,version
bigbench_analogies,multiple_choice_grade,0.3769230769230769,0.04266800546039561,0
bigbench_arithmetic_1_digit_addition,multiple_choice_grade,0.35,0.047937248544110196,0
bigbench_arithmetic_1_digit_division,multiple_choice_grade,0.6521739130434783,0.10154334054280735,0
bigbench_arithmetic_1_digit_multiplication,multiple_choice_grade,0.29,0.045604802157206845,0
bigbench_arithmetic_1_digit_subtraction,multiple_choice_grade,0.25,0.04351941398892446,0
bigbench_arithmetic_2_digit_addition,multiple_choice_grade,0.0,0.0,0
bigbench_arithmetic_2_digit_division,multiple_choice_grade,0.55,0.049999999999999996,0
bigbench_arithmetic_2_digit_multiplication,multiple_choice_grade,0.11,0.031446603773522035,0
bigbench_arithmetic_2_digit_subtraction,multiple_choice_grade,0.23,0.04229525846816505,0
bigbench_arithmetic_3_digit_addition,multiple_choice_grade,0.24,0.042923469599092816,0
bigbench_arithmetic_3_digit_division,multiple_choice_grade,0.22,0.041633319989322695,0
bigbench_arithmetic_3_digit_multiplication,multiple_choice_grade,0.07,0.02564323999762428,0
bigbench_arithmetic_3_digit_subtraction,multiple_choice_grade,0.2,0.04020151261036845,0
bigbench_arithmetic_4_digit_addition,multiple_choice_grade,0.1,0.030151134457776348,0
bigbench_arithmetic_4_digit_division,multiple_choice_grade,0.2,0.04020151261036845,0
bigbench_arithmetic_4_digit_multiplication,multiple_choice_grade,0.05,0.021904291355759026,0
bigbench_arithmetic_4_digit_subtraction,multiple_choice_grade,0.11,0.031446603773522035,0
bigbench_arithmetic_5_digit_addition,multiple_choice_grade,0.07,0.02564323999762429,0
bigbench_arithmetic_5_digit_division,multiple_choice_grade,0.13,0.03379976689896309,0
bigbench_arithmetic_5_digit_multiplication,multiple_choice_grade,0.0,0.0,0
bigbench_arithmetic_5_digit_subtraction,multiple_choice_grade,0.12,0.03265986323710906,0
bigbench_cause_and_effect_one_sentence,multiple_choice_grade,0.49019607843137253,0.07069708383262727,0
bigbench_cause_and_effect_one_sentence_no_prompt,multiple_choice_grade,0.7058823529411765,0.06443794794178427,0
bigbench_cause_and_effect_two_sentences,multiple_choice_grade,0.43137254901960786,0.07004145529212454,0
bigbench_emotions,multiple_choice_grade,0.18125,0.03055034379985447,0
bigbench_empirical_judgments,multiple_choice_grade,0.31313131313131315,0.0468475702186087,0
bigbench_general_knowledge,multiple_choice_grade,0.34285714285714286,0.05714285714285712,0
bigbench_hhh_alignment_harmless,multiple_choice_grade,0.3793103448275862,0.06426835284800642,0
bigbench_hhh_alignment_helpful,multiple_choice_grade,0.3050847457627119,0.060459168847106955,0
bigbench_hhh_alignment_honest,multiple_choice_grade,0.3728813559322034,0.0634959746661109,0
bigbench_hhh_alignment_other,multiple_choice_grade,0.6046511627906976,0.07544284088704808,0
bigbench_intent_recognition,multiple_choice_grade,0.1936416184971098,0.015032263521320763,0
bigbench_misconceptions,multiple_choice_grade,0.4626865671641791,0.04323460286839717,0
bigbench_paraphrase,multiple_choice_grade,0.525,0.0353997274497642,0
bigbench_sentence_ambiguity,multiple_choice_grade,0.5333333333333333,0.06494964005966064,0
bigbench_similarities_abstraction,multiple_choice_grade,0.5263157894736842,0.05765500605317536,0