evaluation / gpt3_finnish_8b_bigbench_2shot.csv
Muennighoff's picture
Add
40c3147
task,metric,value,err,version
bigbench_analogies,multiple_choice_grade,0.4230769230769231,0.04349844871480441,0
bigbench_arithmetic_1_digit_addition,multiple_choice_grade,0.48,0.05021167315686779,0
bigbench_arithmetic_1_digit_division,multiple_choice_grade,0.6956521739130435,0.09810018692482896,0
bigbench_arithmetic_1_digit_multiplication,multiple_choice_grade,0.41,0.04943110704237102,0
bigbench_arithmetic_1_digit_subtraction,multiple_choice_grade,0.62,0.048783173121456316,0
bigbench_arithmetic_2_digit_addition,multiple_choice_grade,0.5,0.050251890762960605,0
bigbench_arithmetic_2_digit_division,multiple_choice_grade,0.47,0.050161355804659205,0
bigbench_arithmetic_2_digit_multiplication,multiple_choice_grade,0.22,0.041633319989322695,0
bigbench_arithmetic_2_digit_subtraction,multiple_choice_grade,0.51,0.05024183937956912,0
bigbench_arithmetic_3_digit_addition,multiple_choice_grade,0.45,0.05,0
bigbench_arithmetic_3_digit_division,multiple_choice_grade,0.3,0.046056618647183814,0
bigbench_arithmetic_3_digit_multiplication,multiple_choice_grade,0.23,0.04229525846816506,0
bigbench_arithmetic_3_digit_subtraction,multiple_choice_grade,0.55,0.04999999999999999,0
bigbench_arithmetic_4_digit_addition,multiple_choice_grade,0.4,0.04923659639173309,0
bigbench_arithmetic_4_digit_division,multiple_choice_grade,0.25,0.04351941398892446,0
bigbench_arithmetic_4_digit_multiplication,multiple_choice_grade,0.3,0.046056618647183814,0
bigbench_arithmetic_4_digit_subtraction,multiple_choice_grade,0.4,0.04923659639173309,0
bigbench_arithmetic_5_digit_addition,multiple_choice_grade,0.54,0.05009082659620333,0
bigbench_arithmetic_5_digit_division,multiple_choice_grade,0.28,0.04512608598542129,0
bigbench_arithmetic_5_digit_multiplication,multiple_choice_grade,0.3,0.046056618647183814,0
bigbench_arithmetic_5_digit_subtraction,multiple_choice_grade,0.53,0.05016135580465919,0
bigbench_cause_and_effect_one_sentence,multiple_choice_grade,0.49019607843137253,0.07069708383262727,0
bigbench_cause_and_effect_one_sentence_no_prompt,multiple_choice_grade,0.9215686274509803,0.03802101848953982,0
bigbench_cause_and_effect_two_sentences,multiple_choice_grade,0.5490196078431373,0.07037003311735827,0
bigbench_emotions,multiple_choice_grade,0.48125,0.039624688757383286,0
bigbench_empirical_judgments,multiple_choice_grade,0.37373737373737376,0.04887069039502487,0
bigbench_general_knowledge,multiple_choice_grade,0.5,0.0601929265428846,0
bigbench_hhh_alignment_harmless,multiple_choice_grade,0.46551724137931033,0.06606893520605228,0
bigbench_hhh_alignment_helpful,multiple_choice_grade,0.3389830508474576,0.062155747381159164,0
bigbench_hhh_alignment_honest,multiple_choice_grade,0.3898305084745763,0.06403968100905791,0
bigbench_hhh_alignment_other,multiple_choice_grade,0.5581395348837209,0.07662832288817804,0
bigbench_intent_recognition,multiple_choice_grade,0.43641618497109824,0.018866470325851908,0
bigbench_misconceptions,multiple_choice_grade,0.5298507462686567,0.04327816419216089,0
bigbench_paraphrase,multiple_choice_grade,0.55,0.03526639466921486,0
bigbench_sentence_ambiguity,multiple_choice_grade,0.4666666666666667,0.06494964005966064,0
bigbench_similarities_abstraction,multiple_choice_grade,0.6052631578947368,0.056441080498755805,0