evaluation / gpt2_finnish_bigbench_0shot.csv
Muennighoff's picture
Add
79b60e5
task,metric,value,err,version
bigbench_analogies,multiple_choice_grade,0.23846153846153847,0.03751977598816766,0
bigbench_arithmetic_1_digit_addition,multiple_choice_grade,0.35,0.04793724854411019,0
bigbench_arithmetic_1_digit_division,multiple_choice_grade,0.6521739130434783,0.10154334054280736,0
bigbench_arithmetic_1_digit_multiplication,multiple_choice_grade,0.16,0.03684529491774707,0
bigbench_arithmetic_1_digit_subtraction,multiple_choice_grade,0.29,0.04560480215720683,0
bigbench_arithmetic_2_digit_addition,multiple_choice_grade,0.09,0.028762349126466146,0
bigbench_arithmetic_2_digit_division,multiple_choice_grade,0.46,0.05009082659620333,0
bigbench_arithmetic_2_digit_multiplication,multiple_choice_grade,0.06,0.023868325657594166,0
bigbench_arithmetic_2_digit_subtraction,multiple_choice_grade,0.14,0.034873508801977725,0
bigbench_arithmetic_3_digit_addition,multiple_choice_grade,0.03,0.017144660799776557,0
bigbench_arithmetic_3_digit_division,multiple_choice_grade,0.17,0.0377525168068637,0
bigbench_arithmetic_3_digit_multiplication,multiple_choice_grade,0.04,0.019694638556693216,0
bigbench_arithmetic_3_digit_subtraction,multiple_choice_grade,0.02,0.01407052941362896,0
bigbench_arithmetic_4_digit_addition,multiple_choice_grade,0.16,0.0368452949177471,0
bigbench_arithmetic_4_digit_division,multiple_choice_grade,0.18,0.03861229196653694,0
bigbench_arithmetic_4_digit_multiplication,multiple_choice_grade,0.1,0.030151134457776334,0
bigbench_arithmetic_4_digit_subtraction,multiple_choice_grade,0.08,0.027265992434429086,0
bigbench_arithmetic_5_digit_addition,multiple_choice_grade,0.13,0.033799766898963086,0
bigbench_arithmetic_5_digit_division,multiple_choice_grade,0.1,0.030151134457776334,0
bigbench_arithmetic_5_digit_multiplication,multiple_choice_grade,0.08,0.0272659924344291,0
bigbench_arithmetic_5_digit_subtraction,multiple_choice_grade,0.03,0.017144660799776525,0
bigbench_cause_and_effect_one_sentence,multiple_choice_grade,0.45098039215686275,0.0703700331173583,0
bigbench_cause_and_effect_one_sentence_no_prompt,multiple_choice_grade,0.6862745098039216,0.06562039423796669,0
bigbench_cause_and_effect_two_sentences,multiple_choice_grade,0.6666666666666666,0.06666666666666664,0
bigbench_emotions,multiple_choice_grade,0.2,0.031722063428725716,0
bigbench_empirical_judgments,multiple_choice_grade,0.32323232323232326,0.047245903445151234,0
bigbench_general_knowledge,multiple_choice_grade,0.21428571428571427,0.049397433914866055,0
bigbench_hhh_alignment_harmless,multiple_choice_grade,0.4482758620689655,0.06587130109529918,0
bigbench_hhh_alignment_helpful,multiple_choice_grade,0.2711864406779661,0.05837517703884876,0
bigbench_hhh_alignment_honest,multiple_choice_grade,0.3728813559322034,0.0634959746661109,0
bigbench_hhh_alignment_other,multiple_choice_grade,0.6046511627906976,0.07544284088704808,0
bigbench_intent_recognition,multiple_choice_grade,0.18063583815028902,0.014635292876381762,0
bigbench_misconceptions,multiple_choice_grade,0.48507462686567165,0.04333617784312701,0
bigbench_paraphrase,multiple_choice_grade,0.525,0.035399727449764204,0
bigbench_sentence_ambiguity,multiple_choice_grade,0.5333333333333333,0.06494964005966064,0
bigbench_similarities_abstraction,multiple_choice_grade,0.47368421052631576,0.057655006053175376,0