evaluation / gpt3_finnish_large_bigbench_1shot.csv
Muennighoff's picture
Add
40c3147
raw
history blame
3.32 kB
task,metric,value,err,version
bigbench_analogies,multiple_choice_grade,0.36923076923076925,0.042490254996217565,0
bigbench_arithmetic_1_digit_addition,multiple_choice_grade,0.39,0.04902071300001975,0
bigbench_arithmetic_1_digit_division,multiple_choice_grade,0.6956521739130435,0.09810018692482896,0
bigbench_arithmetic_1_digit_multiplication,multiple_choice_grade,0.37,0.048523658709391,0
bigbench_arithmetic_1_digit_subtraction,multiple_choice_grade,0.41,0.049431107042371025,0
bigbench_arithmetic_2_digit_addition,multiple_choice_grade,0.21,0.040936018074033256,0
bigbench_arithmetic_2_digit_division,multiple_choice_grade,0.4,0.04923659639173309,0
bigbench_arithmetic_2_digit_multiplication,multiple_choice_grade,0.15,0.0358870281282637,0
bigbench_arithmetic_2_digit_subtraction,multiple_choice_grade,0.29,0.045604802157206845,0
bigbench_arithmetic_3_digit_addition,multiple_choice_grade,0.39,0.049020713000019756,0
bigbench_arithmetic_3_digit_division,multiple_choice_grade,0.22,0.041633319989322695,0
bigbench_arithmetic_3_digit_multiplication,multiple_choice_grade,0.24,0.042923469599092816,0
bigbench_arithmetic_3_digit_subtraction,multiple_choice_grade,0.37,0.04852365870939099,0
bigbench_arithmetic_4_digit_addition,multiple_choice_grade,0.28,0.04512608598542127,0
bigbench_arithmetic_4_digit_division,multiple_choice_grade,0.24,0.04292346959909282,0
bigbench_arithmetic_4_digit_multiplication,multiple_choice_grade,0.27,0.044619604333847394,0
bigbench_arithmetic_4_digit_subtraction,multiple_choice_grade,0.31,0.04648231987117316,0
bigbench_arithmetic_5_digit_addition,multiple_choice_grade,0.43,0.049756985195624284,0
bigbench_arithmetic_5_digit_division,multiple_choice_grade,0.2,0.04020151261036845,0
bigbench_arithmetic_5_digit_multiplication,multiple_choice_grade,0.2,0.04020151261036845,0
bigbench_arithmetic_5_digit_subtraction,multiple_choice_grade,0.45,0.049999999999999996,0
bigbench_cause_and_effect_one_sentence,multiple_choice_grade,0.5490196078431373,0.07037003311735827,0
bigbench_cause_and_effect_one_sentence_no_prompt,multiple_choice_grade,0.8431372549019608,0.05143089038668236,0
bigbench_cause_and_effect_two_sentences,multiple_choice_grade,0.47058823529411764,0.07058823529411762,0
bigbench_emotions,multiple_choice_grade,0.26875,0.035156741348767645,0
bigbench_empirical_judgments,multiple_choice_grade,0.40404040404040403,0.04956872738042619,0
bigbench_general_knowledge,multiple_choice_grade,0.35714285714285715,0.05768377522051773,0
bigbench_hhh_alignment_harmless,multiple_choice_grade,0.41379310344827586,0.06523484847771846,0
bigbench_hhh_alignment_helpful,multiple_choice_grade,0.3050847457627119,0.060459168847106955,0
bigbench_hhh_alignment_honest,multiple_choice_grade,0.3728813559322034,0.0634959746661109,0
bigbench_hhh_alignment_other,multiple_choice_grade,0.5813953488372093,0.07612251984976479,0
bigbench_intent_recognition,multiple_choice_grade,0.11127167630057803,0.011962936633615834,0
bigbench_misconceptions,multiple_choice_grade,0.4626865671641791,0.043234602868397164,0
bigbench_paraphrase,multiple_choice_grade,0.46,0.03533045720097816,0
bigbench_sentence_ambiguity,multiple_choice_grade,0.4666666666666667,0.06494964005966064,0
bigbench_similarities_abstraction,multiple_choice_grade,0.6447368421052632,0.055263157894736854,0