Datasets:

Tasks:
Other
Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Tags:
License:
P3 / README.md
VictorSanh's picture
update point of contact + data splits sizes
0d6d527
metadata
YAML tags: null
annotations_creators:
  - crowdsourced
  - expert-generated
languages:
  - en
licenses:
  - apache-2.0
multilinguality:
  - monolingual
pretty_name: P3
size_categories:
  - 100M<n<1B
task_categories:
  - other

Dataset Card for P3

Table of Contents

Dataset Description

Dataset Summary

P3 (Pubic Pool of Prompts) is a collection of prompted English datasets covering a diverse set of NLP tasks. A prompt is the combination of an input template and a target template. The templates are functions mapping a data example into natural language for the input and target sequences. For example, in the case of an NLI dataset, the data example would include fields for Premise, Hypothesis, Label. An input template would be If {Premise} is true, is it also true that {Hypothesis}?, whereas a target template can be defined with the label choices Choices[label]. Here Choices is prompt-specific metadata that consists of the options yes, maybe, no corresponding to label being entailment (0), neutral (1) or contradiction (2).

Prompts are collected using Promptsource, an interface to interactively write prompts on datasets, and collect prompt-specific metadata such as evaluation metrics. As of October 13th, there are 2'000 prompts collected for 270+ data(sub)sets. The collection of prompts of P3 is publicly available on Promptsource.

To train T0*, we used a subset of the prompts available in Promptsource (see details here). However, some of the prompts use random.choice, a method that selects uniformly at random an option in a list of valid possibilities. For reproducibility purposes, we release the collection of prompted examples used to train T0*. The data available here are the materialized version of the prompted datasets used in {Multitask Prompted Training Enables Zero-Shot Task Generalization which represent only a subset of the datasets for which there is at least one prompt on Promptsource.

Supported Tasks and Leaderboards

The tasks represented in P3 cover a diverse set of NLP tasks including multiple-choice QA, sentiment analysis or natural language inference. We detail the full list of datasets in Source Data.

Languages

The data in P3 are in English (BCP-47 en).

Dataset Structure

Data Instances

An example of "train" looks as follows:

{
  'answer_choices': ['safe', 'trolley'],
  'inputs': [86, 8, 7142, 666, 6, 405, 8, 3, 834, 1518, 21, 1346, 42, 31682, 58, 37, 3, 929, 9, 3042, 63, 2765, 808, 8, 2045, 6448, 326, 13, 8, 31682, 11, 3, 24052, 135, 16, 8, 1346, 552, 8, 3, 834, 47, 6364, 5], 'inputs_pretokenized': 'In the sentence below, does the _ stand for safe or trolley?\nThe treasury workers took the gold bars off of the trolley and stacked them in the safe until the _ was empty.',
  'targets': [31682, 1],
  'targets_pretokenized': '\ntrolley'
}

In the case of rank classification (letting the model select its the prediction the option with the highest log-likelihood), an example looks as follows:

{
  'idx': [5, 0],
  'inputs': [86, 8, 7142, 666, 6, 405, 8, 3, 834, 1518, 21, 19454, 42, 22227, 58, 19454, 744, 31, 17, 2112, 4553, 17742, 7, 12, 1953, 6, 298, 22227, 966, 373, 405, 5, 3, 834, 19, 72, 952, 12, 619, 16, 3, 9, 17742, 3298, 5],
  'inputs_pretokenized': "In the sentence below, does the _ stand for Kyle or Logan?\nKyle doesn't wear leg warmers to bed, while Logan almost always does. _ is more likely to live in a warmer climate.",
  'is_correct': True,
  'targets': [19454, 1],
  'targets_pretokenized': 'Kyle',
  'weight': 1.0
}

To check all the prompted examples, you can use the Promptsource hosted tool and choose the Prompted dataset viewer mode in the left panel.

Data Fields

The data fields are the same among all splits:

  • answer_choices: the choices (in natural language) available to the model
  • inputs_pretokenized: the natural language input fed to the model
  • targets_pretokenized: the natural language target that the model has to generate
  • inputs: the tokenized input with T5's tokenizer
  • targets: the tokenized target with T5's tokenizer
  • idx: identifier of the (example, answer_option_id) in the case of rank classification
  • weight: a weight for the example produced by seqio (always set to 1.0 in practise)
  • is_correct: whether the (example, answer_option_id) is the correct one

Data Splits

Data(sub)set Number of examples per splits
adversarial_qa_dbert_answer_the_following_q {'train': 10000, 'validation': 1000}
adversarial_qa_dbert_based_on {'train': 10000, 'validation': 1000}
adversarial_qa_dbert_generate_question {'train': 10000, 'validation': 1000, 'test': 1000}
adversarial_qa_dbert_question_context_answer {'train': 10000, 'validation': 1000}
adversarial_qa_dbert_tell_what_it_is {'train': 10000, 'validation': 1000}
adversarial_qa_dbidaf_answer_the_following_q {'train': 10000, 'validation': 1000}
adversarial_qa_dbidaf_based_on {'train': 10000, 'validation': 1000}
adversarial_qa_dbidaf_generate_question {'train': 10000, 'validation': 1000, 'test': 1000}
adversarial_qa_dbidaf_question_context_answer {'train': 10000, 'validation': 1000}
adversarial_qa_dbidaf_tell_what_it_is {'train': 10000, 'validation': 1000}
adversarial_qa_droberta_answer_the_following_q {'train': 10000, 'validation': 1000}
adversarial_qa_droberta_based_on {'train': 10000, 'validation': 1000}
adversarial_qa_droberta_generate_question {'train': 10000, 'validation': 1000, 'test': 1000}
adversarial_qa_droberta_question_context_answer {'train': 10000, 'validation': 1000}
adversarial_qa_droberta_tell_what_it_is {'train': 10000, 'validation': 1000}
ag_news_classify {'train': 120000, 'test': 7600}
ag_news_classify_question_first {'train': 120000, 'test': 7600}
ag_news_classify_with_choices {'train': 120000, 'test': 7600}
ag_news_classify_with_choices_question_first {'train': 120000, 'test': 7600}
ag_news_recommend {'train': 120000, 'test': 7600}
ag_news_which_section {'train': 120000, 'test': 7600}
ag_news_which_section_choices {'train': 120000, 'test': 7600}
ai2_arc_ARC_Challenge_heres_a_problem {'train': 1119, 'validation': 299, 'test': 1172}
ai2_arc_ARC_Challenge_i_am_hesitating {'train': 1119, 'validation': 299, 'test': 1172}
ai2_arc_ARC_Challenge_multiple_choice {'train': 1119, 'validation': 299, 'test': 1172}
ai2_arc_ARC_Challenge_pick_false_options {'train': 1119, 'validation': 299, 'test': 1172}
ai2_arc_ARC_Challenge_pick_the_most_correct_option {'train': 1119, 'validation': 299, 'test': 1172}
ai2_arc_ARC_Challenge_qa_options {'train': 1119, 'validation': 299, 'test': 1172}
ai2_arc_ARC_Easy_heres_a_problem {'train': 2251, 'validation': 570, 'test': 2376}
ai2_arc_ARC_Easy_i_am_hesitating {'train': 2251, 'validation': 570, 'test': 2376}
ai2_arc_ARC_Easy_multiple_choice {'train': 2251, 'validation': 570, 'test': 2376}
ai2_arc_ARC_Easy_pick_false_options {'train': 2251, 'validation': 570, 'test': 2376}
ai2_arc_ARC_Easy_pick_the_most_correct_option {'train': 2251, 'validation': 570, 'test': 2376}
ai2_arc_ARC_Easy_qa_options {'train': 2251, 'validation': 570, 'test': 2376}
amazon_polarity_Is_this_product_review_positive {'train': 3600000, 'test': 400000}
amazon_polarity_Is_this_review {'train': 3600000, 'test': 400000}
amazon_polarity_Is_this_review_negative {'train': 3600000, 'test': 400000}
amazon_polarity_User_recommend_this_product {'train': 3600000, 'test': 400000}
amazon_polarity_convey_negative_or_positive_sentiment {'train': 3600000, 'test': 400000}
amazon_polarity_flattering_or_not {'train': 3600000, 'test': 400000}
amazon_polarity_negative_or_positive_tone {'train': 3600000, 'test': 400000}
amazon_polarity_user_satisfied {'train': 3600000, 'test': 400000}
amazon_polarity_would_you_buy {'train': 3600000, 'test': 400000}
anli_GPT_3_style_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_GPT_3_style_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_GPT_3_style_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_GPT_3_style_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_GPT_3_style_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_GPT_3_style_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_MNLI_crowdsource_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_MNLI_crowdsource_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_MNLI_crowdsource_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_MNLI_crowdsource_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_MNLI_crowdsource_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_MNLI_crowdsource_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_always_sometimes_never_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_always_sometimes_never_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_always_sometimes_never_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_always_sometimes_never_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_always_sometimes_never_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_always_sometimes_never_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_based_on_the_previous_passage_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_based_on_the_previous_passage_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_based_on_the_previous_passage_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_based_on_the_previous_passage_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_based_on_the_previous_passage_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_based_on_the_previous_passage_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_can_we_infer_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_can_we_infer_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_can_we_infer_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_can_we_infer_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_can_we_infer_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_can_we_infer_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_claim_true_false_inconclusive_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_claim_true_false_inconclusive_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_claim_true_false_inconclusive_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_claim_true_false_inconclusive_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_claim_true_false_inconclusive_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_claim_true_false_inconclusive_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_consider_always_sometimes_never_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_consider_always_sometimes_never_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_consider_always_sometimes_never_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_consider_always_sometimes_never_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_consider_always_sometimes_never_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_consider_always_sometimes_never_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_does_it_follow_that_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_does_it_follow_that_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_does_it_follow_that_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_does_it_follow_that_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_does_it_follow_that_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_does_it_follow_that_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_does_this_imply_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_does_this_imply_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_does_this_imply_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_does_this_imply_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_does_this_imply_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_does_this_imply_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_guaranteed_possible_impossible_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_guaranteed_possible_impossible_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_guaranteed_possible_impossible_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_guaranteed_possible_impossible_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_guaranteed_possible_impossible_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_guaranteed_possible_impossible_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_guaranteed_true_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_guaranteed_true_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_guaranteed_true_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_guaranteed_true_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_guaranteed_true_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_guaranteed_true_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_justified_in_saying_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_justified_in_saying_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_justified_in_saying_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_justified_in_saying_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_justified_in_saying_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_justified_in_saying_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_must_be_true_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_must_be_true_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_must_be_true_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_must_be_true_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_must_be_true_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_must_be_true_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_should_assume_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_should_assume_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_should_assume_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_should_assume_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_should_assume_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_should_assume_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
anli_take_the_following_as_truth_r1 {'train': 16946, 'validation': 1000, 'test': 1000}
anli_take_the_following_as_truth_r1_score_eval {'train': 50838, 'validation': 3000, 'test': 3000}
anli_take_the_following_as_truth_r2 {'train': 45460, 'validation': 1000, 'test': 1000}
anli_take_the_following_as_truth_r2_score_eval {'train': 136380, 'validation': 3000, 'test': 3000}
anli_take_the_following_as_truth_r3 {'train': 100459, 'validation': 1200, 'test': 1200}
anli_take_the_following_as_truth_r3_score_eval {'train': 301377, 'validation': 3600, 'test': 3600}
app_reviews_categorize_rating_using_review {'train': 288065}
app_reviews_convert_to_rating {'train': 288065}
app_reviews_convert_to_star_rating {'train': 288065}
app_reviews_generate_review {'train': 288065}
cnn_dailymail_3.0.0_2_or_3_sentences {'train': 287113, 'validation': 13368, 'test': 11490}
cnn_dailymail_3.0.0_generate_story {'train': 287113, 'validation': 13368, 'test': 11490}
cnn_dailymail_3.0.0_news_card_view {'train': 287113, 'validation': 13368, 'test': 11490}
cnn_dailymail_3.0.0_news_stock {'train': 287113, 'validation': 13368, 'test': 11490}
cnn_dailymail_3.0.0_news_summary {'train': 287113, 'validation': 13368, 'test': 11490}
cnn_dailymail_3.0.0_spice_up_story {'train': 287113, 'validation': 13368, 'test': 11490}
cnn_dailymail_3.0.0_sum_in_brief {'train': 287113, 'validation': 13368, 'test': 11490}
cnn_dailymail_3.0.0_tldr_summary {'train': 287113, 'validation': 13368, 'test': 11490}
cnn_dailymail_3.0.0_write_an_outline {'train': 287113, 'validation': 13368, 'test': 11490}
common_gen_Example_prompt {'train': 67389, 'validation': 4018, 'test': 1497}
common_gen_Given_concepts_type_1 {'train': 67389, 'validation': 4018, 'test': 1497}
common_gen_Given_concepts_type_2 {'train': 67389, 'validation': 4018, 'test': 1497}
common_gen_Put_together {'train': 67389, 'validation': 4018, 'test': 1497}
common_gen_choice_in_concept_centric_sentence_generation {'train': 67389, 'validation': 4018, 'test': 1497}
common_gen_random_task_template_prompt {'train': 67389, 'validation': 4018, 'test': 1497}
common_gen_sentence_to_concepts {'train': 67389, 'validation': 4018, 'test': 1497}
common_gen_topic_to_sentence {'train': 67389, 'validation': 4018, 'test': 1497}
common_gen_topics_from_the_sentence {'train': 67389, 'validation': 4018, 'test': 1497}
cos_e_v1.11_aligned_with_common_sense {'train': 9741, 'validation': 1221}
cos_e_v1.11_description_question_option_id {'train': 9741, 'validation': 1221}
cos_e_v1.11_description_question_option_text {'train': 9741, 'validation': 1221}
cos_e_v1.11_explain_why_human {'train': 9741, 'validation': 1221}
cos_e_v1.11_generate_explanation_given_text {'train': 9741, 'validation': 1221}
cos_e_v1.11_i_think {'train': 9741, 'validation': 1221}
cos_e_v1.11_question_description_option_id {'train': 9741, 'validation': 1221}
cos_e_v1.11_question_description_option_text {'train': 9741, 'validation': 1221}
cos_e_v1.11_question_option_description_id {'train': 9741, 'validation': 1221}
cos_e_v1.11_question_option_description_text {'train': 9741, 'validation': 1221}
cos_e_v1.11_rationale {'train': 9741, 'validation': 1221}
cosmos_qa_context_answer_to_question {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_context_description_question_answer_id {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_context_description_question_answer_text {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_context_description_question_text {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_context_question_description_answer_id {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_context_question_description_answer_text {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_context_question_description_text {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_description_context_question_answer_id {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_description_context_question_answer_text {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_description_context_question_text {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_no_prompt_id {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_no_prompt_text {'train': 25262, 'validation': 2985, 'test': 6963}
cosmos_qa_only_question_answer {'train': 25262, 'validation': 2985, 'test': 6963}
dbpedia_14_given_a_choice_of_categories_ {'train': 560000, 'test': 70000}
dbpedia_14_given_a_list_of_category_what_does_the_title_belong_to {'train': 560000, 'test': 70000}
dbpedia_14_given_list_what_category_does_the_paragraph_belong_to {'train': 560000, 'test': 70000}
dbpedia_14_pick_one_category_for_the_following_text {'train': 560000, 'test': 70000}
dream_answer_to_dialogue {'train': 6116, 'validation': 2040, 'test': 2041}
dream_baseline {'train': 6116, 'validation': 2040, 'test': 2041}
dream_generate_first_utterance {'train': 6116, 'validation': 2040, 'test': 2041}
dream_generate_last_utterance {'train': 6116, 'validation': 2040, 'test': 2041}
dream_read_the_following_conversation_and_answer_the_question {'train': 6116, 'validation': 2040, 'test': 2041}
duorc_ParaphraseRC_answer_question {'train': 69524, 'validation': 15591, 'test': 15857}
duorc_ParaphraseRC_build_story_around_qa {'train': 58752, 'validation': 13111, 'test': 13449}
duorc_ParaphraseRC_decide_worth_it {'train': 69524, 'validation': 15591, 'test': 15857}
duorc_ParaphraseRC_extract_answer {'train': 69524, 'validation': 15591, 'test': 15857}
duorc_ParaphraseRC_generate_question {'train': 69524, 'validation': 15591, 'test': 15857}
duorc_ParaphraseRC_generate_question_by_answer {'train': 58752, 'validation': 13111, 'test': 13449}
duorc_ParaphraseRC_movie_director {'train': 69524, 'validation': 15591, 'test': 15857}
duorc_ParaphraseRC_question_answering {'train': 69524, 'validation': 15591, 'test': 15857}
duorc_ParaphraseRC_title_generation {'train': 69524, 'validation': 15591, 'test': 15857}
duorc_SelfRC_answer_question {'train': 60721, 'validation': 12961, 'test': 12559}
duorc_SelfRC_build_story_around_qa {'train': 60094, 'validation': 12845, 'test': 12415}
duorc_SelfRC_decide_worth_it {'train': 60721, 'validation': 12961, 'test': 12559}
duorc_SelfRC_extract_answer {'train': 60721, 'validation': 12961, 'test': 12559}
duorc_SelfRC_generate_question {'train': 60721, 'validation': 12961, 'test': 12559}
duorc_SelfRC_generate_question_by_answer {'train': 60094, 'validation': 12845, 'test': 12415}
duorc_SelfRC_movie_director {'train': 60721, 'validation': 12961, 'test': 12559}
duorc_SelfRC_question_answering {'train': 60721, 'validation': 12961, 'test': 12559}
duorc_SelfRC_title_generation {'train': 60721, 'validation': 12961, 'test': 12559}
gigaword_TLDR {'train': 3803957, 'validation': 189651, 'test': 1951}
gigaword_first_sentence_title {'train': 3803957, 'validation': 189651, 'test': 1951}
gigaword_generate_summary_for_this {'train': 3803957, 'validation': 189651, 'test': 1951}
gigaword_in_a_nutshell {'train': 3803957, 'validation': 189651, 'test': 1951}
gigaword_make_a_title {'train': 3803957, 'validation': 189651, 'test': 1951}
gigaword_reverse_writing {'train': 3803957, 'validation': 189651, 'test': 1951}
gigaword_write_a_title_for_this_sentence {'train': 3803957, 'validation': 189651, 'test': 1951}
gigaword_write_an_article {'train': 3803957, 'validation': 189651, 'test': 1951}
gigaword_write_its_sentence {'train': 3803957, 'validation': 189651, 'test': 1951}
glue_mrpc_equivalent {'train': 3668, 'validation': 408, 'test': 1725}
glue_mrpc_generate_paraphrase {'train': 2474, 'validation': 279, 'test': 1147}
glue_mrpc_generate_sentence {'train': 2474, 'validation': 279, 'test': 1147}
glue_mrpc_paraphrase {'train': 3668, 'validation': 408, 'test': 1725}
glue_mrpc_replace {'train': 3668, 'validation': 408, 'test': 1725}
glue_mrpc_same_thing {'train': 3668, 'validation': 408, 'test': 1725}
glue_mrpc_want_to_know {'train': 3668, 'validation': 408, 'test': 1725}
glue_qqp_answer {'train': 363846, 'validation': 40430, 'test': 390965}
glue_qqp_duplicate {'train': 363846, 'validation': 40430, 'test': 390965}
glue_qqp_duplicate_or_not {'train': 363846, 'validation': 40430, 'test': 390965}
glue_qqp_meaning {'train': 363846, 'validation': 40430, 'test': 390965}
glue_qqp_quora {'train': 363846, 'validation': 40430, 'test': 390965}
glue_qqp_same_thing {'train': 363846, 'validation': 40430, 'test': 390965}
hellaswag_Appropriate_continuation_Yes_or_No {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_Open_ended_completion {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_Open_ended_start {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_Predict_ending_with_hint {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_Predict_ending_with_hint_score_eval {'train': 159620, 'validation': 40168, 'test': 40012}
hellaswag_Randomized_prompts_template {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_Randomized_prompts_template_score_eval {'train': 159620, 'validation': 40168, 'test': 40012}
hellaswag_Reversed_appropriate_continuation_Yes_or_No {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_Topic_of_the_context {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_Topic_without_the_ending_answer {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_complete_first_then {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_complete_first_then_score_eval {'train': 159620, 'validation': 40168, 'test': 40012}
hellaswag_how_ends {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_if_begins_how_continues {'train': 39905, 'validation': 10042, 'test': 10003}
hellaswag_if_begins_how_continues_score_eval {'train': 159620, 'validation': 40168, 'test': 40012}
imdb_Movie_Expressed_Sentiment {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Movie_Expressed_Sentiment_2 {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Negation_template_for_positive_and_negative {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Reviewer_Enjoyment {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Reviewer_Enjoyment_Yes_No {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Reviewer_Expressed_Sentiment {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Reviewer_Opinion_bad_good_choices {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Reviewer_Sentiment_Feeling {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Sentiment_with_choices_ {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Text_Expressed_Sentiment {'train': 25000, 'test': 25000, 'unsupervised': 50000}
imdb_Writer_Expressed_Sentiment {'train': 25000, 'test': 25000, 'unsupervised': 50000}
kilt_tasks_hotpotqa_combining_facts {'train': 88869, 'validation': 5600}
kilt_tasks_hotpotqa_complex_question {'train': 88869, 'validation': 5600}
kilt_tasks_hotpotqa_final_exam {'train': 88869, 'validation': 5600}
kilt_tasks_hotpotqa_formulate {'train': 88869, 'validation': 5600}
kilt_tasks_hotpotqa_straighforward_qa {'train': 88869, 'validation': 5600}
multi_news_distill {'train': 44972, 'validation': 5622, 'test': 5622}
multi_news_expand_reverse_task_ {'train': 44972, 'validation': 5622, 'test': 5622}
multi_news_summarize {'train': 44972, 'validation': 5622, 'test': 5622}
multi_news_summary_scenario {'train': 44972, 'validation': 5622, 'test': 5622}
multi_news_synthesize {'train': 44972, 'validation': 5622, 'test': 5622}
multi_news_what_are_the_key_points {'train': 44972, 'validation': 5622, 'test': 5622}
openbookqa_main_choices {'train': 4957, 'validation': 500, 'test': 500}
openbookqa_main_choose_an_answer_with_options {'train': 4957, 'validation': 500, 'test': 500}
openbookqa_main_only_options {'train': 4957, 'validation': 500, 'test': 500}
openbookqa_main_pick_answer_with_options {'train': 4957, 'validation': 500, 'test': 500}
openbookqa_main_pick_using_id {'train': 4957, 'validation': 500, 'test': 500}
openbookqa_main_which_correct {'train': 4957, 'validation': 500, 'test': 500}
openbookqa_main_which_correct_inverse {'train': 4957, 'validation': 500, 'test': 500}
paws_labeled_final_Concatenation {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_Concatenation_no_label {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_Meaning {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_Meaning_no_label {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_PAWS_ANLI_GPT3 {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_PAWS_ANLI_GPT3_no_label {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_Rewrite {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_Rewrite_no_label {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_context_question {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_context_question_no_label {'train': 49401, 'validation': 8000, 'test': 8000}
paws_labeled_final_paraphrase_task {'train': 21829, 'validation': 3539, 'test': 3536}
paws_labeled_final_task_description_no_label {'train': 49401, 'validation': 8000, 'test': 8000}
piqa_Correct_the_solution {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_Correct_the_solution_if_false_from_sol_1 {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_Correct_the_solution_if_false_from_sol_2 {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_Does_this_solution_make_sense_sol1 {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_Does_this_solution_make_sense_sol2 {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_choose_the_most_appropriate_solution {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_finish_sentence_with_correct_choice {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_no_prompt_needed {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_pick_correct_choice_index {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_pick_correct_choice_with_choice_given_before_goal {'train': 16113, 'validation': 1838, 'test': 3084}
piqa_what_is_the_correct_ending {'train': 16113, 'validation': 1838, 'test': 3084}
qasc_is_correct_1 {'train': 8134, 'validation': 926, 'test': 920}
qasc_is_correct_2 {'train': 8134, 'validation': 926, 'test': 920}
qasc_qa_with_combined_facts_1 {'train': 8134, 'validation': 926, 'test': 920}
qasc_qa_with_separated_facts_1 {'train': 8134, 'validation': 926, 'test': 920}
qasc_qa_with_separated_facts_2 {'train': 8134, 'validation': 926, 'test': 920}
qasc_qa_with_separated_facts_3 {'train': 8134, 'validation': 926, 'test': 920}
qasc_qa_with_separated_facts_4 {'train': 8134, 'validation': 926, 'test': 920}
qasc_qa_with_separated_facts_5 {'train': 8134, 'validation': 926, 'test': 920}
quail_context_description_question_answer_id {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_context_description_question_answer_text {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_context_description_question_text {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_context_question_answer_description_id {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_context_question_answer_description_text {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_context_question_description_answer_id {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_context_question_description_answer_text {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_context_question_description_text {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_description_context_question_answer_id {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_description_context_question_answer_text {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_description_context_question_text {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_no_prompt_id {'train': 10246, 'validation': 2164, 'challenge': 556}
quail_no_prompt_text {'train': 10246, 'validation': 2164, 'challenge': 556}
quarel_choose_between {'train': 1941, 'validation': 278, 'test': 552}
quarel_do_not_use {'train': 1941, 'validation': 278, 'test': 552}
quarel_heres_a_story {'train': 1941, 'validation': 278, 'test': 552}
quarel_logic_test {'train': 1941, 'validation': 278, 'test': 552}
quarel_testing_students {'train': 1941, 'validation': 278, 'test': 552}
quartz_answer_question_based_on {'train': 2696, 'validation': 384, 'test': 784}
quartz_answer_question_below {'train': 2696, 'validation': 384, 'test': 784}
quartz_given_the_fact_answer_the_q {'train': 2696, 'validation': 384, 'test': 784}
quartz_having_read_above_passage {'train': 2696, 'validation': 384, 'test': 784}
quartz_paragraph_question_plain_concat {'train': 2696, 'validation': 384, 'test': 784}
quartz_read_passage_below_choose {'train': 2696, 'validation': 384, 'test': 784}
quartz_use_info_from_paragraph_question {'train': 2696, 'validation': 384, 'test': 784}
quartz_use_info_from_question_paragraph {'train': 2696, 'validation': 384, 'test': 784}
quoref_Answer_Friend_Question {'train': 19399, 'validation': 2418}
quoref_Answer_Question_Given_Context {'train': 19399, 'validation': 2418}
quoref_Answer_Test {'train': 19399, 'validation': 2418}
quoref_Context_Contains_Answer {'train': 19399, 'validation': 2418}
quoref_Find_Answer {'train': 19399, 'validation': 2418}
quoref_Found_Context_Online {'train': 19399, 'validation': 2418}
quoref_Given_Context_Answer_Question {'train': 19399, 'validation': 2418}
quoref_Guess_Answer {'train': 19399, 'validation': 2418}
quoref_Guess_Title_For_Context {'train': 19399, 'validation': 2418}
quoref_Read_And_Extract_ {'train': 19399, 'validation': 2418}
quoref_What_Is_The_Answer {'train': 19399, 'validation': 2418}
race_high_Is_this_the_right_answer {'train': 62445, 'validation': 3451, 'test': 3498}
race_high_Read_the_article_and_answer_the_question_no_option_ {'train': 62445, 'validation': 3451, 'test': 3498}
race_high_Select_the_best_answer {'train': 62445, 'validation': 3451, 'test': 3498}
race_high_Select_the_best_answer_generate_span_ {'train': 62445, 'validation': 3451, 'test': 3498}
race_high_Select_the_best_answer_no_instructions_ {'train': 62445, 'validation': 3451, 'test': 3498}
race_high_Taking_a_test {'train': 62445, 'validation': 3451, 'test': 3498}
race_high_Write_a_multi_choice_question_for_the_following_article {'train': 62445, 'validation': 3451, 'test': 3498}
race_high_Write_a_multi_choice_question_options_given_ {'train': 62445, 'validation': 3451, 'test': 3498}
race_middle_Is_this_the_right_answer {'train': 25421, 'validation': 1436, 'test': 1436}
race_middle_Read_the_article_and_answer_the_question_no_option_ {'train': 25421, 'validation': 1436, 'test': 1436}
race_middle_Select_the_best_answer {'train': 25421, 'validation': 1436, 'test': 1436}
race_middle_Select_the_best_answer_generate_span_ {'train': 25421, 'validation': 1436, 'test': 1436}
race_middle_Select_the_best_answer_no_instructions_ {'train': 25421, 'validation': 1436, 'test': 1436}
race_middle_Taking_a_test {'train': 25421, 'validation': 1436, 'test': 1436}
race_middle_Write_a_multi_choice_question_for_the_following_article {'train': 25421, 'validation': 1436, 'test': 1436}
race_middle_Write_a_multi_choice_question_options_given_ {'train': 25421, 'validation': 1436, 'test': 1436}
ropes_background_new_situation_answer {'train': 10924, 'validation': 1688}
ropes_background_situation_middle {'train': 10924, 'validation': 1688}
ropes_given_background_situation {'train': 10924, 'validation': 1688}
ropes_new_situation_background_answer {'train': 10924, 'validation': 1688}
ropes_plain_background_situation {'train': 10924, 'validation': 1688}
ropes_plain_bottom_hint {'train': 10924, 'validation': 1688}
ropes_plain_no_background {'train': 10924, 'validation': 1688}
ropes_prompt_beginning {'train': 10924, 'validation': 1688}
ropes_prompt_bottom_hint_beginning {'train': 10924, 'validation': 1688}
ropes_prompt_bottom_no_hint {'train': 10924, 'validation': 1688}
ropes_prompt_mix {'train': 10924, 'validation': 1688}
ropes_read_background_situation {'train': 10924, 'validation': 1688}
rotten_tomatoes_Movie_Expressed_Sentiment {'train': 8530, 'validation': 1066, 'test': 1066}
rotten_tomatoes_Movie_Expressed_Sentiment_2 {'train': 8530, 'validation': 1066, 'test': 1066}
rotten_tomatoes_Reviewer_Enjoyment {'train': 8530, 'validation': 1066, 'test': 1066}
rotten_tomatoes_Reviewer_Enjoyment_Yes_No {'train': 8530, 'validation': 1066, 'test': 1066}
rotten_tomatoes_Reviewer_Expressed_Sentiment {'train': 8530, 'validation': 1066, 'test': 1066}
rotten_tomatoes_Reviewer_Opinion_bad_good_choices {'train': 8530, 'validation': 1066, 'test': 1066}
rotten_tomatoes_Reviewer_Sentiment_Feeling {'train': 8530, 'validation': 1066, 'test': 1066}
rotten_tomatoes_Sentiment_with_choices_ {'train': 8530, 'validation': 1066, 'test': 1066}
rotten_tomatoes_Text_Expressed_Sentiment {'train': 8530, 'validation': 1066, 'test': 1066}
rotten_tomatoes_Writer_Expressed_Sentiment {'train': 8530, 'validation': 1066, 'test': 1066}
samsum_Generate_a_summary_for_this_dialogue {'train': 14732, 'validation': 818, 'test': 819}
samsum_Given_the_above_dialogue_write_a_summary {'train': 14732, 'validation': 818, 'test': 819}
samsum_Sum_up_the_following_dialogue {'train': 14732, 'validation': 818, 'test': 819}
samsum_Summarize_ {'train': 14732, 'validation': 818, 'test': 819}
samsum_Summarize_this_dialogue_ {'train': 14732, 'validation': 818, 'test': 819}
samsum_To_sum_up_this_dialog {'train': 14732, 'validation': 818, 'test': 819}
samsum_Write_a_dialogue_that_match_this_summary {'train': 14732, 'validation': 818, 'test': 819}
sciq_Direct_Question {'train': 11679, 'validation': 1000, 'test': 1000}
sciq_Direct_Question_Closed_Book_ {'train': 11679, 'validation': 1000, 'test': 1000}
sciq_Multiple_Choice {'train': 11679, 'validation': 1000, 'test': 1000}
sciq_Multiple_Choice_Closed_Book_ {'train': 11679, 'validation': 1000, 'test': 1000}
sciq_Multiple_Choice_Question_First {'train': 11679, 'validation': 1000, 'test': 1000}
social_i_qa_Check_if_a_random_answer_is_valid_or_not {'train': 33410, 'validation': 1954}
social_i_qa_Generate_answer {'train': 33410, 'validation': 1954}
social_i_qa_Generate_the_question_from_the_answer {'train': 33410, 'validation': 1954}
social_i_qa_I_was_wondering {'train': 33410, 'validation': 1954}
social_i_qa_Show_choices_and_generate_answer {'train': 33410, 'validation': 1954}
social_i_qa_Show_choices_and_generate_index {'train': 33410, 'validation': 1954}
squad_v2_Jeopardy_with_Context {'train': 86821, 'validation': 5928}
squad_v2_Jeopardy_without_Context {'train': 86821, 'validation': 5928}
squad_v2_Questions_with_Context {'train': 130319, 'validation': 11873}
squad_v2_Questions_with_Context_Without_Prompt_Keywords {'train': 130319, 'validation': 11873}
squad_v2_Questions_with_Context_Without_Prompt_Keywords_unanswerable {'train': 130319, 'validation': 11873}
squad_v2_Questions_with_Context_unanswerable {'train': 130319, 'validation': 11873}
squad_v2_Topic_Prediction_Context {'train': 130319, 'validation': 11873}
squad_v2_Topic_Prediction_Context_with_randomized_prompt_options {'train': 130319, 'validation': 11873}
squad_v2_Topic_Prediction_Context_with_randomized_prompt_options_placed_in_the_end {'train': 130319, 'validation': 11873}
squad_v2_Topic_Prediction_Question_and_Answer_Pair {'train': 86821, 'validation': 5928}
squad_v2_Trivia {'train': 86821, 'validation': 5928}
squad_v2_Unanwerable_question {'train': 130319, 'validation': 11873}
super_glue_boolq_GPT_3_Style {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_boolq_I_wonder_ {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_boolq_after_reading {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_boolq_based_on_the_following_passage {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_boolq_based_on_the_previous_passage {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_boolq_could_you_tell_me_ {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_boolq_exam {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_boolq_exercise {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_boolq_valid_binary {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_boolq_yes_no_question {'train': 9427, 'validation': 3270, 'test': 3245}
super_glue_cb_GPT_3_style {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_GPT_3_style_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_MNLI_crowdsource {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_MNLI_crowdsource_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_always_sometimes_never {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_always_sometimes_never_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_based_on_the_previous_passage {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_based_on_the_previous_passage_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_can_we_infer {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_can_we_infer_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_claim_true_false_inconclusive {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_claim_true_false_inconclusive_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_consider_always_sometimes_never {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_consider_always_sometimes_never_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_does_it_follow_that {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_does_it_follow_that_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_does_this_imply {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_does_this_imply_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_guaranteed_possible_impossible {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_guaranteed_possible_impossible_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_guaranteed_true {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_guaranteed_true_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_justified_in_saying {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_justified_in_saying_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_must_be_true {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_must_be_true_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_should_assume {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_should_assume_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_cb_take_the_following_as_truth {'train': 250, 'validation': 56, 'test': 250}
super_glue_cb_take_the_following_as_truth_score_eval {'train': 750, 'validation': 168, 'test': 750}
super_glue_copa_C1_or_C2_premise_so_because_ {'train': 400, 'validation': 100, 'test': 500}
super_glue_copa_C1_or_C2_premise_so_because__score_eval {'train': 800, 'validation': 200, 'test': 1000}
super_glue_copa__As_a_result_C1_or_C2_ {'train': 202, 'validation': 48, 'test': 250}
super_glue_copa__As_a_result_C1_or_C2__score_eval {'train': 404, 'validation': 96, 'test': 500}
super_glue_copa__What_could_happen_next_C1_or_C2_ {'train': 202, 'validation': 48, 'test': 250}
super_glue_copa__What_could_happen_next_C1_or_C2__score_eval {'train': 404, 'validation': 96, 'test': 500}
super_glue_copa__which_may_be_caused_by {'train': 198, 'validation': 52, 'test': 250}
super_glue_copa__which_may_be_caused_by_score_eval {'train': 396, 'validation': 104, 'test': 500}
super_glue_copa__why_C1_or_C2 {'train': 198, 'validation': 52, 'test': 250}
super_glue_copa__why_C1_or_C2_score_eval {'train': 396, 'validation': 104, 'test': 500}
super_glue_copa_best_option {'train': 400, 'validation': 100, 'test': 500}
super_glue_copa_best_option_score_eval {'train': 800, 'validation': 200, 'test': 1000}
super_glue_copa_cause_effect {'train': 400, 'validation': 100, 'test': 500}
super_glue_copa_cause_effect_score_eval {'train': 800, 'validation': 200, 'test': 1000}
super_glue_copa_choose {'train': 400, 'validation': 100, 'test': 500}
super_glue_copa_choose_score_eval {'train': 800, 'validation': 200, 'test': 1000}
super_glue_copa_exercise {'train': 400, 'validation': 100, 'test': 500}
super_glue_copa_exercise_score_eval {'train': 800, 'validation': 200, 'test': 1000}
super_glue_copa_i_am_hesitating {'train': 400, 'validation': 100, 'test': 500}
super_glue_copa_i_am_hesitating_score_eval {'train': 800, 'validation': 200, 'test': 1000}
super_glue_copa_more_likely {'train': 400, 'validation': 100, 'test': 500}
super_glue_copa_more_likely_score_eval {'train': 800, 'validation': 200, 'test': 1000}
super_glue_copa_plausible_alternatives {'train': 400, 'validation': 100, 'test': 500}
super_glue_copa_plausible_alternatives_score_eval {'train': 800, 'validation': 200, 'test': 1000}
super_glue_multirc_I_was_going_to_say_ {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_multirc_Would_it_be_good_to_answer_ {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_multirc_confirm {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_multirc_correct {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_multirc_decide_valid {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_multirc_found_this_answer {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_multirc_grading {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_multirc_is_a_correct_answer_ {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_multirc_is_the_correct_answer_ {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_multirc_paragraph_question_is_it_ {'train': 27243, 'validation': 4848, 'test': 9693}
super_glue_record_Add_sentence_after_after_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_Add_sentence_after_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_Can_you_figure_out_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_GPT_3_style_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_GPT_3_style_summary_only_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_GPT_3_style_with_labels_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_GPT_3_style_with_labels_without_hyphens_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_GPT_3_style_without_hyphens_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_In_the_question_above_the_placeholder_stands_for {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_New_highlight_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_News_article_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_Summary_first_continuation_choices_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_What_could_the_placeholder_be_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_Which_one_is_the_placeholder_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_choose_between {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_corrupted {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_exercise {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_pick_one_option {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_the_placeholder_refers_to_ {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_record_trying_to_decide {'train': 100730, 'validation': 10000, 'test': 10000}
super_glue_rte_GPT_3_style {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_GPT_3_style_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_rte_MNLI_crowdsource {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_MNLI_crowdsource_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_rte_based_on_the_previous_passage {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_based_on_the_previous_passage_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_rte_can_we_infer {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_can_we_infer_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_rte_does_it_follow_that {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_does_it_follow_that_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_rte_does_this_imply {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_does_this_imply_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_rte_guaranteed_true {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_guaranteed_true_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_rte_justified_in_saying {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_justified_in_saying_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_rte_must_be_true {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_must_be_true_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_rte_should_assume {'train': 2490, 'validation': 277, 'test': 3000}
super_glue_rte_should_assume_score_eval {'train': 4980, 'validation': 554, 'test': 6000}
super_glue_wic_GPT_3_prompt {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_GPT_3_prompt_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wic_GPT_3_prompt_with_label {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_GPT_3_prompt_with_label_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wic_affirmation_true_or_false {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_affirmation_true_or_false_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wic_grammar_homework {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_grammar_homework_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wic_polysemous {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_polysemous_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wic_question_context {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_question_context_meaning {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_question_context_meaning_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wic_question_context_meaning_with_label {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_question_context_meaning_with_label_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wic_question_context_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wic_same_sense {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_same_sense_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wic_similar_sense {'train': 5428, 'validation': 638, 'test': 1400}
super_glue_wic_similar_sense_score_eval {'train': 10856, 'validation': 1276, 'test': 2800}
super_glue_wsc.fixed_GPT_3_Style {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_GPT_3_Style_score_eval {'train': 1108, 'validation': 208, 'test': 292}
super_glue_wsc.fixed_I_think_they_mean {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_I_think_they_mean_score_eval {'train': 1108, 'validation': 208, 'test': 292}
super_glue_wsc.fixed_Who_or_what_is_are {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_Who_or_what_is_are_score_eval {'train': 1108, 'validation': 208, 'test': 292}
super_glue_wsc.fixed_by_p_they_mean {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_by_p_they_mean_score_eval {'train': 1108, 'validation': 208, 'test': 292}
super_glue_wsc.fixed_does_p_stand_for {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_does_p_stand_for_score_eval {'train': 1108, 'validation': 208, 'test': 292}
super_glue_wsc.fixed_does_the_pronoun_refer_to {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_does_the_pronoun_refer_to_score_eval {'train': 1108, 'validation': 208, 'test': 292}
super_glue_wsc.fixed_in_other_words {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_in_other_words_score_eval {'train': 1108, 'validation': 208, 'test': 292}
super_glue_wsc.fixed_p_is_are_r {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_p_is_are_r_score_eval {'train': 1108, 'validation': 208, 'test': 292}
super_glue_wsc.fixed_replaced_with {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_replaced_with_score_eval {'train': 1108, 'validation': 208, 'test': 292}
super_glue_wsc.fixed_the_pronoun_refers_to {'train': 554, 'validation': 104, 'test': 146}
super_glue_wsc.fixed_the_pronoun_refers_to_score_eval {'train': 1108, 'validation': 208, 'test': 292}
trec_fine_grained_ABBR {'train': 86, 'test': 9}
trec_fine_grained_ABBR_context_first {'train': 86, 'test': 9}
trec_fine_grained_DESC {'train': 1162, 'test': 138}
trec_fine_grained_DESC_context_first {'train': 1162, 'test': 138}
trec_fine_grained_ENTY {'train': 1250, 'test': 94}
trec_fine_grained_HUM {'train': 1223, 'test': 65}
trec_fine_grained_HUM_context_first {'train': 1223, 'test': 65}
trec_fine_grained_LOC {'train': 835, 'test': 81}
trec_fine_grained_LOC_context_first {'train': 835, 'test': 81}
trec_fine_grained_NUM {'train': 896, 'test': 113}
trec_fine_grained_NUM_context_first {'train': 896, 'test': 113}
trec_fine_grained_open {'train': 5452, 'test': 500}
trec_fine_grained_open_context_first {'train': 5452, 'test': 500}
trec_pick_the_best_descriptor {'train': 5452, 'test': 500}
trec_trec1 {'train': 5452, 'test': 500}
trec_trec2 {'train': 5452, 'test': 500}
trec_what_category_best_describe {'train': 5452, 'test': 500}
trec_which_category_best_describes {'train': 5452, 'test': 500}
trivia_qa_unfiltered_first_person_context {'train': 87622, 'validation': 11313, 'test': 10832}
trivia_qa_unfiltered_formal_description {'train': 87622, 'validation': 11313, 'test': 10832}
trivia_qa_unfiltered_guess_question {'train': 87622, 'validation': 11313}
trivia_qa_unfiltered_question_answer {'train': 87622, 'validation': 11313, 'test': 10832}
trivia_qa_unfiltered_question_with_instruction {'train': 87622, 'validation': 11313, 'test': 10832}
web_questions_get_the_answer {'train': 3778, 'test': 2032}
web_questions_potential_correct_answer {'train': 3778, 'test': 2032}
web_questions_question_answer {'train': 3778, 'test': 2032}
web_questions_short_general_knowledge_q {'train': 3778, 'test': 2032}
web_questions_whats_the_answer {'train': 3778, 'test': 2032}
wiki_bio_comprehension {'train': 582639, 'test': 72829, 'val': 72831}
wiki_bio_guess_person {'train': 582639, 'test': 72829, 'val': 72831}
wiki_bio_key_content {'train': 582639, 'test': 72829, 'val': 72831}
wiki_bio_what_content {'train': 582639, 'test': 72829, 'val': 72831}
wiki_bio_who {'train': 582639, 'test': 72829, 'val': 72831}
wiki_hop_original_choose_best_object_affirmative_1 {'train': 43738, 'validation': 5129}
wiki_hop_original_choose_best_object_affirmative_2 {'train': 43738, 'validation': 5129}
wiki_hop_original_choose_best_object_affirmative_3 {'train': 43738, 'validation': 5129}
wiki_hop_original_choose_best_object_interrogative_1 {'train': 43738, 'validation': 5129}
wiki_hop_original_choose_best_object_interrogative_2 {'train': 43738, 'validation': 5129}
wiki_hop_original_explain_relation {'train': 43738, 'validation': 5129}
wiki_hop_original_generate_object {'train': 43738, 'validation': 5129}
wiki_hop_original_generate_subject {'train': 43738, 'validation': 5129}
wiki_hop_original_generate_subject_and_object {'train': 43738, 'validation': 5129}
wiki_qa_Decide_good_answer {'train': 20360, 'validation': 2733, 'test': 6165}
wiki_qa_Direct_Answer_to_Question {'train': 1040, 'validation': 140, 'test': 293}
wiki_qa_Generate_Question_from_Topic {'train': 1040, 'validation': 140, 'test': 293}
wiki_qa_Is_This_True_ {'train': 20360, 'validation': 2733, 'test': 6165}
wiki_qa_Jeopardy_style {'train': 1040, 'validation': 140, 'test': 293}
wiki_qa_Topic_Prediction_Answer_Only {'train': 1040, 'validation': 140, 'test': 293}
wiki_qa_Topic_Prediction_Question_Only {'train': 1040, 'validation': 140, 'test': 293}
wiki_qa_Topic_Prediction_Question_and_Answer_Pair {'train': 1040, 'validation': 140, 'test': 293}
wiki_qa_automatic_system {'train': 20360, 'validation': 2733, 'test': 6165}
wiki_qa_exercise {'train': 20360, 'validation': 2733, 'test': 6165}
wiki_qa_found_on_google {'train': 20360, 'validation': 2733, 'test': 6165}
winogrande_winogrande_debiased_Replace {'train': 9248, 'validation': 1267, 'test': 1767}
winogrande_winogrande_debiased_Replace_score_eval {'train': 18496, 'validation': 2534, 'test': 3534}
winogrande_winogrande_debiased_does_underscore_refer_to {'train': 9248, 'validation': 1267, 'test': 1767}
winogrande_winogrande_debiased_does_underscore_refer_to_score_eval {'train': 18496, 'validation': 2534, 'test': 3534}
winogrande_winogrande_debiased_fill_in_the_blank {'train': 9248, 'validation': 1267, 'test': 1767}
winogrande_winogrande_debiased_fill_in_the_blank_score_eval {'train': 18496, 'validation': 2534, 'test': 3534}
winogrande_winogrande_debiased_stand_for {'train': 9248, 'validation': 1267, 'test': 1767}
winogrande_winogrande_debiased_stand_for_score_eval {'train': 18496, 'validation': 2534, 'test': 3534}
winogrande_winogrande_debiased_underscore_refer_to {'train': 9248, 'validation': 1267, 'test': 1767}
winogrande_winogrande_debiased_underscore_refer_to_score_eval {'train': 18496, 'validation': 2534, 'test': 3534}
winogrande_winogrande_xl_Replace {'train': 40398, 'validation': 1267, 'test': 1767}
winogrande_winogrande_xl_Replace_score_eval {'train': 80796, 'validation': 2534, 'test': 3534}
winogrande_winogrande_xl_does_underscore_refer_to {'train': 40398, 'validation': 1267, 'test': 1767}
winogrande_winogrande_xl_does_underscore_refer_to_score_eval {'train': 80796, 'validation': 2534, 'test': 3534}
winogrande_winogrande_xl_fill_in_the_blank {'train': 40398, 'validation': 1267, 'test': 1767}
winogrande_winogrande_xl_fill_in_the_blank_score_eval {'train': 80796, 'validation': 2534, 'test': 3534}
winogrande_winogrande_xl_stand_for {'train': 40398, 'validation': 1267, 'test': 1767}
winogrande_winogrande_xl_stand_for_score_eval {'train': 80796, 'validation': 2534, 'test': 3534}
winogrande_winogrande_xl_underscore_refer_to {'train': 40398, 'validation': 1267, 'test': 1767}
winogrande_winogrande_xl_underscore_refer_to_score_eval {'train': 80796, 'validation': 2534, 'test': 3534}
wiqa_does_the_supposed_perturbation_have_an_effect {'train': 29808, 'validation': 6894, 'test': 3003}
wiqa_effect_with_label_answer {'train': 29808, 'validation': 6894, 'test': 3003}
wiqa_effect_with_string_answer {'train': 29808, 'validation': 6894, 'test': 3003}
wiqa_what_is_the_final_step_of_the_following_process {'train': 29808, 'validation': 6894, 'test': 3003}
wiqa_what_is_the_missing_first_step {'train': 29808, 'validation': 6894, 'test': 3003}
wiqa_what_might_be_the_first_step_of_the_process {'train': 29808, 'validation': 6894, 'test': 3003}
wiqa_what_might_be_the_last_step_of_the_process {'train': 29808, 'validation': 6894, 'test': 3003}
wiqa_which_of_the_following_is_the_supposed_perturbation {'train': 29808, 'validation': 6894, 'test': 3003}
xsum_DOC_boils_down_to_simple_idea_that {'train': 204045, 'validation': 11332, 'test': 11334}
xsum_DOC_given_above_write_one_sentence {'train': 204045, 'validation': 11332, 'test': 11334}
xsum_DOC_how_would_you_rephrase_few_words {'train': 204045, 'validation': 11332, 'test': 11334}
xsum_DOC_tldr {'train': 204045, 'validation': 11332, 'test': 11334}
xsum_DOC_write_summary_of_above {'train': 204045, 'validation': 11332, 'test': 11334}
xsum_article_DOC_summary {'train': 204045, 'validation': 11332, 'test': 11334}
xsum_college_roommate_asked_DOC_so_I_recap {'train': 204045, 'validation': 11332, 'test': 11334}
xsum_read_below_DOC_write_abstract {'train': 204045, 'validation': 11332, 'test': 11334}
xsum_summarize_DOC {'train': 204045, 'validation': 11332, 'test': 11334}
xsum_summarize_this_DOC_summary {'train': 204045, 'validation': 11332, 'test': 11334}
yelp_review_full_based_on_that {'train': 650000, 'test': 50000}
yelp_review_full_format_rating {'train': 650000, 'test': 50000}
yelp_review_full_format_score {'train': 650000, 'test': 50000}
yelp_review_full_format_star {'train': 650000, 'test': 50000}
yelp_review_full_on_a_scale {'train': 650000, 'test': 50000}
yelp_review_full_so_i_would {'train': 650000, 'test': 50000}
yelp_review_full_this_place {'train': 650000, 'test': 50000}

Dataset Creation

Curation Rationale

The Public Pool of Prompts relies on the Hugging Face Dataset library. Any public dataset in the Datasets library can be prompted. We select the datasets that have at least one subset in English and excluded datasets containing (predominantly) non-natural language examples.

We conservatively decided not to prompt datasets that contain potentially harmful content (for instance, datasets built on social media content). However, we sometimes prompt datasets that are purposefully built to measure bias and fairness of trained models, and reserve these prompted datasets (the validation or test sets) for evaluation purposes.

Source Data

Here's the full list of the datasets present in the materialized version of P3:

  • Multiple-Choice QA
    • CommonsenseQA
    • DREAM
    • QUAIL
    • QuaRTz
    • Social IQA
    • WiQA
    • Cosmos
    • QASC
    • Quarel
    • SciQ
    • Wiki Hop
    • ARC
    • OpenBookQA
    • MultiRC
    • PIQA
    • RACE
    • HellaSwag
    • BoolQ
  • Extractive QA
    • Adversarial QA
    • Quoref
    • DuoRC
    • ROPES
    • SQuAD v2
    • ReCoRD
  • Close-book QA
    • Hotpot QA
    • Wiki QA
    • Trivia QA
    • Web Questions
  • Structure-to-text
    • Common Gen
    • Wiki Bio
  • Sentiment
    • Amazon
    • App Reviews
    • IMDB
    • Rotten Tomatoes
    • Yelp
  • Summarization
    • CNN Daily Mail
    • Gigaword
    • MultiNews
    • SamSum
    • XSum
  • Topic Classification
    • AG News
    • DBPedia
    • TREC
  • Paraphrase Identification
    • MRPC
    • PAWS
    • QQP
  • Natural Language Inference
    • ANLI
    • CB
    • RTE
  • Coreference Resolution
    • WSC
    • Winogrande
  • Word Sense disambiguation
    • WiC
  • Sentence Completion
    • COPA
    • HellaSwag
    • Story Cloze

Annotations

The prompts available in Promptsource are collected as part of BigScience, one-year long research workshop on large multilingual models and datasets. 36 contributors affiliated with 24 institutions in 8 countries participated to the prompt collection. Contributors are in majority machine learning researchers or machine learning engineers.

The main annotation guideline was that prompts needed to be grammatical and understandable by a native English speaker with no prior experience of the tasks. Additionally, prompts that required explicit counting or numerical indexing were removed in favor of natural language variants, e.g., instead of predicting indices of a span to extract (e.g. in extractive question answering), the model was expected to copy the span's text instead. With these minimal constraints, prompt writers were encouraged to use both formal and creative prompts and various orderings of the data. Most of the prompts correspond directly to a version of the original proposed task, although we also allowed prompts that permuted the original task (for instance, generating a document from its summary) or allowed for ambiguous output (for instance, not indicating a list of available choices).

The full annotation given to the contributors can be found here. *Note to self: the link is currently being updated with the)

Additional Information

Licensing Information

The dataset is released under Apache 2.0.

Citation Information

TODO

Contributions

Thanks to the contributors of promptsource for adding this dataset.