flan2021-full / README.md
aslawliet's picture
Update README.md
d862c92 verified
metadata
license: cc-by-4.0
task_categories:
  - text-generation
  - text-classification
  - token-classification
  - question-answering
  - zero-shot-classification
  - translation
  - summarization
language:
  - en
size_categories:
  - 10M<n<100M

Task Name

  • FLAN-2021 -> 70
{
  "ag_news_subset": null,
  "ai2_arc_ARC-Challenge": null,
  "ai2_arc_ARC-Easy": null,
  "aeslc": null,
  "anli_r1": null,
  "anli_r2": null,
  "anli_r3": null, 
  "bool_q": null,
  "cnn_dailymail": null,
  "coqa": null,
  "cosmos_qa": null,
  "definite_pronoun_resolution": null,
  "drop": null,
  "fix_punct": null,
  "gem_common_gen": null,
  "gem_dart": null,
  "gem_e2e_nlg": null,
  "gem_web_nlg_en": null,
  "gem_wiki_lingua_english_en": null,
  "gigaword": null,
  "glue_cola": null,
  "glue_mnli": null,
  "glue_mrpc": null,
  "glue_qnli": null,
  "glue_qqp": null,
  "glue_sst2": null,
  "glue_stsb": null,
  "glue_wnli": null,
  "hellaswag": null,
  "huggingface_xsum": null, 
  "imdb_reviews_plain_text": null,
  "lambada": null,
  "math_dataset_algebra__linear_1d": null,
  "multi_news": null,
  "natural_questions_open": null,
  "newsroom": null,
  "openbookqa": null,
  "opinion_abstracts_idebate": null,
  "opinion_abstracts_rotten_tomatoes": null,
  "para_crawl_enes": null,
  "paws_wiki": null,
  "piqa": null,
  "quac": null,
  "samsum": null,
  "sentiment140": null,
  "snli": null,
  "squad_v1_1": null,
  "squad_v2_0": null, 
  "story_cloze_2016": null,
  "super_glue_cb": null,
  "super_glue_copa": null,
  "super_glue_multirc": null,
  "super_glue_record": null,
  "super_glue_rte": null,
  "super_glue_wic": null,
  "super_glue_wsc_fixed": null,
  "trec": null,
  "trivia_qa_rc": null,
  "true_case": null,
  "unified_qa_science_inst": null,
  "winogrande": null,
  "word_segment": null, 
  "wmt14_translate_fr-en": null,
  "wmt16_translate_cs-en": null,
  "wmt16_translate_de-en": null,
  "wmt16_translate_fi-en": null,
  "wmt16_translate_ro-en": null,
  "wmt16_translate_ru-en": null,
  "wmt16_translate_tr-en": null,
  "yelp_polarity_reviews": null
}
text_classification = [
    "ag_news_subset",
    "glue_cola",
    "glue_sst2",
    "imdb_reviews_plain_text",
    "yelp_polarity_reviews"
]

question_answering = [
    "ai2_arc_ARC-Challenge",
    "ai2_arc_ARC-Easy",
    "bool_q",
    "coqa",
    "cosmos_qa",
    "drop",
    "natural_questions_open",
    "openbookqa",
    "quac",
    "squad_v1_1",
    "squad_v2_0",
    "trivia_qa_rc"
]

text_generation = [
    "aeslc",
    "cnn_dailymail",
    "gem_common_gen",
    "gem_dart",
    "gem_e2e_nlg",
    "gem_web_nlg_en",
    "gem_wiki_lingua_english_en",
    "gigaword",
    "huggingface_xsum",
    "lambada",
    "multi_news",
    "newsroom",
    "samsum"
]

translation = [
    "wmt14_translate_fr-en",
    "wmt16_translate_cs-en",
    "wmt16_translate_de-en",
    "wmt16_translate_fi-en",
    "wmt16_translate_ro-en",
    "wmt16_translate_ru-en",
    "wmt16_translate_tr-en"
]

sentiment_analysis = [
    "sentiment140"
]

textual_entailment = [
    "anli_r1",
    "anli_r2",
    "anli_r3",
    "glue_mnli",
    "glue_rte",
    "snli",
    "super_glue_cb",
    "super_glue_copa",
    "super_glue_rte"
]

paraphrase_detection = [
    "glue_mrpc",
    "glue_qqp",
    "paws_wiki"
]

commonsense_reasoning = [
    "hellaswag",
    "piqa",
    "super_glue_multirc",
    "super_glue_record",
    "super_glue_wic",
    "super_glue_wsc_fixed",
    "winogrande"
]

textual_similarity = [
    "glue_stsb"
]

named_entity_recognition = [
    "glue_wnli"
]

text_correction = [
    "fix_punct",
    "true_case"
]

text_segmentation = [
    "word_segment"
]

argument_mining = [
    "opinion_abstracts_idebate",
    "opinion_abstracts_rotten_tomatoes"
]

machine_reading_comprehension = [
    "glue_qnli"
]

text_summarization = [
    "trec"
]

language_modelling = [
    "story_cloze_2016"
]

math_problem_solving = [
    "math_dataset_algebra__linear_1d",
    "unified_qa_science_inst"
]

cross_lingual_information_retrieval = [
    "para_crawl_enes"
]