Heng666's picture
Update README.md
afd0445 verified
|
raw
history blame
38.9 kB
metadata
dataset_info:
  - config_name: aya_dataset
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 1703870
        num_examples: 4909
    download_size: 998499
    dataset_size: 1703870
  - config_name: templated_ntx_llm
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 402016
        num_examples: 256
    download_size: 49658
    dataset_size: 402016
  - config_name: templated_uner_llm
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 8259554
        num_examples: 7992
      - name: test
        num_bytes: 2349655
        num_examples: 1997
      - name: validation
        num_bytes: 1073059
        num_examples: 998
    download_size: 2001352
    dataset_size: 11682268
  - config_name: templated_xcsqa
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: validation
        num_bytes: 297261
        num_examples: 1000
    download_size: 104307
    dataset_size: 297261
  - config_name: templated_xlel_wd
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 40044968
        num_examples: 50949
      - name: test
        num_bytes: 5192254
        num_examples: 6637
      - name: validation
        num_bytes: 4999716
        num_examples: 6399
    download_size: 31048776
    dataset_size: 50236938
  - config_name: templated_xwikis
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 147846215
        num_examples: 31276
      - name: test
        num_bytes: 34828372
        num_examples: 6998
      - name: validation
        num_bytes: 65882386
        num_examples: 14998
    download_size: 168472726
    dataset_size: 248556973
  - config_name: translated_adversarial_qa
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 13844756
        num_examples: 20000
      - name: test
        num_bytes: 1505596
        num_examples: 2000
      - name: validation
        num_bytes: 1392335
        num_examples: 2000
    download_size: 6348021
    dataset_size: 16742687
  - config_name: translated_cnn_dailymail
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 284764459
        num_examples: 200000
      - name: test
        num_bytes: 33310796
        num_examples: 22980
      - name: validation
        num_bytes: 39146476
        num_examples: 26736
    download_size: 228280361
    dataset_size: 357221731
  - config_name: translated_dolly
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 21966259
        num_examples: 29616
    download_size: 12062356
    dataset_size: 21966259
  - config_name: translated_flan_coqa
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 26455946
        num_examples: 12818
    download_size: 16080989
    dataset_size: 26455946
  - config_name: translated_flan_cot
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 72401586
        num_examples: 183820
    download_size: 34564552
    dataset_size: 72401586
  - config_name: translated_flan_gem_wiki
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 84073165
        num_examples: 54294
    download_size: 52363999
    dataset_size: 84073165
  - config_name: translated_flan_lambada
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 2236958
        num_examples: 8558
    download_size: 1156298
    dataset_size: 2236958
  - config_name: translated_flan_qa
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 355221
        num_examples: 1080
    download_size: 166753
    dataset_size: 355221
  - config_name: translated_hotpotqa
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 140638090
        num_examples: 710952
    download_size: 51704787
    dataset_size: 140638090
  - config_name: translated_joke_explaination
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 789677
        num_examples: 1508
    download_size: 343907
    dataset_size: 789677
  - config_name: translated_mintaka
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 5661977
        num_examples: 28000
      - name: test
        num_bytes: 1608641
        num_examples: 8000
      - name: validation
        num_bytes: 829827
        num_examples: 4000
    download_size: 1732447
    dataset_size: 8100445
  - config_name: translated_mlqa
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: test
        num_bytes: 36734954
        num_examples: 46360
      - name: validation
        num_bytes: 3563390
        num_examples: 4592
    download_size: 21052230
    dataset_size: 40298344
  - config_name: translated_nqopen
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 56945959
        num_examples: 351700
    download_size: 19578926
    dataset_size: 56945959
  - config_name: translated_paws
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 34768722
        num_examples: 98802
      - name: test
        num_bytes: 5625712
        num_examples: 16000
      - name: validation
        num_bytes: 5721186
        num_examples: 16000
    download_size: 8291033
    dataset_size: 46115620
  - config_name: translated_piqa
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 11873740
        num_examples: 32226
    download_size: 4759498
    dataset_size: 11873740
  - config_name: translated_wikiqa
    features:
      - name: id
        dtype: int64
      - name: inputs
        dtype: string
      - name: targets
        dtype: string
      - name: dataset_name
        dtype: string
      - name: sub_dataset_name
        dtype: string
      - name: task_type
        dtype: string
      - name: template_id
        dtype: string
      - name: language_code
        dtype: string
      - name: split
        dtype: string
      - name: script
        dtype: string
    splits:
      - name: train
        num_bytes: 551856
        num_examples: 2080
      - name: test
        num_bytes: 154223
        num_examples: 586
      - name: validation
        num_bytes: 78316
        num_examples: 280
    download_size: 387555
    dataset_size: 784395
configs:
  - config_name: aya_dataset
    data_files:
      - split: train
        path: aya_dataset/train-*
  - config_name: templated_ntx_llm
    data_files:
      - split: train
        path: templated_ntx_llm/train-*
  - config_name: templated_uner_llm
    data_files:
      - split: train
        path: templated_uner_llm/train-*
      - split: test
        path: templated_uner_llm/test-*
      - split: validation
        path: templated_uner_llm/validation-*
  - config_name: templated_xcsqa
    data_files:
      - split: validation
        path: templated_xcsqa/validation-*
  - config_name: templated_xlel_wd
    data_files:
      - split: train
        path: templated_xlel_wd/train-*
      - split: test
        path: templated_xlel_wd/test-*
      - split: validation
        path: templated_xlel_wd/validation-*
  - config_name: templated_xwikis
    data_files:
      - split: train
        path: templated_xwikis/train-*
      - split: test
        path: templated_xwikis/test-*
      - split: validation
        path: templated_xwikis/validation-*
  - config_name: translated_adversarial_qa
    data_files:
      - split: train
        path: translated_adversarial_qa/train-*
      - split: test
        path: translated_adversarial_qa/test-*
      - split: validation
        path: translated_adversarial_qa/validation-*
  - config_name: translated_cnn_dailymail
    data_files:
      - split: train
        path: translated_cnn_dailymail/train-*
      - split: test
        path: translated_cnn_dailymail/test-*
      - split: validation
        path: translated_cnn_dailymail/validation-*
  - config_name: translated_dolly
    data_files:
      - split: train
        path: translated_dolly/train-*
  - config_name: translated_flan_coqa
    data_files:
      - split: train
        path: translated_flan_coqa/train-*
  - config_name: translated_flan_cot
    data_files:
      - split: train
        path: translated_flan_cot/train-*
  - config_name: translated_flan_gem_wiki
    data_files:
      - split: train
        path: translated_flan_gem_wiki/train-*
  - config_name: translated_flan_lambada
    data_files:
      - split: train
        path: translated_flan_lambada/train-*
  - config_name: translated_flan_qa
    data_files:
      - split: train
        path: translated_flan_qa/train-*
  - config_name: translated_hotpotqa
    data_files:
      - split: train
        path: translated_hotpotqa/train-*
  - config_name: translated_joke_explaination
    data_files:
      - split: train
        path: translated_joke_explaination/train-*
  - config_name: translated_mintaka
    data_files:
      - split: train
        path: translated_mintaka/train-*
      - split: test
        path: translated_mintaka/test-*
      - split: validation
        path: translated_mintaka/validation-*
  - config_name: translated_mlqa
    data_files:
      - split: test
        path: translated_mlqa/test-*
      - split: validation
        path: translated_mlqa/validation-*
  - config_name: translated_nqopen
    data_files:
      - split: train
        path: translated_nqopen/train-*
  - config_name: translated_paws
    data_files:
      - split: train
        path: translated_paws/train-*
      - split: test
        path: translated_paws/test-*
      - split: validation
        path: translated_paws/validation-*
  - config_name: translated_piqa
    data_files:
      - split: train
        path: translated_piqa/train-*
  - config_name: translated_wikiqa
    data_files:
      - split: train
        path: translated_wikiqa/train-*
      - split: test
        path: translated_wikiqa/test-*
      - split: validation
        path: translated_wikiqa/validation-*
license: apache-2.0
task_categories:
  - question-answering
  - translation
  - summarization
  - zero-shot-classification
language:
  - zh
pretty_name: ' Traditional_Chinese-aya_collection'
size_categories:
  - 1M<n<10M

Traditional_Chinese_Aya Header

資料集描述

繁體中文 Aya (Traditional Chinese Aya Chinese;TCA):專注於繁體中文處理的 Aya 集合的精選子集

概述

繁體中文 Aya 是一個精心策劃的資料集,源自 CohereForAI 的綜合 Aya 集合,特別關注繁體中文文本資料。 此資料集結合了來自 CohereForAI/aya_collection,過濾掉除繁體中文、簡體中文內容之外的所有內容。

目標

繁體中文 Aya 的目標是為研究人員、技術專家和語言學家提供即用型繁體中文文本資源,顯著減少專注於繁體中文的 NLP 和 AI 專案中數據預處理所需的時間和精力。

資料集來源與資訊

  • 資料來源: 從 CohereForAI/aya_collection 64 個子集而來。
  • 語言: 繁體中文、簡體中文('zho')
  • 應用: 非常適合語言建模、文本分類、情感分析、和機器翻譯等任務。
  • 論文連結: 2402.06619
  • 維護人: Heng666
  • License: Apache-2.0

使用方法

此資料集是開始繁體中文語言專案(從學術研究到商業應用)的基礎工具。 透過提供預先過濾的繁體中文文本來源,繁體中文 Aya 讓研究人員、技術專家和開發人員能夠直接進行模型訓練、分析和應用程式開發,而無需進行資料清理和語言過濾的初步麻煩。

展示範例

from datasets import load_dataset
dataset = load_dataset("Heng666/Traditional_Chinese-aya_collection", "aya_dataset")

在上面的程式碼片段中,「aya_dataset」指的是原始 「aya_collection」中「aya_dataset」子集的繁體中文版本(100k行)。 您可以透過在載入資料集時指定其名稱來載入其他子集。

訪問和貢獻

可在 Heng666/Traditional_Chinese-aya_collection 下的 Hugging Face Hub 上獲取, 繁體中文 Aya 邀請社區做出貢獻。鼓勵用戶提供回饋、提出改進建議。

支持與合作

我們致力於圍繞繁體中文人工智慧和 NLP 研究創造一個包容和支持的環境。如需支援、協作或有關資料集的疑問,請透過 Hugging Face Hub 的討論部分進行聯絡。

Original Dataset Card of Aya by CohereForAI

Aya Header

Dataset Summary

The Aya Collection is a massive multilingual collection consisting of 513 million instances of prompts and completions covering a wide range of tasks. This collection incorporates instruction-style templates from fluent speakers and applies them to a curated list of datasets, as well as translations of instruction-style datasets into 101 languages. Aya Dataset, a human-curated multilingual instruction and response dataset, is also part of this collection. See our paper for more details regarding the collection.

  • Curated by: Contributors of Aya Open Science Intiative

  • Language(s): 115 languages

  • License: Apache 2.0

  • Aya Datasets Family:

    Name Explanation
    aya_dataset Human-annotated multilingual instruction finetuning dataset, comprising over 204K instances across 65 languages.
    aya_collection Created by applying instruction-style templates from fluent speakers to 44 datasets, including translations of 19 instruction-style datasets into 101 languages.
    aya_evaluation_suite A diverse evaluation set for multilingual open-ended generation, featuring 250 culturally grounded prompts in 7 languages, 200 translated prompts in 24 languages, and human-edited versions selected for cross-cultural relevance from English Dolly in 6 languages.

Dataset

The Aya Collection is a comprehensive, large corpus of datasets that can be used by researchers around the world to train multilingual models. Our goal is only to include datasets with permissive licensing for manipulation and redistribution. The Aya Collection consists of three different sources of data:

  1. Templated data: We collaborated with fluent speakers to create templates that allowed for the automatic expansion of existing datasets into various languages.
  2. Translated data: We translated a hand-selected subset of 19 datasets into 101 languages (114 dialects) using the NLLB 3.3B parameter machine translation model.
  3. Aya Dataset: We release the Aya Dataset as a subset of the overall collection. This is the only dataset in the collection that is human-annotated in its entirety.

Load with Datasets

To load this dataset with Datasets, you'll need to install Datasets as pip install datasets --upgrade and then use the following code:

from datasets import load_dataset
dataset = load_dataset("CohereForAI/aya_collection", "templated_mintaka")

In the above code snippet, "templated_mintaka" refers to a subset of the aya_collection. You can load other subsets by specifying its name at the time of loading the dataset.

Data Instances

An example of a train instance looks as follows:

{'id': 246001,
 'inputs': 'The following query in English is taken from the geography category. What could be the answer to the question?\nWhat is the seventh tallest mountain in North America?',
 'targets': 'The answer is Mount Lucania.',
 'dataset_name': 'Mintaka-inst',
 'sub_dataset_name': '-',
 'task_type': 'question-answering',
 'template_id': 3,
 'language': 'eng',
 'split': 'train',
 'script': 'Latn'
}

Data Fields

The data fields are the same among all splits:

  • id: Unique id of the data point
  • inputs: Prompt or input to the language model.
  • targets: Completion or output of the language model.
  • dataset_name: The name of the source dataset that the data point was taken from
  • sub_dataset_name: If the source is a collection, this field indicates which part of that collection the data point was taken from. If it is not a collection, this field is left blank.
  • task_type: The task type that this conversation belongs to.
  • template_id: The id of the template applied to this data point.
  • language: The ISO code of the dialect of the conversation.
  • script: The script of the language.
  • split: Indicates whether the data point is part of the train or the test split.

Statistics

The total number of data points, including the Aya Dataset` is 513,758,189. To view the breakdown of dialect codes and the respective templated and translated data point counts in the Aya Collection , refer to the toggled table below.

Breakdown of Aya Collection data point counts grouped by dialects |dialect code|language|translated data point count|templated data point count|total count | |------------|--------|---------------------------|--------------------------|---------------| |ace |Achinese|8240684 |2000 |8242684 | |acm |Arabic |4120342 |0 |4120342 | |acq |Arabic |4120342 |0 |4120342 | |aeb |Arabic |4120342 |0 |4120342 | |afr |Afrikaans|4120342 |6108 |4126450 | |ajp |Arabic |4120342 |0 |4120342 | |als |Albanian|4120342 |0 |4120342 | |amh |Amharic |4120342 |25327 |4145669 | |apc |Arabic |4120342 |0 |4120342 | |arb |Arabic |6424999 |216430 |6641429 | |ars |Arabic |4120342 |0 |4120342 | |ary |Arabic |4120342 |18076 |4138418 | |arz |Arabic |4120342 |0 |4120342 | |azb |Azerbaijani|4120342 |0 |4120342 | |azj |Azerbaijani|4120342 |0 |4120342 | |bel |Belarusian|4120342 |21273 |4141615 | |ben |Bengali |4120342 |30661 |4151003 | |bjn |Banjar |8240684 |2000 |8242684 | |bul |Bulgarian|4120342 |37722 |4158064 | |cat |Catalan |4120342 |66900 |4187242 | |ceb |Cebuano |4120342 |0 |4120342 | |ces |Czech |4120342 |179604 |4299946 | |ckb |Kurdish |4120342 |0 |4120342 | |cym |Welsh |4120342 |0 |4120342 | |dan |Danish |4120342 |36310 |4156652 | |deu |German |4120342 |1326722 |5447064 | |ell |Greek |4120342 |40291 |4160633 | |eng |English |9771427 |8066678 |17838105 | |epo |Esperanto|4120342 |0 |4120342 | |est |Estonian|4120342 |0 |4120342 | |eus |Basque |4120342 |0 |4120342 | |fin |Finnish |4120342 |457895 |4578237 | |fra |French |4120342 |835520 |4955862 | |gla |Scottish Gaelic|4120342 |0 |4120342 | |gle |Irish |4120342 |0 |4120342 | |glg |Galician|4120342 |0 |4120342 | |guj |Gujarati|4120342 |2157 |4122499 | |hat |Haitian Creole|4120342 |0 |4120342 | |hau |Hausa |4120342 |51396 |4171738 | |heb |Hebrew |4120342 |103466 |4223808 | |hin |Hindi |4120342 |260387 |4380729 | |hun |Hungarian|4120342 |82039 |4202381 | |hye |Armenian|4120342 |7080 |4127422 | |ibo |Igbo |4120342 |36312 |4156654 | |ind |Indonesian|4120342 |45709 |4166051 | |isl |Icelandic|4120342 |0 |4120342 | |ita |Italian |4120342 |405682 |4526024 | |jav |Javanese|4120342 |829 |4121171 | |jpn |Japanese|4120342 |2693177 |6813519 | |kan |Kannada |4120342 |1156 |4121498 | |kas |Kashmiri|4120342 |0 |4120342 | |kat |Georgian|4120342 |0 |4120342 | |kaz |Kazakh |4120342 |0 |4120342 | |khk |Mongolian|4120342 |0 |4120342 | |khm |Khmer |4120342 |0 |4120342 | |kir |Kyrgyz |4120342 |0 |4120342 | |kmr |Kurdish |4120342 |0 |4120342 | |knc |Kanuri |8240684 |0 |8240684 | |kor |Korean |4120342 |41011 |4161353 | |lao |Lao |4120342 |0 |4120342 | |lit |Lithuanian|4120342 |0 |4120342 | |ltz |Luxembourgish|4120342 |0 |4120342 | |lvs |Latvian |4120342 |0 |4120342 | |mal |Malayalam|4120342 |4347 |4124689 | |mar |Marathi |4120342 |3678 |4124020 | |min |Minangkabau|6753788 |2000 |6755788 | |mkd |Macedonian|4120342 |0 |4120342 | |mlt |Maltese |4120342 |0 |4120342 | |mni |Manipuri|4120342 |0 |4120342 | |mri |Maori |4120342 |0 |4120342 | |mya |Burmese |4120342 |0 |4120342 | |nld |Dutch |4120342 |220181 |4340523 | |nno |Norwegian|4120342 |0 |4120342 | |nob |Norwegian|4120342 |0 |4120342 | |npi |Nepali |4120342 |0 |4120342 | |nso |Northern Sotho|4120342 |0 |4120342 | |pbt |Pashto |4120342 |0 |4120342 | |pes |Persian |4120342 |245520 |4365862 | |plt |Malagasy|4120342 |0 |4120342 | |pol |Polish |4120342 |332503 |4452845 | |por |Portuguese|4120342 |287432 |4407774 | |ron |Romanian|4120342 |36359 |4156701 | |rus |Russian |4120342 |545920 |4666262 | |sin |Sinhala |4120342 |195 |4120537 | |slk |Slovak |4120342 |27845 |4148187 | |slv |Slovenian|4120342 |25731 |4146073 | |smo |Samoan |4120342 |0 |4120342 | |sna |Shona |4120342 |3684 |4124026 | |snd |Sindhi |4120342 |0 |4120342 | |som |Somali |4120342 |2926 |4123268 | |sot |Southern Sotho|4120342 |0 |4120342 | |spa |Spanish |4120342 |379194 |4499536 | |srp |Serbian |4120342 |77124 |4197466 | |sun |Sundanese|4120342 |2208 |4122550 | |swe |Swedish |4120342 |76486 |4196828 | |swh |Swahili |4120342 |12726 |4133068 | |tam |Tamil |4120342 |11462 |4131804 | |taq |Tamasheq|4120342 |0 |4120342 | |tel |Telugu |4120342 |477821 |4598163 | |tgk |Tajik |4120342 |0 |4120342 | |tha |Thai |4120342 |2125180 |6245522 | |tur |Turkish |4120342 |59932 |4180274 | |ukr |Ukrainian|4120342 |189384 |4309726 | |urd |Urdu |4120342 |337739 |4458081 | |uzn |Uzbek |4120342 |0 |4120342 | |vie |Vietnamese|4120342 |42232 |4162574 | |xho |Xhosa |4120342 |2952 |4123294 | |ydd |Yiddish |4120342 |0 |4120342 | |yor |Yoruba |4120342 |4907 |4125249 | |yue |Chinese |4120342 |0 |4120342 | |zho-Hans |Chinese |4120342 |54528 |4174870 | |zho-Hant |Chinese |4120342 |0 |4120342 | |zsm |Malay |4120342 |13950 |4134292 | |zul |Zulu |4120342 |786 |4121128 | |arq |Arabic |0 |6046 |6046 | |ban |Balinese|0 |2000 |2000 | |bbc |Toba Batak|0 |2000 |2000 | |bem |Bemba |0 |776 |776 | |fil |Filipino|0 |220 |220 | |fon |Fon |0 |845 |845 | |hrv |Croatian|0 |9007 |9007 | |kin |Kinyarwanda|0 |11165 |11165 | |lij |Ligurian|0 |6409 |6409 | |mad |Madurese|0 |2000 |2000 | |nij |Ngaju |0 |2000 |2000 | |nor |Norwegian|0 |72352 |72352 | |pan |Punjabi |0 |2156 |2156 | |twi |Twi |0 |10840 |10840 | |wol |Wolof |0 |785 |785 | |zho |Chinese |0 |74972 |74972 | PS: Templated data also includes Mozambican Portuguese, which doesn't have its own ISO language code.

Motivations & Intentions

  • Curation Rationale: Automatic augmentation of existing datasets serves to enhance the available linguistic resources for multiple languages. The list of languages was initially established from mT5 and aligned with the annotators’ language list and NLLB translation model. The datasets were translated directly from English for all languages.

Additional Information

Provenance

  • Methods Used: A combination of crowd-sourced templating and automatic translation was employed to source this dataset.
  • Methodology Details:
    • Source: Existing NLP datasets
    • Dates of Collection: May 2023 - Dec 2023

Dataset Version and Maintenance

  • Maintenance Status: Actively Maintained
  • Version Details:
    • Current version: 1.0
    • Last Update: 02/2024
    • First Release: 02/2024

Authorship

Licensing Information

This dataset can be used for any purpose, whether academic or commercial, under the terms of the Apache 2.0 License.

Citation Information

@misc{singh2024aya,
      title={Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning}, 
      author={Shivalika Singh and Freddie Vargus and Daniel Dsouza and Börje F. Karlsson and Abinaya Mahendiran and Wei-Yin Ko and Herumb Shandilya and Jay Patel and Deividas Mataciunas and Laura OMahony and Mike Zhang and Ramith Hettiarachchi and Joseph Wilson and Marina Machado and Luisa Souza Moura and Dominik Krzemiński and Hakimeh Fadaei and Irem Ergün and Ifeoma Okoh and Aisha Alaagib and Oshan Mudannayake and Zaid Alyafeai and Vu Minh Chien and Sebastian Ruder and Surya Guthikonda and Emad A. Alghamdi and Sebastian Gehrmann and Niklas Muennighoff and Max Bartolo and Julia Kreutzer and Ahmet Üstün and Marzieh Fadaee and Sara Hooker},
      year={2024},
      eprint={2402.06619},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}