diff --git a/README.md b/README.md deleted file mode 100644 index f07851d95a2c56b4de8716949e128502cc26818c..0000000000000000000000000000000000000000 --- a/README.md +++ /dev/null @@ -1,278 +0,0 @@ ---- -annotations_creators: -- expert-generated -language: -- en -- fr -- am -- bm -- bbj -- ee -- fon -- ha -- ig -- lg -- mos -- ny -- pcm -- rw -- sn -- sw -- tn -- tw -- wo -- xh -- yo -- zu -language_creators: -- expert-generated -license: -- cc-by-nc-4.0 -multilinguality: -- translation -- multilingual -pretty_name: mafand -size_categories: -- 1K>> from datasets import load_dataset ->>> data = load_dataset('masakhane/mafand', 'en-yor') - -{"translation": {"src": "President Buhari will determine when to lift lockdown – Minister", "tgt": "Ààrẹ Buhari ló lè yóhùn padà lórí ètò kónílégbélé – Mínísítà"}} - - -{"translation": {"en": "President Buhari will determine when to lift lockdown – Minister", "yo": "Ààrẹ Buhari ló lè yóhùn padà lórí ètò kónílégbélé – Mínísítà"}} -``` - -### Data Fields - -- "translation": name of the task -- "src" : source language e.g en -- "tgt": target language e.g yo - -### Data Splits - -Train/dev/test split - -language| Train| Dev |Test --|-|-|- -amh |-|899|1037 -bam |3302|1484|1600 -bbj |2232|1133|1430 -ewe |2026|1414|1563 -fon |2637|1227|1579 -hau |5865|1300|1500 -ibo |6998|1500|1500 -kin |-|460|1006 -lug |4075|1500|1500 -luo |4262|1500|1500 -mos |2287|1478|1574 -nya |-|483|1004 -pcm |4790|1484|1574 -sna |-|556|1005 -swa |30782|1791|1835 -tsn |2100|1340|1835 -twi |3337|1284|1500 -wol |3360|1506|1500| -xho |-|486|1002| -yor |6644|1544|1558| -zul |3500|1239|998| - - -## Dataset Creation - -### Curation Rationale - -MAFAND was created from the news domain, translated from English or French to an African language - -### Source Data - -#### Initial Data Collection and Normalization - -[Needs More Information] - -#### Who are the source language producers? - -- [Masakhane](https://github.com/masakhane-io/lafand-mt) -- [Igbo](https://github.com/IgnatiusEzeani/IGBONLP/tree/master/ig_en_mt) -- [Swahili](https://opus.nlpl.eu/GlobalVoices.php) -- [Hausa](https://www.statmt.org/wmt21/translation-task.html) -- [Yoruba](https://github.com/uds-lsv/menyo-20k_MT) - -### Annotations - -#### Annotation process - -[Needs More Information] - -#### Who are the annotators? - -Masakhane members - -### Personal and Sensitive Information - -[Needs More Information] - -## Considerations for Using the Data - -### Social Impact of Dataset - -[Needs More Information] - -### Discussion of Biases - -[Needs More Information] - -### Other Known Limitations - -[Needs More Information] - -## Additional Information - -### Dataset Curators - -[Needs More Information] - -### Licensing Information - -[CC-BY-4.0-NC](https://creativecommons.org/licenses/by-nc/4.0/) - -### Citation Information - -``` -@inproceedings{adelani-etal-2022-thousand, - title = "A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation", - author = "Adelani, David and - Alabi, Jesujoba and - Fan, Angela and - Kreutzer, Julia and - Shen, Xiaoyu and - Reid, Machel and - Ruiter, Dana and - Klakow, Dietrich and - Nabende, Peter and - Chang, Ernie and - Gwadabe, Tajuddeen and - Sackey, Freshia and - Dossou, Bonaventure F. P. and - Emezue, Chris and - Leong, Colin and - Beukman, Michael and - Muhammad, Shamsuddeen and - Jarso, Guyo and - Yousuf, Oreen and - Niyongabo Rubungo, Andre and - Hacheme, Gilles and - Wairagala, Eric Peter and - Nasir, Muhammad Umair and - Ajibade, Benjamin and - Ajayi, Tunde and - Gitau, Yvonne and - Abbott, Jade and - Ahmed, Mohamed and - Ochieng, Millicent and - Aremu, Anuoluwapo and - Ogayo, Perez and - Mukiibi, Jonathan and - Ouoba Kabore, Fatoumata and - Kalipe, Godson and - Mbaye, Derguene and - Tapo, Allahsera Auguste and - Memdjokam Koagne, Victoire and - Munkoh-Buabeng, Edwin and - Wagner, Valencia and - Abdulmumin, Idris and - Awokoya, Ayodele and - Buzaaba, Happy and - Sibanda, Blessing and - Bukula, Andiswa and - Manthalu, Sam", - booktitle = "Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", - month = jul, - year = "2022", - address = "Seattle, United States", - publisher = "Association for Computational Linguistics", - url = "https://aclanthology.org/2022.naacl-main.223", - doi = "10.18653/v1/2022.naacl-main.223", - pages = "3053--3070", - abstract = "Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.", -} -``` \ No newline at end of file diff --git a/dataset_infos.json b/dataset_infos.json deleted file mode 100644 index 222d194f45b0b817ece170c1bf030b7ca6f36eda..0000000000000000000000000000000000000000 --- a/dataset_infos.json +++ /dev/null @@ -1 +0,0 @@ -{"amh": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "amh", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 639911, "num_examples": 1750, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 92753, "num_examples": 250, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 184271, "num_examples": 500, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/amh/train.txt": {"num_bytes": 399218, "checksum": "4cd4bb953f2d2a47172f5589133324657c785ed577acc3ebd7d5a74a106b0883"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/amh/dev.txt": {"num_bytes": 58077, "checksum": "0ba1bb30f7519c255341a9438fda19e0f61f15aa344e5ba038eb50144147e21d"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/amh/test.txt": {"num_bytes": 114656, "checksum": "8957bf06668ea28842acaca3caaf139aac1e078b56bf7aec593c1bb8fff7d938"}}, "download_size": 571951, "post_processing_size": null, "dataset_size": 916935, "size_in_bytes": 1488886}, "hau": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "hau", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 929848, "num_examples": 1912, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 139503, "num_examples": 276, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 282971, "num_examples": 552, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/hau/train.txt": {"num_bytes": 436199, "checksum": "f09523f77aacc7d89e935ce8fdd0d7d6afb7ba06e4a6040a431f295b0b937939"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/hau/dev.txt": {"num_bytes": 65392, "checksum": "a4695528b190c629f417e66feda127650611ee8d0f0d0ce14dee8f1506da0f7f"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/hau/test.txt": {"num_bytes": 131781, "checksum": "1f636716cb1633490bc475951721d81a637b033045123f3ee84fee2a1ea06c70"}}, "download_size": 633372, "post_processing_size": null, "dataset_size": 1352322, "size_in_bytes": 1985694}, "ibo": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "ibo", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 749196, "num_examples": 2235, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 110572, "num_examples": 320, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 222192, "num_examples": 638, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/ibo/train.txt": {"num_bytes": 354547, "checksum": "9731a5268749b43c5f1c7ac1f420aa7b856cbe2a0002b14dd1d2f809257612bb"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/ibo/dev.txt": {"num_bytes": 53684, "checksum": "059e81794392a659efd65218ae988ff4ae680cd5beceae84caa5a8658767e8ee"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/ibo/test.txt": {"num_bytes": 107184, "checksum": "ecfc765813ac82b5807e3cf00e26020ec971e5ebd93c71e172c7f48df427d818"}}, "download_size": 515415, "post_processing_size": null, "dataset_size": 1081960, "size_in_bytes": 1597375}, "kin": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "kin", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 878746, "num_examples": 2116, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 120998, "num_examples": 302, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 258638, "num_examples": 605, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/kin/train.txt": {"num_bytes": 442520, "checksum": "46960b853d33b08759ec04dea5ba390ac196573b9fb868556aa3ed9ecd613d5a"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/kin/dev.txt": {"num_bytes": 61150, "checksum": "2adff8ef6a84927a9ebf68760feaffba8ff4d220fde89d6da88e084287c6cae9"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/kin/test.txt": {"num_bytes": 129354, "checksum": "d49c012b50ef2e1e5f6e424f4348d18ebc171298dd0eb34ec5a000a353f3d98f"}}, "download_size": 633024, "post_processing_size": null, "dataset_size": 1258382, "size_in_bytes": 1891406}, "lug": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "lug", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 611917, "num_examples": 1428, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 70058, "num_examples": 200, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 183063, "num_examples": 407, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/lug/train.txt": {"num_bytes": 315679, "checksum": "5884e3c4be037b37dec9c1d60125638dad3510c81f903402d9c31f023deaf7fa"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/lug/dev.txt": {"num_bytes": 35254, "checksum": "654a457c9c7aaba690a6cab0a72390605288fd46668e3c17fa5f84596bbc740c"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/lug/test.txt": {"num_bytes": 94822, "checksum": "6b8b327cecc61d7228be95331e6bf3bbdeea84a3d023172359f0aa1c89b52ecc"}}, "download_size": 445755, "post_processing_size": null, "dataset_size": 865038, "size_in_bytes": 1310793}, "luo": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "luo", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 314995, "num_examples": 644, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 43506, "num_examples": 92, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 87716, "num_examples": 186, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/luo/train.txt": {"num_bytes": 150174, "checksum": "a676dba1a88bfbf352150145855252f7c47c271c37da8c2d09fb3560e899fcd9"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/luo/dev.txt": {"num_bytes": 20777, "checksum": "37a270c722d1c5462907222bf3ff1cf2453075490f00867a4c8d90838080416a"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/luo/test.txt": {"num_bytes": 42330, "checksum": "9342c9a99a17b08daee32fe9ed414ab1e4810db886ba4e8b1c9f5522438479df"}}, "download_size": 213281, "post_processing_size": null, "dataset_size": 446217, "size_in_bytes": 659498}, "pcm": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "pcm", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 868229, "num_examples": 2124, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 126829, "num_examples": 306, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 262185, "num_examples": 600, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/pcm/train.txt": {"num_bytes": 395332, "checksum": "7fbb8bc4f0b456e6283624d05a1bfd0e331c580ebf9564815d6f5b7453588272"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/pcm/dev.txt": {"num_bytes": 57178, "checksum": "34e2dc4868718fbb3646f2d65d3e7802866d4c17bde005ff28bfc3960de2f0bc"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/pcm/test.txt": {"num_bytes": 119544, "checksum": "de0dc15f4f74adff9486d03897354ef30ded9bedb73e058a96e3d1bd0a0c85fc"}}, "download_size": 572054, "post_processing_size": null, "dataset_size": 1257243, "size_in_bytes": 1829297}, "swa": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "swa", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1001120, "num_examples": 2109, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 128563, "num_examples": 300, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 272108, "num_examples": 604, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/swa/train.txt": {"num_bytes": 490917, "checksum": "3fbefb76d97a652a29c6f856946af60e31a92ad82927bb748092c5be1b96cb08"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/swa/dev.txt": {"num_bytes": 63078, "checksum": "ee742c38cba9139f1a147baffe7d294d6e38599c67a66ec896500ac2666f53d3"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/swa/test.txt": {"num_bytes": 132318, "checksum": "f91b97def896572dfaaf6db247a01cc1a8b7bd2ce537c19956ec0a0a1ac5913c"}}, "download_size": 686313, "post_processing_size": null, "dataset_size": 1401791, "size_in_bytes": 2088104}, "wol": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "wol", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 602076, "num_examples": 1871, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 71535, "num_examples": 267, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 191484, "num_examples": 539, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/wol/train.txt": {"num_bytes": 252788, "checksum": "160b946c5944313135cd0893030f9f2e5f305e1195e763124bfffd30eac787e8"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/wol/dev.txt": {"num_bytes": 29058, "checksum": "48517a2ad733938805be4bdd285518bed75f14d62068abed74c7eece550e9630"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/wol/test.txt": {"num_bytes": 82617, "checksum": "4791d2b105f1022c6b9180b7d96d10d4daa864d7f7087f1822b55ddbdd22b931"}}, "download_size": 364463, "post_processing_size": null, "dataset_size": 865095, "size_in_bytes": 1229558}, "yor": {"description": "MasakhaNER is the first large publicly available high-quality dataset for named entity recognition (NER) in ten African languages.\n\nNamed entities are phrases that contain the names of persons, organizations, locations, times and quantities.\n\nExample:\n[PER Wolff] , currently a journalist in [LOC Argentina] , played with [PER Del Bosque] in the final years of the seventies in [ORG Real Madrid] .\nMasakhaNER is a named entity dataset consisting of PER, ORG, LOC, and DATE entities annotated by Masakhane for ten African languages:\n- Amharic\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Nigerian-Pidgin\n- Swahili\n- Wolof\n- Yoruba\n\nThe train/validation/test sets are available for all the ten languages.\n\nFor more details see https://arxiv.org/abs/2103.11811\n", "citation": "@article{Adelani2021MasakhaNERNE,\n title={MasakhaNER: Named Entity Recognition for African Languages},\n author={D. Adelani and Jade Abbott and Graham Neubig and Daniel D'Souza and Julia Kreutzer and Constantine Lignos\n and Chester Palen-Michel and Happy Buzaaba and Shruti Rijhwani and Sebastian Ruder and Stephen Mayhew and\n Israel Abebe Azime and S. Muhammad and Chris C. Emezue and Joyce Nakatumba-Nabende and Perez Ogayo and\n Anuoluwapo Aremu and Catherine Gitau and Derguene Mbaye and J. Alabi and Seid Muhie Yimam and Tajuddeen R. Gwadabe and\n Ignatius Ezeani and Rubungo Andre Niyongabo and Jonathan Mukiibi and V. Otiende and Iroro Orife and Davis David and\n Samba Ngom and Tosin P. Adewumi and Paul Rayson and Mofetoluwa Adeyemi and Gerald Muriuki and Emmanuel Anebi and\n C. Chukwuneke and N. Odu and Eric Peter Wairagala and S. Oyerinde and Clemencia Siro and Tobius Saul Bateesa and\n Temilola Oloyede and Yvonne Wambui and Victor Akinode and Deborah Nabagereka and Maurice Katusiime and\n Ayodele Awokoya and Mouhamadane Mboup and D. Gebreyohannes and Henok Tilaye and Kelechi Nwaike and Degaga Wolde and\n Abdoulaye Faye and Blessing Sibanda and Orevaoghene Ahia and Bonaventure F. P. Dossou and Kelechi Ogueji and\n Thierno Ibrahima Diop and A. Diallo and Adewale Akinfaderin and T. Marengereke and Salomey Osei},\n journal={ArXiv},\n year={2021},\n volume={abs/2103.11811}\n}\n", "homepage": "https://arxiv.org/abs/2103.11811", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-DATE", "I-DATE"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "masakhaner", "config_name": "yor", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1016741, "num_examples": 2171, "dataset_name": "masakhaner"}, "validation": {"name": "validation", "num_bytes": 127415, "num_examples": 305, "dataset_name": "masakhaner"}, "test": {"name": "test", "num_bytes": 359519, "num_examples": 645, "dataset_name": "masakhaner"}}, "download_checksums": {"https://github.com/masakhane-io/masakhane-ner/raw/main/data/yor/train.txt": {"num_bytes": 506380, "checksum": "8baf42a3231ab6f10ce7ed5ed434a5a5447634d1cdf979ac932b2496e3566784"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/yor/dev.txt": {"num_bytes": 62819, "checksum": "6fd8ba4c586606a08669ca56954185092f419ddf9c2c4d950815a1cb03af3832"}, "https://github.com/masakhane-io/masakhane-ner/raw/main/data/yor/test.txt": {"num_bytes": 182311, "checksum": "fbace84a8b7cc0c4078f1600a2408b06b3d182977bb64d55318590cba56c2058"}}, "download_size": 751510, "post_processing_size": null, "dataset_size": 1503675, "size_in_bytes": 2255185}, "en-amh": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "amh"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "amh"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-amh", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"validation": {"name": "validation", "num_bytes": 142259, "num_examples": 899, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 114922, "num_examples": 1037, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-amh/dev.json": {"num_bytes": 386223, "checksum": "9d650265aeaccafeb67296c0020a2984bdbc062221c5a0aa88241e47512b5796"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-amh/test.json": {"num_bytes": 317617, "checksum": "e011f98888b10729c76e9b02f06164e86211396aa32ea459c8c0d013dd5c420a"}}, "download_size": 703840, "post_processing_size": null, "dataset_size": 257181, "size_in_bytes": 961021}, "en-hau": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "hau"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "hau"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-hau", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 878882, "num_examples": 5865, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 198905, "num_examples": 1300, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 220740, "num_examples": 1500, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-hau/train.json": {"num_bytes": 2016183, "checksum": "a9af9184e6de5b1d7a9ec89bc7f2c86add6d6c9b5e11bde7dd9b3ae4bfa01c5a"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-hau/dev.json": {"num_bytes": 462239, "checksum": "10b429752c35eb1fc869800968990b4008d43fff4e4025a0ea65ea7313cfeab2"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-hau/test.json": {"num_bytes": 505781, "checksum": "7f0b44884ab9a22882f3f06c26b43b5fc3ff01b2b33cf3f6090f3861354694f3"}}, "download_size": 2984203, "post_processing_size": null, "dataset_size": 1298527, "size_in_bytes": 4282730}, "en-ibo": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "ibo"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "ibo"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-ibo", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 848487, "num_examples": 6998, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 136650, "num_examples": 1500, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 133708, "num_examples": 1500, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-ibo/train.json": {"num_bytes": 1945470, "checksum": "5632edde056ba170a864da712624c27189efb612895543cd1073acee108605e2"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-ibo/dev.json": {"num_bytes": 327129, "checksum": "a1577df2a3d475aec5f2d7b256577dcef5d5c2de717b7bdd4d8d9f33836daaba"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-ibo/test.json": {"num_bytes": 318470, "checksum": "1f08bd80e43e2961effc422e5d94fe70d63b99ba53fb57a9507952f54ebdadaa"}}, "download_size": 2591069, "post_processing_size": null, "dataset_size": 1118845, "size_in_bytes": 3709914}, "en-kin": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "kin"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "kin"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-kin", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"validation": {"name": "validation", "num_bytes": 98435, "num_examples": 460, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 185010, "num_examples": 1006, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-kin/dev.json": {"num_bytes": 211873, "checksum": "7588202c3b685fdecc48c2669bf5bc06affe2873d20eb63c9c78bf29b7577fb3"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-kin/test.json": {"num_bytes": 402546, "checksum": "878be516c3940d8642a806a6600dfcc21b58b1f9ac3f61a0cb9c445e88987e18"}}, "download_size": 614419, "post_processing_size": null, "dataset_size": 283445, "size_in_bytes": 897864}, "en-lug": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "lug"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "lug"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-lug", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 505991, "num_examples": 4075, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 168539, "num_examples": 1500, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 164940, "num_examples": 1500, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-lug/train.json": {"num_bytes": 1111251, "checksum": "9a44376795fe82cf1cff2f4ca6003bcd9a7ccb64d70290eed0d983d4c55e15fe"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-lug/dev.json": {"num_bytes": 378421, "checksum": "ad0383d5c406409e8ccd61e9a048867d8961e5236993ff67cfdf4d87b346890b"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-lug/test.json": {"num_bytes": 372528, "checksum": "50bc3aa1e6ae49877b6611b999b843745f5a1e53097fd3e6f9f3319fb368ca1a"}}, "download_size": 1862200, "post_processing_size": null, "dataset_size": 839470, "size_in_bytes": 2701670}, "en-nya": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "nya"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "nya"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-nya", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"validation": {"name": "validation", "num_bytes": 104269, "num_examples": 483, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 199485, "num_examples": 1004, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-nya/dev.json": {"num_bytes": 234271, "checksum": "bbe84163a83ccf8c493d773330ab522b6194f85fa5646d538ae70e211183a535"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-nya/test.json": {"num_bytes": 438316, "checksum": "4f3cdfd8e865c452652378dbde33db7c93012a3b064548bd4ca605850b05811f"}}, "download_size": 672587, "post_processing_size": null, "dataset_size": 303754, "size_in_bytes": 976341}, "en-pcm": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "pcm"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "pcm"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-pcm", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1253498, "num_examples": 4790, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 411057, "num_examples": 1484, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 445529, "num_examples": 1564, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-pcm/train.json": {"num_bytes": 1401994, "checksum": "626f201900a9a95d3144b13d55ee7c8afbae70a8d65d73223e89be6a2303551f"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-pcm/dev.json": {"num_bytes": 457055, "checksum": "6c2aef216115d19f26df311bbd3bd4d8eb00d6610e3a9595b2f7854f5dc33b8b"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-pcm/test.json": {"num_bytes": 494006, "checksum": "e57cc6894fdec342c30a86a6df0e40b32e90339e4fabf64d6b85ccc66a714dc8"}}, "download_size": 2353055, "post_processing_size": null, "dataset_size": 2110084, "size_in_bytes": 4463139}, "en-sna": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "sna"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "sna"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-sna", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"validation": {"name": "validation", "num_bytes": 75754, "num_examples": 556, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 141706, "num_examples": 1005, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-sna/dev.json": {"num_bytes": 192626, "checksum": "0675fdfcc1d0a1ffb769004552326c1764cef44165791d8947069bf8fd8a2aac"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-sna/test.json": {"num_bytes": 361833, "checksum": "6ce16925d71d2c025072fdd127ffc91925992d3591eef42a3494b289c0ed36b7"}}, "download_size": 554459, "post_processing_size": null, "dataset_size": 217460, "size_in_bytes": 771919}, "en-swa": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "swa"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "swa"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-swa", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 3698641, "num_examples": 30782, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 229818, "num_examples": 1791, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 244617, "num_examples": 1835, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-swa/train.json": {"num_bytes": 8447937, "checksum": "c0242f1a34c080502a68521fb71c51fd72046f9f6e9424635f4702c3ad95c0ad"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-swa/dev.json": {"num_bytes": 511062, "checksum": "ab53290c9409255d90fb3962f8242d943cffb42823a528839fbf15837ea4f81b"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-swa/test.json": {"num_bytes": 552699, "checksum": "dbea0373caab6f517d9103130839e38fb78e56316c24b4686f38a25800114bd2"}}, "download_size": 9511698, "post_processing_size": null, "dataset_size": 4173076, "size_in_bytes": 13684774}, "en-tsn": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "tsn"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "tsn"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-tsn", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 269740, "num_examples": 2100, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 71281, "num_examples": 540, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 196623, "num_examples": 1500, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-tsn/train.json": {"num_bytes": 629359, "checksum": "6d370e1cc7b64285edffe0747c73d8c5132de88f46955c0c29706d83d2048e2b"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-tsn/dev.json": {"num_bytes": 169021, "checksum": "806184d52362e2e56e0824bd307812671e67c36791e154a005863cbfcb209665"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-tsn/test.json": {"num_bytes": 450522, "checksum": "fa5daa725b44f49491fd4b268032ab3e9736cdf561066dacf4d5358a6b701da5"}}, "download_size": 1248902, "post_processing_size": null, "dataset_size": 537644, "size_in_bytes": 1786546}, "en-twi": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "twi"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "twi"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-twi", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 418381, "num_examples": 3337, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 214606, "num_examples": 1284, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 257776, "num_examples": 1500, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-twi/train.json": {"num_bytes": 916305, "checksum": "8c91cdfd9d4b8fc9b95e56f3f36a5357383762caf11b15a362c5661711e2ae7e"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-twi/dev.json": {"num_bytes": 438911, "checksum": "d513e8ad49ebd28bb96529864bada539bef35ab222d3afd23d345db8b0c65419"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-twi/test.json": {"num_bytes": 536213, "checksum": "853f6c797666d7d6aa27d26b13238ce79f3244d473caca5be6d0bca229453022"}}, "download_size": 1891429, "post_processing_size": null, "dataset_size": 890763, "size_in_bytes": 2782192}, "en-xho": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "xho"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "xho"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-xho", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"validation": {"name": "validation", "num_bytes": 66793, "num_examples": 486, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 136894, "num_examples": 1002, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-xho/dev.json": {"num_bytes": 151023, "checksum": "dd7282362c5aa8bc0e34b559b2ebb8436973d137e31789682ed2139e9a868822"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-xho/test.json": {"num_bytes": 305836, "checksum": "fbee16af922728de4f5099a7a5d69d470562c34aa030101d7fb00a3281456324"}}, "download_size": 456859, "post_processing_size": null, "dataset_size": 203687, "size_in_bytes": 660546}, "en-yor": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "yor"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "yor"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-yor", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 879851, "num_examples": 6644, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 240226, "num_examples": 1544, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 220664, "num_examples": 1558, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-yor/train.json": {"num_bytes": 2320170, "checksum": "fd71596355a45a7eb36895270597769157c178f9a9d23ff62caf7b7f9e5369d0"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-yor/dev.json": {"num_bytes": 628950, "checksum": "43c714c3fe502ac7959c5105db9d44bdeee3bf5e5e302a16a7eb932d4a5685a9"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-yor/test.json": {"num_bytes": 577949, "checksum": "dd9c520c9beb27c2db4990fb94529f18ae245627b37e7be3df2d9627c5471bec"}}, "download_size": 3527069, "post_processing_size": null, "dataset_size": 1340741, "size_in_bytes": 4867810}, "en-zul": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["en", "zul"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "en", "output": "zul"}, "task_templates": null, "builder_name": "mafand", "config_name": "en-zul", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 526615, "num_examples": 3500, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 165136, "num_examples": 1239, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 144763, "num_examples": 998, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-zul/train.json": {"num_bytes": 1192112, "checksum": "ba948c35be642b189a870c07c6d9d47be949b9058cd674bcd1e20712d655c7f5"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-zul/dev.json": {"num_bytes": 397973, "checksum": "8b558001679bf62a391c63ba65e5ecfb83f214a653300c471029ed348a587cde"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/en-zul/test.json": {"num_bytes": 328644, "checksum": "4fe50b6bca98319cc46b469c532f2c15fa2d0a5e9b599703aabbd28c4b106159"}}, "download_size": 1918729, "post_processing_size": null, "dataset_size": 836514, "size_in_bytes": 2755243}, "fr-bam": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["fr", "bam"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "fr", "output": "bam"}, "task_templates": null, "builder_name": "mafand", "config_name": "fr-bam", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 443448, "num_examples": 3013, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 232961, "num_examples": 1500, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 218896, "num_examples": 1500, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-bam/train.json": {"num_bytes": 931245, "checksum": "9df33ceeb5ab8c141792df9829c818261ee0d215998daf5127934eeb0e20b40a"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-bam/dev.json": {"num_bytes": 493456, "checksum": "63fcc13207ab3af36287bea11471d737294aaaa7bdfd94c738e152a5a23a3d0e"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-bam/test.json": {"num_bytes": 469529, "checksum": "6b12d3ad802bc4063df5231fe23b9457642ce172027ab9eafbd21aa119d462f0"}}, "download_size": 1894230, "post_processing_size": null, "dataset_size": 895305, "size_in_bytes": 2789535}, "fr-bbj": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["fr", "bbj"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "fr", "output": "bbj"}, "task_templates": null, "builder_name": "mafand", "config_name": "fr-bbj", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 411629, "num_examples": 2232, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 200234, "num_examples": 1133, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 287713, "num_examples": 1430, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-bbj/train.json": {"num_bytes": 480929, "checksum": "5f5d52a1fb9595fb28e806e4d540a2445dae7ad4b8ae39573caaaf9eb146ecc8"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-bbj/dev.json": {"num_bytes": 235365, "checksum": "7ad65f0b71d5ccdc2a457790075930ab57b85c55e6ad0f70b83aaeb9de67a98b"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-bbj/test.json": {"num_bytes": 332183, "checksum": "db43fb00c0a3a963c25972872ff3a702bb7066e31d2eb8f4145ca311568575c5"}}, "download_size": 1048477, "post_processing_size": null, "dataset_size": 899576, "size_in_bytes": 1948053}, "fr-ewe": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["fr", "ewe"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "fr", "output": "ewe"}, "task_templates": null, "builder_name": "mafand", "config_name": "fr-ewe", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 287677, "num_examples": 2026, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 193047, "num_examples": 1414, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 212914, "num_examples": 1563, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-ewe/train.json": {"num_bytes": 506055, "checksum": "3b24d35151586f5e6ffa50c68378f0b084ac6ebe7a15b5130665837c0b3ee492"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-ewe/dev.json": {"num_bytes": 379574, "checksum": "7d756d1da5e1ac23b3d68ce9a2ee9c1047c4fe462e51635b2fd204c31c57f454"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-ewe/test.json": {"num_bytes": 422559, "checksum": "e58164a8ce2fa1e9fae85a01db848f6f41b199ef1496af16eb163f876151e823"}}, "download_size": 1308188, "post_processing_size": null, "dataset_size": 693638, "size_in_bytes": 2001826}, "fr-fon": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["fr", "fon"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "fr", "output": "fon"}, "task_templates": null, "builder_name": "mafand", "config_name": "fr-fon", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 888908, "num_examples": 2637, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 435346, "num_examples": 1227, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 569141, "num_examples": 1579, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-fon/train.json": {"num_bytes": 972727, "checksum": "5a0a5068a0d60320cfa6621c7543a4533ee8cb4b9efb40efd43e183d79dfbd9d"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-fon/dev.json": {"num_bytes": 473769, "checksum": "f2c0a281ab9b9cb1aa4ca4790261087c56a2276b8a649459d1aa5f40a95e3755"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-fon/test.json": {"num_bytes": 618374, "checksum": "aad1cda5ca84966ffc22d970a2f367665158424b5870c642f9eaaf4046ce19d3"}}, "download_size": 2064870, "post_processing_size": null, "dataset_size": 1893395, "size_in_bytes": 3958265}, "fr-mos": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["fr", "mos"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "fr", "output": "mos"}, "task_templates": null, "builder_name": "mafand", "config_name": "fr-mos", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 732837, "num_examples": 2287, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 489672, "num_examples": 1478, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 417424, "num_examples": 1574, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-mos/train.json": {"num_bytes": 803909, "checksum": "a9eda967fdeb0026873bcdb2e8fc2c3c3d6ac693d9e86e41dc362915067a25bf"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-mos/dev.json": {"num_bytes": 535525, "checksum": "85a51897b535ba999b0c6b56b8cb602294485071960cad2557ee186109bef5dc"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-mos/test.json": {"num_bytes": 466305, "checksum": "ec5607df386fb26ac5060a65f3f6968cb89e0f123bc468c97f2cb23d8f0db6f1"}}, "download_size": 1805739, "post_processing_size": null, "dataset_size": 1639933, "size_in_bytes": 3445672}, "fr-wol": {"description": "MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: \n- Amharic\n- Bambara\n- Ghomala\n- Ewe\n- Fon\n- Hausa\n- Igbo\n- Kinyarwanda\n- Luganda\n- Luo\n- Mossi\n- Nigerian-Pidgin\n- Chichewa\n- Shona\n- Swahili\n- Setswana\n- Twi\n- Wolof\n- Xhosa\n- Yoruba\n- Zulu\n\nThe train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho\n\nFor more details see https://aclanthology.org/2022.naacl-main.223/\n", "citation": "@inproceedings{adelani-etal-2022-thousand,\n title = \"A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation\",\n author = \"Adelani, David and\n Alabi, Jesujoba and\n Fan, Angela and\n Kreutzer, Julia and\n Shen, Xiaoyu and\n Reid, Machel and\n Ruiter, Dana and\n Klakow, Dietrich and\n Nabende, Peter and\n Chang, Ernie and\n Gwadabe, Tajuddeen and\n Sackey, Freshia and\n Dossou, Bonaventure F. P. and\n Emezue, Chris and\n Leong, Colin and\n Beukman, Michael and\n Muhammad, Shamsuddeen and\n Jarso, Guyo and\n Yousuf, Oreen and\n Niyongabo Rubungo, Andre and\n Hacheme, Gilles and\n Wairagala, Eric Peter and\n Nasir, Muhammad Umair and\n Ajibade, Benjamin and\n Ajayi, Tunde and\n Gitau, Yvonne and\n Abbott, Jade and\n Ahmed, Mohamed and\n Ochieng, Millicent and\n Aremu, Anuoluwapo and\n Ogayo, Perez and\n Mukiibi, Jonathan and\n Ouoba Kabore, Fatoumata and\n Kalipe, Godson and\n Mbaye, Derguene and\n Tapo, Allahsera Auguste and\n Memdjokam Koagne, Victoire and\n Munkoh-Buabeng, Edwin and\n Wagner, Valencia and\n Abdulmumin, Idris and\n Awokoya, Ayodele and\n Buzaaba, Happy and\n Sibanda, Blessing and\n Bukula, Andiswa and\n Manthalu, Sam\",\n booktitle = \"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jul,\n year = \"2022\",\n address = \"Seattle, United States\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2022.naacl-main.223\",\n doi = \"10.18653/v1/2022.naacl-main.223\",\n pages = \"3053--3070\",\n abstract = \"Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.\",\n}\n", "homepage": "https://github.com/masakhane-io/lafand-mt", "license": "", "features": {"translation": {"languages": ["fr", "wol"], "id": null, "_type": "Translation"}}, "post_processed": null, "supervised_keys": {"input": "fr", "output": "wol"}, "task_templates": null, "builder_name": "mafand", "config_name": "fr-wol", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 524623, "num_examples": 3360, "dataset_name": "mafand"}, "validation": {"name": "validation", "num_bytes": 239103, "num_examples": 1506, "dataset_name": "mafand"}, "test": {"name": "test", "num_bytes": 239236, "num_examples": 1500, "dataset_name": "mafand"}}, "download_checksums": {"https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-wol/train.json": {"num_bytes": 1053446, "checksum": "2d3ee90a53bbdea73a8fe9c95b1219918d88cd37b74efe868b866ca532ea7bb3"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-wol/dev.json": {"num_bytes": 476674, "checksum": "7f5ddc9f17fe05769e9a0f1c469e92652140106aa448a4536fa6916a6bcbe8d8"}, "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/fr-wol/test.json": {"num_bytes": 492765, "checksum": "4f84711bddc1ae6ad91ffc925104cec51e10de1e5cea6d57b448613f69e57fbe"}}, "download_size": 2022885, "post_processing_size": null, "dataset_size": 1002962, "size_in_bytes": 3025847}} \ No newline at end of file diff --git a/en-amh/mafand-test.parquet b/en-amh/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..163747807c9b49ffc1fdb133264dc28734bc1692 --- /dev/null +++ b/en-amh/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39cfb6440580ac77dc8473bf65e3ffc3930421f65e4d582c198e858b1d118249 +size 166128 diff --git a/en-amh/mafand-validation.parquet b/en-amh/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e309f3bde5abdf5a81dfd23c212260df3de7e071 --- /dev/null +++ b/en-amh/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec3e8c9d4a3e6b065ed083b982c10fb4e18634647ba6084fb7c7846a10a33c6e +size 198088 diff --git a/en-hau/mafand-test.parquet b/en-hau/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ba6e9b892292a7f8c44dd7fc25072df596e667e8 --- /dev/null +++ b/en-hau/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c20d8144b3d3120cf780906305c61f01eb3757e8d02523589001f62d6315eac6 +size 246998 diff --git a/en-hau/mafand-train.parquet b/en-hau/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..491d56bd56a106d305e843161a40c6c6b316bc47 --- /dev/null +++ b/en-hau/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ece59799141e57e81f5e43d0a42d1fea2d9cf4907aae6584d3c076cf74f5318 +size 1050251 diff --git a/en-hau/mafand-validation.parquet b/en-hau/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..6aada25789faa97dd8a394c089648598b53526b6 --- /dev/null +++ b/en-hau/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62ff2a9ff475ecc11696319e6171648350257fcd8b8524f8d8e5784a99b225f0 +size 233226 diff --git a/en-ibo/mafand-test.parquet b/en-ibo/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..c003190dc5d57585289724fb2411ebef2b5b67c1 --- /dev/null +++ b/en-ibo/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1fd44c7a1d3f8e95a91ef4e10908211d42c159cc09cacd1345144c723c0560f9 +size 156063 diff --git a/en-ibo/mafand-train.parquet b/en-ibo/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..d66ceb5f2228638ac292e8130b4d6db7a1ef0588 --- /dev/null +++ b/en-ibo/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da567777089ba50bc6966498016caace0eae70c2d837bd98850453f63bae02df +size 960037 diff --git a/en-ibo/mafand-validation.parquet b/en-ibo/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..c4e37f84a59b21c63e52e9b2c6808292226e2ea8 --- /dev/null +++ b/en-ibo/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:557acf994cab8d45b0e74818781e3336ee3ad9323a4766ca184add73a32609b7 +size 162876 diff --git a/en-kin/mafand-test.parquet b/en-kin/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..12973c896a92bd14f4b4b91bf364b49fabc5bb52 --- /dev/null +++ b/en-kin/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94c17c61cf9af51c57edfe90fe13cf631ca31c71a4e2172943228eed6b2d0b26 +size 225887 diff --git a/en-kin/mafand-validation.parquet b/en-kin/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ed7eb9aec6193460580439a33b2c5987982987e2 --- /dev/null +++ b/en-kin/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86afa742f7b506b7ea48f520bb4ecef28fd27f0059070fbe5083e2d4d798a32e +size 122327 diff --git a/en-lug/mafand-test.parquet b/en-lug/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..cd6704cd3d96d88e6ac5a2e8623a1638b3f4eeb6 --- /dev/null +++ b/en-lug/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c60236c62d667669e197cd1e6c9323625449cfc40d88356fea624f58de2b1e75 +size 198254 diff --git a/en-lug/mafand-train.parquet b/en-lug/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..3a66768cd30bdabff756473e260401f0cf70d784 --- /dev/null +++ b/en-lug/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91af20c144ad8913254a614984c4501ab4d0a97fa6f58729cfca2e869cd2a0c4 +size 597005 diff --git a/en-lug/mafand-validation.parquet b/en-lug/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..9e5394eb19c39af54a4bd28d04e0d46abd78afd8 --- /dev/null +++ b/en-lug/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8aa45bd32e439a0a914d20510edd29d8bd84d245cb911824457d1fc273cdc51d +size 190820 diff --git a/en-nya/mafand-test.parquet b/en-nya/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..026117ef77fce792cac98d9d700f4857e6346367 --- /dev/null +++ b/en-nya/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd36f89eac3640e6aac1cf4eb088a87433db6d01b65a00c9160b251c807d0f4f +size 238943 diff --git a/en-nya/mafand-validation.parquet b/en-nya/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..9ae6b2ee7a2f90452b08778e5505ee38512e99f4 --- /dev/null +++ b/en-nya/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d780a512ab9cc8a21dd0477d61dbaf266119a5943415f33c58fee0b0f1867d44 +size 132874 diff --git a/en-pcm/mafand-test.parquet b/en-pcm/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..f1263bb3ad60f321bde8665f17f935793906e25f --- /dev/null +++ b/en-pcm/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:644f9f2181c1d81aaacdf5f4d26b300363597609fda810e3ac41292af1b2a8ce +size 277958 diff --git a/en-pcm/mafand-train.parquet b/en-pcm/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8f967342a528f8eaa6106437656e749717b511d2 --- /dev/null +++ b/en-pcm/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6857c593547bb9446cfeb35b7c813f6fe8374afef60b8d8058757409b8ec6c8 +size 790467 diff --git a/en-pcm/mafand-validation.parquet b/en-pcm/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..bad8b2489c7bf106c54f3cec7d1dfb6d70d89bf5 --- /dev/null +++ b/en-pcm/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:447a3535467f8c6af1d8c8ac68dd035e4a62c47e7af300bfcf5402469a8054b4 +size 254002 diff --git a/en-sna/mafand-test.parquet b/en-sna/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..a9bc26beca266dbf64e2ca251cee2ed47d013960 --- /dev/null +++ b/en-sna/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37b64f6743593886c025de0c2dfa384e64913ce0ac5af61f54ef071d3590ae3d +size 200410 diff --git a/en-sna/mafand-validation.parquet b/en-sna/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..75f1d9fd7dae3300faf196426b0e92b2ca9b8a4c --- /dev/null +++ b/en-sna/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31e9db64a3f66041e1a8068fc95123389ebafbf44b23197af6bba7a9ba374855 +size 106533 diff --git a/en-swa/mafand-test.parquet b/en-swa/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..78fcc91513d723beff5fd1e37f96a3d72f0465f4 --- /dev/null +++ b/en-swa/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0f140d5ae694501db8f63e046ea0009d3b4f2398627288a5ff58e5d86af201b +size 302868 diff --git a/en-swa/mafand-train.parquet b/en-swa/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..aec3f6f06c979eec626390b60fc0f67c3a3537bb --- /dev/null +++ b/en-swa/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8becc3730b07539bb79bc786413d93e2b088a694b219719e667fb654ae40f5c3 +size 4785192 diff --git a/en-swa/mafand-validation.parquet b/en-swa/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..29443ac7488a0363a9caa89052b852252b1cf291 --- /dev/null +++ b/en-swa/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d72be1877ede0311a77a97f0cdf24c4629986db1072f30ce96cf438c9835c667 +size 282762 diff --git a/en-tsn/mafand-test.parquet b/en-tsn/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..39a71aab669cfc3f791560938ac1ea05826746f0 --- /dev/null +++ b/en-tsn/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47c223b57f248383874fd5307621514bbe42a2f917639285471efbca81c7961a +size 234763 diff --git a/en-tsn/mafand-train.parquet b/en-tsn/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..2ddc3bfeee5ae3231a9d12f3829da1c067b7ee14 --- /dev/null +++ b/en-tsn/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f90d5fcd8d184d56d077105f3a4d3db5b7726f833f33c2e93c9846cc9d1408db +size 332430 diff --git a/en-tsn/mafand-validation.parquet b/en-tsn/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..50583337de3ee832707332d65237ae0583a13c1c --- /dev/null +++ b/en-tsn/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90d682280e35d0eaf73ce8d707a82663d865c9076be7f63773ff282c8a9a6feb +size 218229 diff --git a/en-twi/mafand-test.parquet b/en-twi/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..715f83c0a1b59102d651a07227a1c96788979aae --- /dev/null +++ b/en-twi/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9a9efcb198fae2096c8bfe46609a96e08ef8febf83b71ee7a76fed5de5f5cd5 +size 284946 diff --git a/en-twi/mafand-train.parquet b/en-twi/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..f9cf1d3239bcd3ffd64b0af362d196dde0e9edbb --- /dev/null +++ b/en-twi/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb12e7b18b3dde7cfe50f8227f30ff9dbece23e528e3939acf3a78b44ec3175a +size 535501 diff --git a/en-twi/mafand-validation.parquet b/en-twi/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..acaae45cc145a9d25ce60ec4d41a0daeea828cde --- /dev/null +++ b/en-twi/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92bdacbdf9d59999093215b79cf18a19345aaa18d48dea55b668d6f1b1391a74 +size 225137 diff --git a/en-xho/mafand-test.parquet b/en-xho/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..b8d7b9beb38d8e4dcbf7f382a38a8fb1827a8709 --- /dev/null +++ b/en-xho/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7624f30d8791de62bf9f5d3a639f524afa0b8ab8931ef4db5d635e56e7b8de7d +size 166865 diff --git a/en-xho/mafand-validation.parquet b/en-xho/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..68e1c3734af8c08f58cea9d43a65c7bea289b6a3 --- /dev/null +++ b/en-xho/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84d52c64a435a92993ec8f8a9df9ba0a4f2e793f25498de8169cf0451c206f98 +size 83492 diff --git a/en-yor/mafand-test.parquet b/en-yor/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0609626a9123fc4ff401b83bc9e5ba7c720fce9a --- /dev/null +++ b/en-yor/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4f2a9ed83a8f682155dfb3179ffb47cd04dca3de585dea029c52f481dae5492 +size 290381 diff --git a/en-yor/mafand-train.parquet b/en-yor/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..6bb3b0a4b43f5b5d202fde756d95e51497886a46 --- /dev/null +++ b/en-yor/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bd6d8bd5c01761c9058cc6bfdcd9dd909f62c87ba9c09c28e209fc50909d075 +size 1134705 diff --git a/en-yor/mafand-validation.parquet b/en-yor/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0442a58dc8e390a3d540a9403e9734f06262d245 --- /dev/null +++ b/en-yor/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd1ca14e2da55df99a3d9115668f921502f511f345e020658b2f575559e52a0a +size 303701 diff --git a/en-zul/mafand-test.parquet b/en-zul/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..1093e41f479b7d361137e16e01ea9bbebe2831fa --- /dev/null +++ b/en-zul/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6776f09189b15c3dc13f288940531756ffeaf479535f4aa29dff6415f145b71 +size 216521 diff --git a/en-zul/mafand-train.parquet b/en-zul/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..7bd8e1e70c616b438c079901d00f49de8b1d0e27 --- /dev/null +++ b/en-zul/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b580d3f5e3a4845ee6a578609e1f82ecd9801c7fdfcf3fd673da0b21027e380 +size 787940 diff --git a/en-zul/mafand-validation.parquet b/en-zul/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..2a3b1b023326a3ec8f1d911ceca4653fab1eb665 --- /dev/null +++ b/en-zul/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecb142bc8e269d74986175129317b0945871719428abbc2df9731146a5fadc14 +size 261479 diff --git a/fr-bam/mafand-test.parquet b/fr-bam/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..28e4a2bcd754813a51eb4710534e41523e8563ba --- /dev/null +++ b/fr-bam/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6258df3473f30fc005bcae87dcecc88c32465b6e3ae1df7d38bb35b72c207b3 +size 254208 diff --git a/fr-bam/mafand-train.parquet b/fr-bam/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e4606a7a6c3b0ab50eac742cbf37e8e386066337 --- /dev/null +++ b/fr-bam/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4936f59edbc53ede6506f8f64150aaba8ecc8c226549b3f2b8038b400f60f4f6 +size 501094 diff --git a/fr-bam/mafand-validation.parquet b/fr-bam/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..ecfe7e4926b3572118189d85c24a019bd2b5d8c0 --- /dev/null +++ b/fr-bam/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9817f1654db042acfc49320f11944067038cab3ae4851e194ce05a766982753f +size 261495 diff --git a/fr-bbj/mafand-test.parquet b/fr-bbj/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..84452aefe1c9daf96650a6f5f65cf70d73bfb9bf --- /dev/null +++ b/fr-bbj/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a4659049cfe2b911d257e219ad5d41d567db14028cdd6a2c74960b71ccd0777 +size 175446 diff --git a/fr-bbj/mafand-train.parquet b/fr-bbj/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..4b8669707a1393723f110dc20df128da12658fe0 --- /dev/null +++ b/fr-bbj/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e73affb207face01be09ab3f0a2d4ca78f95bfefe23ab6ec84beb8589a9d552 +size 239558 diff --git a/fr-bbj/mafand-validation.parquet b/fr-bbj/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..8c830ae80b531f80fa1525219342aaa937c0d3ee --- /dev/null +++ b/fr-bbj/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a9fa18a6c57ae0b4ee4c919f1f645ce5db4bc4c2248684fa020f762c7b3b359 +size 123600 diff --git a/fr-ewe/mafand-test.parquet b/fr-ewe/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..88101ec95efd3f54dd783055bf3979ed12210247 --- /dev/null +++ b/fr-ewe/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ab41a05844e5ad7fcd3ac2c6ad4a09234657347f3170b04777d2ae466538cd7 +size 216737 diff --git a/fr-ewe/mafand-train.parquet b/fr-ewe/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..cc37cef9503454327bb1940c16f8f91767016c07 --- /dev/null +++ b/fr-ewe/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:057834e03780c1c8bcec0e2a6de81eea7113be74951a9eed4983f3929bed19aa +size 292852 diff --git a/fr-ewe/mafand-validation.parquet b/fr-ewe/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..32ed20d4d55dbe69ddefbf85b3f38346f65f8dd1 --- /dev/null +++ b/fr-ewe/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:683e12d94283078ed497c072bd161fb123ad4774f8a7257a65afb46dfc8780d8 +size 188066 diff --git a/fr-fon/mafand-test.parquet b/fr-fon/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..0f71e60b889ba23026ba3f422a447dc18f107b12 --- /dev/null +++ b/fr-fon/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5df504bf6130381a2d9627507694acff0776df407b22445e2875b4bbdbb3c88a +size 298816 diff --git a/fr-fon/mafand-train.parquet b/fr-fon/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..5ee63ceae29f503e62fd0b21071fb7a0c9a233ab --- /dev/null +++ b/fr-fon/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd09f8990a02a6b1b49885516f1cd19dd7dc33b6db452b66b97e019d33d2a2c4 +size 507841 diff --git a/fr-fon/mafand-validation.parquet b/fr-fon/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..a7565dea8b31d842904fc8de30f18e928b00dbed --- /dev/null +++ b/fr-fon/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93a988944cb46d3d28a7b7c66b9e8198cdb96d509d7e0d55423cde10ea5b4090 +size 249666 diff --git a/fr-mos/mafand-test.parquet b/fr-mos/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..881eebd1bff547d272682bc4d088729547b63b84 --- /dev/null +++ b/fr-mos/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6a1b0ce37736fec7059eb69145b7bca06b3b3861f3182cf9ffca56c2329a775 +size 247699 diff --git a/fr-mos/mafand-train.parquet b/fr-mos/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..b83f4b5e623bcf4b52c9c20325774ee6ba00035b --- /dev/null +++ b/fr-mos/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80e0639b822a38b5aa436432c571f05d3642c85ada999cc94cfc7856e748301f +size 457576 diff --git a/fr-mos/mafand-validation.parquet b/fr-mos/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..7e233f4b780e6969571df56bf48679e45099cc1e --- /dev/null +++ b/fr-mos/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5e7e8966e13efb7bff125c395b36f0ab6e274b5225480344cfa4fa07a0b5420 +size 252141 diff --git a/fr-wol/mafand-test.parquet b/fr-wol/mafand-test.parquet new file mode 100644 index 0000000000000000000000000000000000000000..68c05fe3212ef2c2a7dede339503489e8f310b8d --- /dev/null +++ b/fr-wol/mafand-test.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65abfc6e7cf536acd36c9fef98ea51c4b85f9c640e49e119d0b85fbcbd7acd88 +size 303589 diff --git a/fr-wol/mafand-train.parquet b/fr-wol/mafand-train.parquet new file mode 100644 index 0000000000000000000000000000000000000000..e9ab77d49a1b29e9f58573c2ba3d937513f96ba9 --- /dev/null +++ b/fr-wol/mafand-train.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c645dae68c48cb090a197e281be6fc39fc8914e93e4a8e06d8bb2aae12df5778 +size 645569 diff --git a/fr-wol/mafand-validation.parquet b/fr-wol/mafand-validation.parquet new file mode 100644 index 0000000000000000000000000000000000000000..17916b29886b01be5464cd88a181a4cb8a4acc0d --- /dev/null +++ b/fr-wol/mafand-validation.parquet @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1ab829af0a2bf5173eab498521e6988b9074d6e9ccde11ca586120f86db1562 +size 293882 diff --git a/mafand.py b/mafand.py deleted file mode 100644 index 8e527ab2355262a8f42ffbd29316b2cc8c872cfb..0000000000000000000000000000000000000000 --- a/mafand.py +++ /dev/null @@ -1,224 +0,0 @@ -# coding=utf-8 -# Copyright 2020 HuggingFace Datasets Authors. -# -# Licensed under the Apache License, Version 2.0 (the "License"); -# you may not use this file except in compliance with the License. -# You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. - -# Lint as: python3 -"""MAFAND-MT: Masakhane Anglo and Franco Africa News Dataset for Machine Translation""" - -import datasets -import json - -logger = datasets.logging.get_logger(__name__) - -_CITATION = """\ -@inproceedings{adelani-etal-2022-thousand, - title = "A Few Thousand Translations Go a Long Way! Leveraging Pre-trained Models for {A}frican News Translation", - author = "Adelani, David and - Alabi, Jesujoba and - Fan, Angela and - Kreutzer, Julia and - Shen, Xiaoyu and - Reid, Machel and - Ruiter, Dana and - Klakow, Dietrich and - Nabende, Peter and - Chang, Ernie and - Gwadabe, Tajuddeen and - Sackey, Freshia and - Dossou, Bonaventure F. P. and - Emezue, Chris and - Leong, Colin and - Beukman, Michael and - Muhammad, Shamsuddeen and - Jarso, Guyo and - Yousuf, Oreen and - Niyongabo Rubungo, Andre and - Hacheme, Gilles and - Wairagala, Eric Peter and - Nasir, Muhammad Umair and - Ajibade, Benjamin and - Ajayi, Tunde and - Gitau, Yvonne and - Abbott, Jade and - Ahmed, Mohamed and - Ochieng, Millicent and - Aremu, Anuoluwapo and - Ogayo, Perez and - Mukiibi, Jonathan and - Ouoba Kabore, Fatoumata and - Kalipe, Godson and - Mbaye, Derguene and - Tapo, Allahsera Auguste and - Memdjokam Koagne, Victoire and - Munkoh-Buabeng, Edwin and - Wagner, Valencia and - Abdulmumin, Idris and - Awokoya, Ayodele and - Buzaaba, Happy and - Sibanda, Blessing and - Bukula, Andiswa and - Manthalu, Sam", - booktitle = "Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", - month = jul, - year = "2022", - address = "Seattle, United States", - publisher = "Association for Computational Linguistics", - url = "https://aclanthology.org/2022.naacl-main.223", - doi = "10.18653/v1/2022.naacl-main.223", - pages = "3053--3070", - abstract = "Recent advances in the pre-training for language models leverage large-scale datasets to create multilingual models. However, low-resource languages are mostly left out in these datasets. This is primarily because many widely spoken languages that are not well represented on the web and therefore excluded from the large-scale crawls for datasets. Furthermore, downstream users of these models are restricted to the selection of languages originally chosen for pre-training. This work investigates how to optimally leverage existing pre-trained models to create low-resource translation systems for 16 African languages. We focus on two questions: 1) How can pre-trained models be used for languages not included in the initial pretraining? and 2) How can the resulting translation models effectively transfer to new domains? To answer these questions, we create a novel African news corpus covering 16 languages, of which eight languages are not part of any existing evaluation dataset. We demonstrate that the most effective strategy for transferring both additional languages and additional domains is to leverage small quantities of high-quality translation data to fine-tune large pre-trained models.", -} -""" - -_DESCRIPTION = """\ -MAFAND-MT is the largest MT benchmark for African languages in the news domain, covering 21 languages. The languages covered are: -- Amharic -- Bambara -- Ghomala -- Ewe -- Fon -- Hausa -- Igbo -- Kinyarwanda -- Luganda -- Luo -- Mossi -- Nigerian-Pidgin -- Chichewa -- Shona -- Swahili -- Setswana -- Twi -- Wolof -- Xhosa -- Yoruba -- Zulu - -The train/validation/test sets are available for 16 languages, and validation/test set for amh, kin, nya, sna, and xho - -For more details see https://aclanthology.org/2022.naacl-main.223/ -""" - -_URL = "https://raw.githubusercontent.com/masakhane-io/lafand-mt/main/data/json_files/" -_TRAINING_FILE = "train.json" -_DEV_FILE = "dev.json" -_TEST_FILE = "test.json" - - -class MafandConfig(datasets.BuilderConfig): - """BuilderConfig for Mafand""" - - def __init__(self, **kwargs): - """BuilderConfig for Masakhaner. - Args: - **kwargs: keyword arguments forwarded to super. - """ - super(MafandConfig, self).__init__(**kwargs) - - -class Mafand(datasets.GeneratorBasedBuilder): - """Mafand dataset.""" - BUILDER_CONFIGS = [ - MafandConfig(name="en-amh", version=datasets.Version("1.0.0"), - description="Mafand English-Amharic dataset"), - MafandConfig(name="en-hau", version=datasets.Version("1.0.0"), - description="Mafand English-Hausa dataset"), - MafandConfig(name="en-ibo", version=datasets.Version("1.0.0"), - description="Mafand English-Igbo dataset"), - MafandConfig(name="en-kin", version=datasets.Version("1.0.0"), - description="Mafand English-Kinyarwanda dataset"), - MafandConfig(name="en-lug", version=datasets.Version("1.0.0"), - description="Mafand English-Luganda dataset"), - MafandConfig(name="en-nya", version=datasets.Version("1.0.0"), - description="Mafand English-Chichewa dataset"), - MafandConfig(name="en-pcm", version=datasets.Version("1.0.0"), - description="Mafand English-Naija dataset"), - MafandConfig(name="en-sna", version=datasets.Version("1.0.0"), - description="Mafand English-Shona dataset"), - MafandConfig(name="en-swa", version=datasets.Version("1.0.0"), - description="Mafand English-Swahili dataset"), - MafandConfig(name="en-tsn", version=datasets.Version("1.0.0"), - description="Mafand English-Setswana dataset"), - MafandConfig(name="en-twi", version=datasets.Version("1.0.0"), - description="Mafand English-Twi dataset"), - MafandConfig(name="en-xho", version=datasets.Version("1.0.0"), - description="Mafand English-Xhosa dataset"), - MafandConfig(name="en-yor", version=datasets.Version("1.0.0"), - description="Mafand English-Yoruba dataset"), - MafandConfig(name="en-zul", version=datasets.Version("1.0.0"), - description="Mafand English-Zulu dataset"), - MafandConfig(name="fr-bam", version=datasets.Version("1.0.0"), - description="Mafand French-Bambara dataset"), - MafandConfig(name="fr-bbj", version=datasets.Version("1.0.0"), - description="Mafand French-Ghomala dataset"), - MafandConfig(name="fr-ewe", version=datasets.Version("1.0.0"), - description="Mafand French-Ewe dataset"), - MafandConfig(name="fr-fon", version=datasets.Version("1.0.0"), - description="Mafand French-Fon dataset"), - MafandConfig(name="fr-mos", version=datasets.Version("1.0.0"), - description="Mafand French-Mossi dataset"), - MafandConfig(name="fr-wol", version=datasets.Version("1.0.0"), - description="Mafand French-Wolof dataset"), - ] - - def _info(self): - source, target = self.config.name.split('-') - return datasets.DatasetInfo( - description=_DESCRIPTION, - features=datasets.Features({"translation": datasets.features.Translation(languages=(source, target))}), - supervised_keys=(source, target), - homepage="https://github.com/masakhane-io/lafand-mt", - citation=_CITATION, - ) - - def _split_generators(self, dl_manager): - """Returns SplitGenerators.""" - source, target = self.config.name.split('-') - if target in ['amh', 'kin', 'nya', 'sna', 'xho']: - urls_to_download = { - "dev": f"{_URL}{self.config.name}/{_DEV_FILE}", - "test": f"{_URL}{self.config.name}/{_TEST_FILE}", - } - downloaded_files = dl_manager.download_and_extract(urls_to_download) - return [ - datasets.SplitGenerator(name=datasets.Split.VALIDATION, - gen_kwargs={"filepath": downloaded_files["dev"]}), - datasets.SplitGenerator(name=datasets.Split.TEST, - gen_kwargs={"filepath": downloaded_files["test"]}), - ] - else: - urls_to_download = { - "train": f"{_URL}{self.config.name}/{_TRAINING_FILE}", - "dev": f"{_URL}{self.config.name}/{_DEV_FILE}", - "test": f"{_URL}{self.config.name}/{_TEST_FILE}", - } - downloaded_files = dl_manager.download_and_extract(urls_to_download) - - return [ - datasets.SplitGenerator(name=datasets.Split.TRAIN, - gen_kwargs={"filepath": downloaded_files["train"]}), - datasets.SplitGenerator(name=datasets.Split.VALIDATION, - gen_kwargs={"filepath": downloaded_files["dev"]}), - datasets.SplitGenerator(name=datasets.Split.TEST, - gen_kwargs={"filepath": downloaded_files["test"]}), - ] - - def _generate_examples(self, filepath): - logger.info("⏳ Generating examples from = %s", filepath) - with open(filepath, encoding="utf-8") as f: - idx = 0 - for line in f: - src_tgt = json.loads(line) - yield idx, src_tgt - idx += 1