--- language: - en paperswithcode_id: wiki-40b pretty_name: Wiki-40B dataset_info: - config_name: ar features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 773508885 num_examples: 220885 - name: validation num_bytes: 44102674 num_examples: 12198 - name: test num_bytes: 43755879 num_examples: 12271 download_size: 413683528 dataset_size: 861367438 - config_name: bg features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 1413477231 num_examples: 130670 - name: validation num_bytes: 78976448 num_examples: 7259 - name: test num_bytes: 78350414 num_examples: 7289 download_size: 484828696 dataset_size: 1570804093 - config_name: ca features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 784791826 num_examples: 277313 - name: validation num_bytes: 43576907 num_examples: 15362 - name: test num_bytes: 44904134 num_examples: 15568 download_size: 480954417 dataset_size: 873272867 - config_name: cs features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 901187017 num_examples: 235971 - name: validation num_bytes: 49743998 num_examples: 13096 - name: test num_bytes: 49325867 num_examples: 12984 download_size: 493522926 dataset_size: 1000256882 - config_name: da features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 247928023 num_examples: 109486 - name: validation num_bytes: 13937406 num_examples: 6173 - name: test num_bytes: 14401179 num_examples: 6219 download_size: 156696617 dataset_size: 276266608 - config_name: de features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 4988094268 num_examples: 1554910 - name: validation num_bytes: 278101948 num_examples: 86068 - name: test num_bytes: 278024815 num_examples: 86594 download_size: 3174352286 dataset_size: 5544221031 - config_name: el features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 1738534924 num_examples: 93596 - name: validation num_bytes: 97711791 num_examples: 5130 - name: test num_bytes: 99743744 num_examples: 5261 download_size: 621575577 dataset_size: 1935990459 - config_name: en features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 9423468036 num_examples: 2926536 - name: validation num_bytes: 527374301 num_examples: 163597 - name: test num_bytes: 522210646 num_examples: 162274 download_size: 6183831905 dataset_size: 10473052983 - config_name: es features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 2906242601 num_examples: 872541 - name: validation num_bytes: 161381260 num_examples: 48592 - name: test num_bytes: 164110964 num_examples: 48764 download_size: 1783120767 dataset_size: 3231734825 - config_name: et features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 196484412 num_examples: 114464 - name: validation num_bytes: 10987144 num_examples: 6351 - name: test num_bytes: 10691693 num_examples: 6205 download_size: 122192870 dataset_size: 218163249 - config_name: fa features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 1551260324 num_examples: 203145 - name: validation num_bytes: 86108146 num_examples: 11180 - name: test num_bytes: 89064531 num_examples: 11262 download_size: 552712695 dataset_size: 1726433001 - config_name: fi features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 589614484 num_examples: 255822 - name: validation num_bytes: 32645294 num_examples: 13962 - name: test num_bytes: 32869383 num_examples: 14179 download_size: 346601923 dataset_size: 655129161 - config_name: fr features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 3850031120 num_examples: 1227206 - name: validation num_bytes: 216405364 num_examples: 68655 - name: test num_bytes: 215243874 num_examples: 68004 download_size: 2246390244 dataset_size: 4281680358 - config_name: he features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 2834322770 num_examples: 165359 - name: validation num_bytes: 160235180 num_examples: 9231 - name: test num_bytes: 162131949 num_examples: 9344 download_size: 754632129 dataset_size: 3156689899 - config_name: hi features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 947403521 num_examples: 45737 - name: validation num_bytes: 54497912 num_examples: 2596 - name: test num_bytes: 54448878 num_examples: 2643 download_size: 231716300 dataset_size: 1056350311 - config_name: hr features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 247471855 num_examples: 103857 - name: validation num_bytes: 14004242 num_examples: 5792 - name: test num_bytes: 13881533 num_examples: 5724 download_size: 158644264 dataset_size: 275357630 - config_name: it features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 1998187938 num_examples: 732609 - name: validation num_bytes: 109399796 num_examples: 40684 - name: test num_bytes: 108160871 num_examples: 40443 download_size: 1330554944 dataset_size: 2215748605 - config_name: ja features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 7719156890 num_examples: 745392 - name: validation num_bytes: 423396781 num_examples: 41576 - name: test num_bytes: 424775191 num_examples: 41268 download_size: 2914923230 dataset_size: 8567328862 - config_name: ko features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 1424423053 num_examples: 194977 - name: validation num_bytes: 79027067 num_examples: 10805 - name: test num_bytes: 78623281 num_examples: 10802 download_size: 568560655 dataset_size: 1582073401 - config_name: nl features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 906908479 num_examples: 447555 - name: validation num_bytes: 51519150 num_examples: 25201 - name: test num_bytes: 49492508 num_examples: 24776 download_size: 594312303 dataset_size: 1007920137 - config_name: pl features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 1250270240 num_examples: 505191 - name: validation num_bytes: 70048390 num_examples: 28310 - name: test num_bytes: 69957343 num_examples: 27987 download_size: 755556434 dataset_size: 1390275973 - config_name: pt features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 1186541609 num_examples: 406507 - name: validation num_bytes: 65911750 num_examples: 22301 - name: test num_bytes: 65941634 num_examples: 22693 download_size: 725984914 dataset_size: 1318394993 - config_name: ru features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 14041955183 num_examples: 926037 - name: validation num_bytes: 787569099 num_examples: 51287 - name: test num_bytes: 782630173 num_examples: 51885 download_size: 4959684748 dataset_size: 15612154455 - config_name: th features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 1167742322 num_examples: 56798 - name: validation num_bytes: 58604863 num_examples: 3093 - name: test num_bytes: 63235795 num_examples: 3114 download_size: 286569412 dataset_size: 1289582980 - config_name: tr features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 417796625 num_examples: 142576 - name: validation num_bytes: 23829728 num_examples: 7845 - name: test num_bytes: 23573543 num_examples: 7890 download_size: 208571967 dataset_size: 465199896 - config_name: zh-cn features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 902812807 num_examples: 549672 - name: validation num_bytes: 50487729 num_examples: 30299 - name: test num_bytes: 49584239 num_examples: 30355 download_size: 667605463 dataset_size: 1002884775 - config_name: zh-tw features: - name: wikidata_id dtype: string - name: text dtype: string - name: version_id dtype: string splits: - name: train num_bytes: 3254625339 num_examples: 552031 - name: validation num_bytes: 185024571 num_examples: 30739 - name: test num_bytes: 181148137 num_examples: 30670 download_size: 1375185673 dataset_size: 3620798047 configs: - config_name: ar data_files: - split: train path: ar/train-* - split: validation path: ar/validation-* - split: test path: ar/test-* - config_name: bg data_files: - split: train path: bg/train-* - split: validation path: bg/validation-* - split: test path: bg/test-* - config_name: ca data_files: - split: train path: ca/train-* - split: validation path: ca/validation-* - split: test path: ca/test-* - config_name: cs data_files: - split: train path: cs/train-* - split: validation path: cs/validation-* - split: test path: cs/test-* - config_name: da data_files: - split: train path: da/train-* - split: validation path: da/validation-* - split: test path: da/test-* - config_name: de data_files: - split: train path: de/train-* - split: validation path: de/validation-* - split: test path: de/test-* - config_name: el data_files: - split: train path: el/train-* - split: validation path: el/validation-* - split: test path: el/test-* - config_name: en data_files: - split: train path: en/train-* - split: validation path: en/validation-* - split: test path: en/test-* - config_name: es data_files: - split: train path: es/train-* - split: validation path: es/validation-* - split: test path: es/test-* - config_name: et data_files: - split: train path: et/train-* - split: validation path: et/validation-* - split: test path: et/test-* - config_name: fa data_files: - split: train path: fa/train-* - split: validation path: fa/validation-* - split: test path: fa/test-* - config_name: fi data_files: - split: train path: fi/train-* - split: validation path: fi/validation-* - split: test path: fi/test-* - config_name: fr data_files: - split: train path: fr/train-* - split: validation path: fr/validation-* - split: test path: fr/test-* - config_name: he data_files: - split: train path: he/train-* - split: validation path: he/validation-* - split: test path: he/test-* - config_name: hi data_files: - split: train path: hi/train-* - split: validation path: hi/validation-* - split: test path: hi/test-* - config_name: hr data_files: - split: train path: hr/train-* - split: validation path: hr/validation-* - split: test path: hr/test-* - config_name: it data_files: - split: train path: it/train-* - split: validation path: it/validation-* - split: test path: it/test-* - config_name: ja data_files: - split: train path: ja/train-* - split: validation path: ja/validation-* - split: test path: ja/test-* - config_name: ko data_files: - split: train path: ko/train-* - split: validation path: ko/validation-* - split: test path: ko/test-* - config_name: nl data_files: - split: train path: nl/train-* - split: validation path: nl/validation-* - split: test path: nl/test-* - config_name: pl data_files: - split: train path: pl/train-* - split: validation path: pl/validation-* - split: test path: pl/test-* - config_name: pt data_files: - split: train path: pt/train-* - split: validation path: pt/validation-* - split: test path: pt/test-* - config_name: ru data_files: - split: train path: ru/train-* - split: validation path: ru/validation-* - split: test path: ru/test-* - config_name: th data_files: - split: train path: th/train-* - split: validation path: th/validation-* - split: test path: th/test-* - config_name: tr data_files: - split: train path: tr/train-* - split: validation path: tr/validation-* - split: test path: tr/test-* - config_name: zh-cn data_files: - split: train path: zh-cn/train-* - split: validation path: zh-cn/validation-* - split: test path: zh-cn/test-* - config_name: zh-tw data_files: - split: train path: zh-tw/train-* - split: validation path: zh-tw/validation-* - split: test path: zh-tw/test-* --- # Dataset Card for "wiki40b" ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) - [Languages](#languages) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Dataset Creation](#dataset-creation) - [Curation Rationale](#curation-rationale) - [Source Data](#source-data) - [Annotations](#annotations) - [Personal and Sensitive Information](#personal-and-sensitive-information) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Social Impact of Dataset](#social-impact-of-dataset) - [Discussion of Biases](#discussion-of-biases) - [Other Known Limitations](#other-known-limitations) - [Additional Information](#additional-information) - [Dataset Curators](#dataset-curators) - [Licensing Information](#licensing-information) - [Citation Information](#citation-information) - [Contributions](#contributions) ## Dataset Description - **Homepage:** [https://research.google/pubs/pub49029/](https://research.google/pubs/pub49029/) - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) - **Size of downloaded dataset files:** 0.00 MB - **Size of the generated dataset:** 10.47 GB - **Total amount of disk used:** 10.47 GB ### Dataset Summary Clean-up text for 40+ Wikipedia languages editions of pages correspond to entities. The datasets have train/dev/test splits per language. The dataset is cleaned up by page filtering to remove disambiguation pages, redirect pages, deleted pages, and non-entity pages. Each example contains the wikidata id of the entity, and the full Wikipedia article after page processing that removes non-content sections and structured objects. ### Supported Tasks and Leaderboards [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Languages [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Dataset Structure ### Data Instances #### en - **Size of downloaded dataset files:** 0.00 MB - **Size of the generated dataset:** 10.47 GB - **Total amount of disk used:** 10.47 GB An example of 'train' looks as follows. ``` ``` ### Data Fields The data fields are the same among all splits. #### en - `wikidata_id`: a `string` feature. - `text`: a `string` feature. - `version_id`: a `string` feature. ### Data Splits |name| train |validation| test | |----|------:|---------:|-----:| |en |2926536| 163597|162274| ## Dataset Creation ### Curation Rationale [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Source Data #### Initial Data Collection and Normalization [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the source language producers? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Annotations #### Annotation process [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) #### Who are the annotators? [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Personal and Sensitive Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Discussion of Biases [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Other Known Limitations [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ## Additional Information ### Dataset Curators [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Licensing Information [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) ### Citation Information ``` ``` ### Contributions Thanks to [@jplu](https://github.com/jplu), [@patrickvonplaten](https://github.com/patrickvonplaten), [@thomwolf](https://github.com/thomwolf), [@albertvillanova](https://github.com/albertvillanova), [@lhoestq](https://github.com/lhoestq) for adding this dataset.