--- dataset_info: features: - name: text dtype: string - name: meta struct: - name: pile_set_name dtype: string splits: - name: train num_bytes: 1604258672 num_examples: 4000 download_size: 972760838 dataset_size: 1604258672 configs: - config_name: default data_files: - split: train path: data/train-* --- ## Dataset Creation Process These subsets were created by streaming over the rows from [`monology/pile-uncopyrighted`](https://huggingface.co/datasets/monology/pile-uncopyrighted) and filtering by the meta column. Each subset is generally limited to the first 100,000 qualifying rows encountered. ## Citations If you use this dataset, please cite the original Pile papers: ```bibtex @article{gao2020pile, title={The Pile: An 800GB dataset of diverse text for language modeling}, author={Gao, Leo and Biderman, Stella and Black, Sid and Golding, Laurence and Hoppe, Travis and Foster, Charles and Phang, Jason and He, Horace and Thite, Anish and Nabeshima, Noa and others}, journal={arXiv preprint arXiv:2101.00027}, year={2020} } @article{biderman2022datasheet, title={Datasheet for the pile}, author={Biderman, Stella and Bicheno, Kieran and Gao, Leo}, journal={arXiv preprint arXiv:2201.07311}, year={2022} } ``` **Note**: Some subsets contain fewer than 100,000 rows due to the sparsity of certain categories in the original dataset. In these cases, the rows may not be the first encountered but are unevenly distributed due to parallelization during collection. ## Future Updates Future updates may include completing the smaller subsets to reach the 100,000 row target, which could result in breaking changes for those specific subsets.