metadata
dataset_info:
features:
- name: id
dtype: string
- name: url
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 21102365479
num_examples: 6623239
download_size: 12161597141
dataset_size: 21102365479
task_categories:
- text-generation
language:
- en
tags:
- LM
size_categories:
- 1M<n<10M
Wikipedia
- Source: https://huggingface.co/datasets/wikipedia
- Num examples: 6,623,239
- Language: English
from datasets import load_dataset
load_dataset("tdtunlp/wikipedia_en")