Datasets:

Modalities:
Text
ArXiv:
Libraries:
Datasets
RedPajama-Data-V2 / README.md
Maurice Weber
update README.md
34b0752
|
raw
history blame
11.2 kB
---
task_categories:
- text-generation
language:
- en
- de
- fr
- es
- it
pretty_name: Red Pajama V2 Data Foundation
---
### Getting Started
The full RedPajama-V2 dataset is a data foundation that includes over 100B text documents coming from 84 CommonCrawl
snapshots and processed using the [CCNet](https://github.com/facebookresearch/cc_net) pipeline. Out of these, there are
30B documents in the corpus that additionally come with quality signals.
Check out our [blog post](XXXXX) for more details on the build process, dataset structure and schema.
To familiarize yourself with the dataset, you can load the sample dataset using:
```python
from datasets import load_dataset
ds = load_dataset("togethercomputer/RedPajama-Data-V2", name="sample")
```
To download a the dataset for a specific combination of `{partition} x {snapshot_id} x {language}`, you can run
```python
from datasets import load_dataset
ds = load_dataset("togethercomputer/RedPajama-Data-V2",
name="sample",
partition="head_middle",
snapshots=["2023-06", "2022-49"],
languages=["en", "de"])
```
Alternatively, you can also directly download the files using the following instructions, using English data from the
`2023-06` snapshot and the `head_middle` partition as an example. The full set of CC snapshots included in the dataset
is given in `_CC_SNAPSHOT_IDS`, and the available partitions are `tail` and `head_middle`. The available language tags
are `en`, `de`, `fr`, `es`, `it`.
```bash
CC_SNAPSHOT="2023-06"
LANG="en"
PARTITION="head_middle"
BASE_URL="https://data.together.xyz/redpajama-data-v2/v1.0.0/"
listings_file="${LANG}-${CC_SNAPSHOT}-${PARTITION}.txt"
wget "${BASE_URL}/listings/${listings_file}"
# download documents
while read line; do
url="${BASE_URL}/documents/${line}.json.gz"
dest="documents/${line}.json.gz"
mkdir -p $(dirname $dest)
wget "$line" -O "$dest"
done <"${LANG}-${CC_SNAPSHOT}-${PARTITION}.txt"
# download other components
COMPS=("quality_signals" "minhash" "duplicates")
for comp in "${COMPS[@]}"; do
while read line; do
url="${BASE_URL}/${comp}/${line}.${comp}.json.gz"
dest="${comp}/${line}.${comp}.json.gz"
mkdir -p $(dirname $dest)
wget "$line" -O "$dest"
done <"${LANG}-${CC_SNAPSHOT}-${PARTITION}.txt"
done
```
A full set of scripts to recreate the dataset, including the quality signals, can be
found [here](https://github.com/togethercomputer/RedPajama-Data).
### Dataset Summary
RedPajama-V2 is an open dataset for training large laguage models and includes over 100B text documents. Out of these,
30B documents come with quality annotations.
#### Quality Annotations
| Annotation Tag | Description | Category | Reference |
|--------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------------|---------------|------------------------------------|
| ccnet_bucket | head, middle or tail bucket of the perplexity score | ccnet | ccnet |
| ccnet_language_score | score of the language identification model | ccnet | ccnet |
| ccnet_length | number of characters | ccnet | ccnet |
| ccnet_nlines | number of lines | ccnet | ccnet |
| ccnet_original_length | number of characters before in-document line deduplication | ccnet | ccnet |
| ccnet_original_nlines | number of lines before in-document line deduplication | ccnet | ccnet |
| ccnet_perplexity | perplexity of an LM trained on Wikipedia | ccnet | ccnet |
| rps_doc_books_importance | Given a bag of {1,2}-wordgram model trained on Books p, and a model trained on the source domain q, This is the logarithm of the ratio p(doc)/q(doc) | ML Heuristics | Importance Resampling (Xie et al.) |
| rps_doc_openwebtext_importance | Given a bag of {1,2}-wordgram model trained on OpenWebText p, and a model trained on the source domain q, this is the logarithm of the ratio p(doc)/q(doc). | ML Heuristics | Importance Resampling (Xie et al.) |
| rps_doc_wikipedia_importance | Given a bag of {1,2}-wordgram model trained on Wikipedia articles p, and a model trained on the source domain q, this is the logarithm of the ratio p(doc)/q(doc). | ML Heuristics | Importance Resampling (Xie et al.) |
| rps_doc_ml_wikiref_score | Fasttext classifier prediction for the document being a Wikipedia | ML Heuristics | LLaMA, RedPajama-1T |
| | reference. This is the same fasttext model used in the RedPajama-1T | | |
| | dataset. Only applies to English data. | | |
| rps_doc_ml_palm_score | Fasttext classifier prediction for the document being a Wikipedia | ML Heuristics | PaLM, GLaM |
| | article, OpenWebText sample or a RedPajama-V1 book. Only for English | | |
| | data. | | |
| rps_doc_ml_wikipedia_score | Fasttext classifier prediction for the document being a Wikipedia | ML Heuristics | - |
| | article. This is used for non-English data | | |
#### Document Counts for the Annotated part of the dataset
| | en | de | fr | es | it | Total |
|-------------|-------|------|------|------|------|-------|
| # Documents | 24.5B | 2.7B | 2.2B | 2.3B | 1.2B | 32.9B |
### Languages
English, German, French, Italian, Spanish
## Dataset Structure
The dataset is structured into four components, each following the same key structure:
```
β”œβ”€β”€ documents
β”œβ”€β”€ 2018-43
β”œβ”€β”€ 0000
β”œβ”€β”€ en_head.json.gz
β”œβ”€β”€ ...
β”œβ”€β”€ it_middle.json.gz
β”œβ”€β”€ quality_signals
β”œβ”€β”€ 2018-43
β”œβ”€β”€ 0000
β”œβ”€β”€ en_head.signals.json.gz
β”œβ”€β”€ ...
β”œβ”€β”€ it_middle.json.gz
β”œβ”€β”€ duplicates
β”œβ”€β”€ 2018-43
β”œβ”€β”€ 0000
β”œβ”€β”€ en_head.duplicates.parquet
β”œβ”€β”€ ...
β”œβ”€β”€ it_middle.duplicates.parquet
β”œβ”€β”€ minhash
β”œβ”€β”€ 2018-43
β”œβ”€β”€ 0000
β”œβ”€β”€ en_head.minhash.parquet
β”œβ”€β”€ ...
β”œβ”€β”€ it_middle.minhash.parquet
```
Documents files, which contain the text, folow the schema defined by CCNet, and the quality signals follow the schema
```json
{
"id": "2018-43/0000/en_head.json.gz/0",
"id_int": 7972430436813205988,
"metadata": {
"cc_segment": "crawl-data/...",
"cc_net_source": "2018-43/0000/en_head.json.gz",
"url": "...",
"source_domain": "...",
"language": "en",
"snapshot_id": "2018-43"
},
"quality_signals": {
"ccnet_original_length": [
[
0,
7033,
8711.0
]
],
...,
"rps_doc_stop_word_fraction": [
[
0,
7033,
0.45121107
]
],
"rps_lines_num_words": [
[
0,
25,
2
],
...,
[
6980,
7033,
10
]
]
}
}
```
where signal scores are encoded as a list of tuples `(start, end, score)`, where `start` and `end` are the locations in
the
`raw_content` string where the `score` applies.
## Dataset Creation
The dataset is based on 84 snapshots provided by Common Crawl.
## Citation
To cite RedPajama-V2, please use:
```
@software{together2023redpajama-v2,
author = {Together Computer},
title = {RedPajama-Data-v2: a living data foundation for training open LLM models},
month = October,
year = 2023,
url = {https://github.com/togethercomputer/RedPajama-Data}
}
```
## License
Please refer to the [Common Crawl Foundation Terms of Use](https://commoncrawl.org/terms-of-use) for the data.
The code used to load and process the dataset is licensed under the Apache 2.0 license.
<!--
### Annotations
#### Annotation process
[More Information Needed]
#### Who are the annotators?
[More Information Needed]
### Personal and Sensitive Information
[More Information Needed]
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
[More Information Needed]
### Licensing Information
[More Information Needed]
### Citation Information
[More Information Needed]
### Contributions
[More Information Needed]
-->