Datasets:

Multilinguality:
multilingual
Language Creators:
found
Annotations Creators:
no-annotation
Source Datasets:
original
ArXiv:
License:
dirkgr commited on
Commit
34c61cd
1 Parent(s): 1ddc917

Updated readme

Browse files
Files changed (1) hide show
  1. README.md +18 -7
README.md CHANGED
@@ -1,12 +1,16 @@
1
  This is the processed version of [Google's C4 dataset](https://www.tensorflow.org/datasets/catalog/c4).
2
 
3
- We prepared three variants of the data: `en`, `en.noclean`, and `realnewslike`. There is a fourth one, `webtextlike`, which was not ready as I am writing this, but we are working on it. If you are interested in the `multilingual` version, please get in touch.
4
 
5
  For reference, these are the sizes of the variants:
6
 
7
- - `en`: 300GB
8
  - `en.noclean`: 2.3TB
 
9
  - `realnewslike`: 15GB
 
 
 
10
 
11
  # How do I download this?
12
 
@@ -16,17 +20,24 @@ Unfortunately we ran out of time making this into a proper Huggingface dataset,
16
  git clone https://huggingface.co/datasets/allenai/c4
17
  ```
18
 
19
- If you want only one of the variants, you need some more commands:
20
 
21
  ```bash
22
- git clone -n https://huggingface.co/datasets/allenai/c4
23
  cd c4
24
- git sparse-checkout init --cone
25
- git sparse-checkout set en
26
  ```
27
 
28
- You can use `git sparse-checkout set` multiple times to select multiple datasets.
 
 
 
 
29
 
30
  # Acknowledgements
31
 
32
  Big ups to the good folks at [Common Crawl](https://commoncrawl.org) whose data made this possible ([consider donating](http://commoncrawl.org/donate/)!), to Google for creating the code that curates and filters the data, and to Huggingface, who had no issue with hosting these 3TB of data for public download!
 
 
 
 
 
1
  This is the processed version of [Google's C4 dataset](https://www.tensorflow.org/datasets/catalog/c4).
2
 
3
+ We prepared five variants of the data: `en`, `en.noclean`, `en.noblocklist`, `realnewslike`, and `multilingual`.
4
 
5
  For reference, these are the sizes of the variants:
6
 
7
+ - `en`: 305GB
8
  - `en.noclean`: 2.3TB
9
+ - `en.noblocklist`: 380GB
10
  - `realnewslike`: 15GB
11
+ - `multilingual`: 9.7TB
12
+
13
+ The `en.noblocklist` variant is exactly the same as the `en` variant, except we turned off the so-called "badwords filter", which removes all documents that contain words from the lists at https://github.com/LDNOOBW/List-of-Dirty-Naughty-Obscene-and-Otherwise-Bad-Words.
14
 
15
  # How do I download this?
16
 
 
20
  git clone https://huggingface.co/datasets/allenai/c4
21
  ```
22
 
23
+ This will download 13TB to your local drive. If you want to be more precise with what you are downloading, follow these commands instead:
24
 
25
  ```bash
26
+ GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/allenai/c4
27
  cd c4
28
+ git lfs pull --include "en/*"
 
29
  ```
30
 
31
+ The `git clone` command in this variant will download a bunch of stub files that Git LFS uses, so you can see all the filenames that exist that way. You can then convert the stubs into their real files with `git lfs pull --include "..."`. For example, if you wanted all the Dutch documents from the multilingual set, you would run
32
+
33
+ ```bash
34
+ git lfs pull --include "multilingual/c4-nl.*.json.gz"
35
+ ```
36
 
37
  # Acknowledgements
38
 
39
  Big ups to the good folks at [Common Crawl](https://commoncrawl.org) whose data made this possible ([consider donating](http://commoncrawl.org/donate/)!), to Google for creating the code that curates and filters the data, and to Huggingface, who had no issue with hosting these 3TB of data for public download!
40
+
41
+ ### License
42
+
43
+ We are releasing this dataset under the terms of [ODC-BY](https://opendatacommons.org/licenses/by/1-0/). By using this, you are also bound by the [Common Crawl terms of use](https://commoncrawl.org/terms-of-use/) in respect of the content contained in the dataset.