Datasets:

Languages:
English
Multilinguality:
multilingual
Size Categories:
100M<n<1B
Language Creators:
found
Annotations Creators:
no-annotation
Source Datasets:
original
ArXiv:
License:
system HF staff commited on
Commit
f128177
1 Parent(s): 84e3e67

Update files from the datasets library (from 1.4.0)

Browse files

Release notes: https://github.com/huggingface/datasets/releases/tag/1.4.0

Files changed (1) hide show
  1. c4.py +4 -2
c4.py CHANGED
@@ -19,7 +19,6 @@
19
  from __future__ import absolute_import, division, print_function
20
 
21
  import json
22
- import logging
23
  import os
24
 
25
  import datasets
@@ -39,6 +38,9 @@ from .c4_utils import (
39
  )
40
 
41
 
 
 
 
42
  _DESCRIPTION = """\
43
  A colossal, cleaned version of Common Crawl's web crawl corpus.
44
 
@@ -215,7 +217,7 @@ class C4(datasets.BeamBasedBuilder):
215
  cc_dir, "*.warc.wet.gz", self.manual_download_instructions
216
  )
217
  )
218
- logging.info("Adding %d WET files for manually downloaded version %s.", len(wet_files), cc_version)
219
  file_paths["wet_files"].extend(wet_files)
220
 
221
  page_content_pcollection = self._get_page_content(pipeline, file_paths, dl_manager)
19
  from __future__ import absolute_import, division, print_function
20
 
21
  import json
 
22
  import os
23
 
24
  import datasets
38
  )
39
 
40
 
41
+ logger = datasets.logging.get_logger(__name__)
42
+
43
+
44
  _DESCRIPTION = """\
45
  A colossal, cleaned version of Common Crawl's web crawl corpus.
46
 
217
  cc_dir, "*.warc.wet.gz", self.manual_download_instructions
218
  )
219
  )
220
+ logger.info("Adding %d WET files for manually downloaded version %s.", len(wet_files), cc_version)
221
  file_paths["wet_files"].extend(wet_files)
222
 
223
  page_content_pcollection = self._get_page_content(pipeline, file_paths, dl_manager)