saheedniyi commited on
Commit
d535aa2
1 Parent(s): ca0f63c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -58,8 +58,8 @@ size_categories:
58
 
59
  ## Data Collection
60
 
61
- The initial fineweb dataset was collected from **Nairaland.com**, extracting **1,795,908 unique posts** from 19 different sections of the site. Additionally, **1,289,195 outbound links** were extracted from these posts. The content of these web pages was extracted using **Trafilatura**, a popular library for web scraping and content extraction.
62
- The full data collection can be found [in this repo](https://github.com/saheedniyi02/Naijaweb), kindly give a star ().
63
  ## Data Cleaning
64
 
65
  The cleaning process was conducted using **[Datatrove](https://github.com/huggingface/datatrove)**, the same library employed in cleaning the **[FineWeb](https://huggingface.co/datasets/HuggingFaceFW/fineweb)** dataset, which is known for its high quality. The data cleaning process involved multiple stages of deduplication, filtering, and normalization to ensure the dataset's quality matches that of other high-performing datasets.
 
58
 
59
  ## Data Collection
60
 
61
+ The initial fineweb dataset was collected from **Nairaland.com**, extracting **about 30 million unique posts** from 19 different sections of the site. Additionally, **1,289,195 outbound links** were extracted from these posts. The content of these web pages was extracted using **Trafilatura**, a popular library for web scraping and content extraction.
62
+ The full data collection can be found [in this repo](https://github.com/saheedniyi02/Naijaweb), kindly give a star ⭐.
63
  ## Data Cleaning
64
 
65
  The cleaning process was conducted using **[Datatrove](https://github.com/huggingface/datatrove)**, the same library employed in cleaning the **[FineWeb](https://huggingface.co/datasets/HuggingFaceFW/fineweb)** dataset, which is known for its high quality. The data cleaning process involved multiple stages of deduplication, filtering, and normalization to ensure the dataset's quality matches that of other high-performing datasets.