diff --git "a/README.md" "b/README.md"
new file mode 100644--- /dev/null
+++ "b/README.md"
@@ -0,0 +1,11047 @@
+---
+annotations_creators:
+- no-annotation
+language_creators:
+- crowdsourced
+pretty_name: Banglawiki
+paperswithcode_id: null
+license:
+- cc-by-sa-3.0
+- gfdl
+task_categories:
+- text-generation
+- fill-mask
+task_ids:
+- language-modeling
+- masked-language-modeling
+source_datasets:
+- original
+multilinguality:
+- multilingual
+size_categories:
+- n<1K
+- 1K
+and identify the date.
+
+### 2. [Optional] Get a refreshed list of languages
+
+This is optional because it not very likely that a new language will have
+suddenly appeared since the last version _and_ have a significant dataset.
+
+Navigate to and copy the
+languages column from the "Detailed list" table (near the end of the page).
+
+Copy that content in the form of a Python list into `lang_def.py` (at the top
+of the repo) under a new date.
+
+### 3. [Optional] Create Media and Category aliases
+
+In order to properly extract links to images and media in all languages, we
+must refresh the two corresponding files. To do so, from the root of the repo,
+run
+
+```sh
+python -m prep.create_aliases
+```
+
+This will create or update these two files at the root of the repo:
+
+- `media_aliases.py`
+- `category_aliases.py`
+
+These files are used in the final step
+
+### 4. Build and prepare the datasets into sharded parquet files
+
+Running this script downloads the wikipedia dumps for each language in
+`lang_def.py` and shards each language dataset into the appropriate number of
+shards (max size ~ 250MB).
+
+```sh
+python -m prep.build --date 20230601
+```
+
+There are other options:
+
+```text
+$ python -m prep.build --help
+usage: Banglawiki Builder [-h] [--date DATE] [--language [LANG ...]] [--cache-dir DIR] [--mirror MIRROR]
+
+Prepares the Banglawiki dataset for each language
+
+optional arguments:
+ -h, --help show this help message and exit
+ --date DATE Banglawiki dump date (e.g. 20230601)
+ --language [LANG ...] Language code (e.g. en). If missing, all languages are processed
+ --cache-dir DIR Cache directory for 🤗 Datasets
+ --mirror MIRROR Mirror URL
+```
+
+For instance, for faster downloads of the dumps, use the mirror option:
+
+```sh
+python -m prep.build \
+ --date 20230601 \
+ --language bs \
+ --mirror https://mirror.accum.se/mirror/wikimedia.org/dumps/
+```
+
+It will download the dumps at around 60MB/s instead of the capped speed
+(~4MB/s) from . The script will skip existing
+directories, allowing you to run the script in several passes.
+
+Notes:
+
+- These instructions build upon the build process of the
+ [Banglawiki](https://huggingface.co/datasets/banglawiki) 🤗 Dataset. HF did a
+ fantastic job, I just pushed it a bit further.
+- Be aware that not all mirrors contain all dumps. For instance mirror.accum.se
+ does not contain dumps for languages such as be-x-old or cbk-zam. My own
+ solution is to run a first pass using the aforementioned mirror, and a second
+ pass with the official `https://dumps.wikimedia.org` site (omitting the
+ `--mirror` parameter).
\ No newline at end of file