Ali-C137 commited on
Commit
3c048ee
•
1 Parent(s): 4c7524f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -0
README.md CHANGED
@@ -233,20 +233,27 @@ Encompassing a wide spectrum of content, ranging from social media conversations
233
  This corpus offers comprehensive insights into the linguistic diversity and cultural nuances of Arabic expression.
234
 
235
  ## Usage
 
236
  If you want to use this dataset you pick one among the available configs:
237
 
238
  `Ara--MBZUAI--Bactrian-X` | `Ara--OpenAssistant--oasst1` | `Ary--AbderrahmanSkiredj1--Darija-Wikipedia`
239
 
240
  `Ara--Wikipedia` | `Ary--Wikipedia` | `Arz--Wikipedia`
241
 
 
 
242
  Example of usage:
 
243
  ```python
244
  dataset = load_dataset('Ali-C137/Mixed-Arabic-Datasets', 'Ara--MBZUAI--Bactrian-X')
245
  ```
 
246
  If you loaded multiple datasets and wanted to merge them together then you can simply laverage `concatenate_datasets()` from `datasets`
 
247
  ```pyhton
248
  dataset3 = concatenate_datasets([dataset1['train'], dataset2['train']])
249
  ```
 
250
  Note : proccess the datasets before merging in order to make sure you have a new dataset that is consistent
251
 
252
  ## Dataset Details
@@ -271,6 +278,8 @@ MAD draws from a diverse array of sources, each contributing to its richness and
271
  - [✔] Arabic Wikipedia : [Dataset Link](https://huggingface.co/datasets/wikipedia)
272
  - [✔] Moroccan Arabic Wikipedia : [Dataset Link](https://huggingface.co/datasets/wikipedia)
273
  - [✔] Egyptian Arabic Wikipedia : [Dataset Link](https://huggingface.co/datasets/wikipedia)
 
 
274
  - [] Pain/ArabicTweets : [Dataset Link](https://huggingface.co/datasets/pain/Arabic-Tweets)
275
  - [] Abu-El-Khair Corpus : [Dataset Link](https://huggingface.co/datasets/arabic_billion_words)
276
  - [] QuranExe : [Dataset Link](https://huggingface.co/datasets/mustapha/QuranExe)
 
233
  This corpus offers comprehensive insights into the linguistic diversity and cultural nuances of Arabic expression.
234
 
235
  ## Usage
236
+
237
  If you want to use this dataset you pick one among the available configs:
238
 
239
  `Ara--MBZUAI--Bactrian-X` | `Ara--OpenAssistant--oasst1` | `Ary--AbderrahmanSkiredj1--Darija-Wikipedia`
240
 
241
  `Ara--Wikipedia` | `Ary--Wikipedia` | `Arz--Wikipedia`
242
 
243
+ `Ary--Ali-C137--Darija-Stories-Dataset` | `Ara--Ali-C137--Hindawi-Books-dataset` | ``
244
+
245
  Example of usage:
246
+
247
  ```python
248
  dataset = load_dataset('Ali-C137/Mixed-Arabic-Datasets', 'Ara--MBZUAI--Bactrian-X')
249
  ```
250
+
251
  If you loaded multiple datasets and wanted to merge them together then you can simply laverage `concatenate_datasets()` from `datasets`
252
+
253
  ```pyhton
254
  dataset3 = concatenate_datasets([dataset1['train'], dataset2['train']])
255
  ```
256
+
257
  Note : proccess the datasets before merging in order to make sure you have a new dataset that is consistent
258
 
259
  ## Dataset Details
 
278
  - [✔] Arabic Wikipedia : [Dataset Link](https://huggingface.co/datasets/wikipedia)
279
  - [✔] Moroccan Arabic Wikipedia : [Dataset Link](https://huggingface.co/datasets/wikipedia)
280
  - [✔] Egyptian Arabic Wikipedia : [Dataset Link](https://huggingface.co/datasets/wikipedia)
281
+ - [✔] Darija Stories Dataset : [Dataset Link](https://huggingface.co/datasets/Ali-C137/Darija-Stories-Dataset)
282
+ - [✔] Hindawi Books Dataset : [Dataset Link](https://huggingface.co/datasets/Ali-C137/Hindawi-Books-dataset)
283
  - [] Pain/ArabicTweets : [Dataset Link](https://huggingface.co/datasets/pain/Arabic-Tweets)
284
  - [] Abu-El-Khair Corpus : [Dataset Link](https://huggingface.co/datasets/arabic_billion_words)
285
  - [] QuranExe : [Dataset Link](https://huggingface.co/datasets/mustapha/QuranExe)