Convert dataset sizes from base 2 to base 10 in the dataset card
#1
by
albertvillanova
HF staff
- opened
README.md
CHANGED
@@ -343,9 +343,9 @@ dataset_info:
|
|
343 |
- **Repository:** https://github.com/neulab/word-embeddings-for-nmt
|
344 |
- **Paper:** [When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation?](https://aclanthology.org/N18-2084/)
|
345 |
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
346 |
-
- **Size of downloaded dataset files:**
|
347 |
-
- **Size of the generated dataset:**
|
348 |
-
- **Total amount of disk used:**
|
349 |
|
350 |
### Dataset Summary
|
351 |
|
@@ -366,9 +366,9 @@ where one is high resource and the other is low resource.
|
|
366 |
|
367 |
#### az_to_en
|
368 |
|
369 |
-
- **Size of downloaded dataset files:**
|
370 |
-
- **Size of the generated dataset:** 1.
|
371 |
-
- **Total amount of disk used:**
|
372 |
|
373 |
An example of 'train' looks as follows.
|
374 |
```
|
@@ -382,9 +382,9 @@ An example of 'train' looks as follows.
|
|
382 |
|
383 |
#### aztr_to_en
|
384 |
|
385 |
-
- **Size of downloaded dataset files:**
|
386 |
-
- **Size of the generated dataset:**
|
387 |
-
- **Total amount of disk used:**
|
388 |
|
389 |
An example of 'train' looks as follows.
|
390 |
```
|
@@ -398,9 +398,9 @@ An example of 'train' looks as follows.
|
|
398 |
|
399 |
#### be_to_en
|
400 |
|
401 |
-
- **Size of downloaded dataset files:**
|
402 |
-
- **Size of the generated dataset:** 1.
|
403 |
-
- **Total amount of disk used:**
|
404 |
|
405 |
An example of 'train' looks as follows.
|
406 |
```
|
@@ -414,9 +414,9 @@ An example of 'train' looks as follows.
|
|
414 |
|
415 |
#### beru_to_en
|
416 |
|
417 |
-
- **Size of downloaded dataset files:**
|
418 |
-
- **Size of the generated dataset:**
|
419 |
-
- **Total amount of disk used:**
|
420 |
|
421 |
An example of 'validation' looks as follows.
|
422 |
```
|
@@ -429,9 +429,9 @@ This example was too long and was cropped:
|
|
429 |
|
430 |
#### es_to_pt
|
431 |
|
432 |
-
- **Size of downloaded dataset files:**
|
433 |
-
- **Size of the generated dataset:**
|
434 |
-
- **Total amount of disk used:**
|
435 |
|
436 |
An example of 'validation' looks as follows.
|
437 |
```
|
|
|
343 |
- **Repository:** https://github.com/neulab/word-embeddings-for-nmt
|
344 |
- **Paper:** [When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation?](https://aclanthology.org/N18-2084/)
|
345 |
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
|
346 |
+
- **Size of downloaded dataset files:** 1.83 GB
|
347 |
+
- **Size of the generated dataset:** 281.66 MB
|
348 |
+
- **Total amount of disk used:** 2.12 GB
|
349 |
|
350 |
### Dataset Summary
|
351 |
|
|
|
366 |
|
367 |
#### az_to_en
|
368 |
|
369 |
+
- **Size of downloaded dataset files:** 131.01 MB
|
370 |
+
- **Size of the generated dataset:** 1.53 MB
|
371 |
+
- **Total amount of disk used:** 132.54 MB
|
372 |
|
373 |
An example of 'train' looks as follows.
|
374 |
```
|
|
|
382 |
|
383 |
#### aztr_to_en
|
384 |
|
385 |
+
- **Size of downloaded dataset files:** 131.01 MB
|
386 |
+
- **Size of the generated dataset:** 40.14 MB
|
387 |
+
- **Total amount of disk used:** 171.15 MB
|
388 |
|
389 |
An example of 'train' looks as follows.
|
390 |
```
|
|
|
398 |
|
399 |
#### be_to_en
|
400 |
|
401 |
+
- **Size of downloaded dataset files:** 131.01 MB
|
402 |
+
- **Size of the generated dataset:** 1.43 MB
|
403 |
+
- **Total amount of disk used:** 132.42 MB
|
404 |
|
405 |
An example of 'train' looks as follows.
|
406 |
```
|
|
|
414 |
|
415 |
#### beru_to_en
|
416 |
|
417 |
+
- **Size of downloaded dataset files:** 131.01 MB
|
418 |
+
- **Size of the generated dataset:** 60.20 MB
|
419 |
+
- **Total amount of disk used:** 191.21 MB
|
420 |
|
421 |
An example of 'validation' looks as follows.
|
422 |
```
|
|
|
429 |
|
430 |
#### es_to_pt
|
431 |
|
432 |
+
- **Size of downloaded dataset files:** 131.01 MB
|
433 |
+
- **Size of the generated dataset:** 9.13 MB
|
434 |
+
- **Total amount of disk used:** 140.14 MB
|
435 |
|
436 |
An example of 'validation' looks as follows.
|
437 |
```
|