Tianyu Gao commited on
Commit
1bba113
1 Parent(s): 5930cf8

update data

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. bin-shard1-8/dict.txt +0 -0
  2. bin-shard1-8/preprocess.log +8 -0
  3. bin-shard1-8/train.bin +3 -0
  4. bin-shard1-8/train.idx +3 -0
  5. bin-shard1-8/valid.bin +3 -0
  6. bin-shard1-8/valid.idx +3 -0
  7. bin-shard10-8/dict.txt +0 -0
  8. bin-shard10-8/preprocess.log +8 -0
  9. bin-shard10-8/train.bin +3 -0
  10. bin-shard10-8/train.idx +3 -0
  11. bin-shard10-8/valid.bin +3 -0
  12. bin-shard10-8/valid.idx +3 -0
  13. bin-shard11-8/dict.txt +0 -0
  14. bin-shard11-8/preprocess.log +8 -0
  15. bin-shard11-8/train.bin +3 -0
  16. bin-shard11-8/train.idx +3 -0
  17. bin-shard11-8/valid.bin +3 -0
  18. bin-shard11-8/valid.idx +3 -0
  19. bin-shard12-8/dict.txt +0 -0
  20. bin-shard12-8/preprocess.log +8 -0
  21. bin-shard12-8/train.bin +3 -0
  22. bin-shard12-8/train.idx +3 -0
  23. bin-shard12-8/valid.bin +3 -0
  24. bin-shard12-8/valid.idx +3 -0
  25. bin-shard13-8/dict.txt +0 -0
  26. bin-shard13-8/preprocess.log +8 -0
  27. bin-shard13-8/train.bin +3 -0
  28. bin-shard13-8/train.idx +3 -0
  29. bin-shard13-8/valid.bin +3 -0
  30. bin-shard13-8/valid.idx +3 -0
  31. bin-shard14-8/dict.txt +0 -0
  32. bin-shard14-8/preprocess.log +8 -0
  33. bin-shard14-8/train.bin +3 -0
  34. bin-shard14-8/train.idx +3 -0
  35. bin-shard14-8/valid.bin +3 -0
  36. bin-shard14-8/valid.idx +3 -0
  37. bin-shard15-8/dict.txt +0 -0
  38. bin-shard15-8/preprocess.log +8 -0
  39. bin-shard15-8/train.bin +3 -0
  40. bin-shard15-8/train.idx +3 -0
  41. bin-shard15-8/valid.bin +3 -0
  42. bin-shard15-8/valid.idx +3 -0
  43. bin-shard2-8/dict.txt +0 -0
  44. bin-shard2-8/preprocess.log +8 -0
  45. bin-shard2-8/train.bin +3 -0
  46. bin-shard2-8/train.idx +3 -0
  47. bin-shard2-8/valid.bin +3 -0
  48. bin-shard2-8/valid.idx +3 -0
  49. bin-shard3-8/dict.txt +0 -0
  50. bin-shard3-8/preprocess.log +8 -0
bin-shard1-8/dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
bin-shard1-8/preprocess.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Namespace(align_suffix=None, alignfile=None, all_gather_list_size=16384, azureml_logging=False, bf16=False, bpe=None, cpu=False, criterion='cross_entropy', dataset_impl='mmap', destdir='/n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard1-8', empty_cache_freq=0, fp16=False, fp16_init_scale=128, fp16_no_flatten_grads=False, fp16_scale_tolerance=0.0, fp16_scale_window=None, joined_dictionary=False, log_format=None, log_interval=100, lr_scheduler='fixed', memory_efficient_bf16=False, memory_efficient_fp16=False, min_loss_scale=0.0001, model_parallel_size=1, no_progress_bar=False, nwordssrc=-1, nwordstgt=-1, only_source=True, optimizer=None, padding_factor=8, plasma_path='/tmp/plasma', profile=False, quantization_config_path=None, reset_logging=False, scoring='bleu', seed=1, simul_type=None, slurm_job_id=None, slurm_job_name=None, source_lang=None, srcdict='/n/fs/nlp-awettig/data-pretraining/gpt2_bpe/dict.txt', suppress_crashes=False, tag=None, target_lang=None, task='translation', tensorboard_logdir=None, testpref=None, tgtdict=None, threshold_loss_scale=None, thresholdsrc=0, thresholdtgt=0, tokenizer=None, tpu=False, trainpref='/n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard1-8', use_plasma_view=False, user_dir=None, validpref='/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe,/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe', wandb_project=None, workers=12)
2
+ [None] Dictionary: 50264 types
3
+ [None] /n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard1-8: 18842955 sents, 538297703 tokens, 0.0% replaced by <unk>
4
+ [None] Dictionary: 50264 types
5
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe: 1801350 sents, 117517428 tokens, 0.0% replaced by <unk>
6
+ [None] Dictionary: 50264 types
7
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe: 3760 sents, 246403 tokens, 0.0% replaced by <unk>
8
+ Wrote preprocessed data to /n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard1-8
bin-shard1-8/train.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:743ea2d6f95ac0ecf840b80cf1638f624b15ff4a414c7f51c091a2097058a21c
3
+ size 1076595406
bin-shard1-8/train.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4d869bdf7a2aee1d40abb5911219d35afea52f87e6974dfca504920bdc06c58
3
+ size 226115486
bin-shard1-8/valid.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c590981bafb9ee39836011daf94fa23e429d3de88dd1c7a7d6c1699aa8147828
3
+ size 492806
bin-shard1-8/valid.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c730179c16607e30876530e1886e62ef9cdd0daf228f8b86de8d5eb752a91c
3
+ size 45146
bin-shard10-8/dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
bin-shard10-8/preprocess.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Namespace(align_suffix=None, alignfile=None, all_gather_list_size=16384, azureml_logging=False, bf16=False, bpe=None, cpu=False, criterion='cross_entropy', dataset_impl='mmap', destdir='/n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard10-8', empty_cache_freq=0, fp16=False, fp16_init_scale=128, fp16_no_flatten_grads=False, fp16_scale_tolerance=0.0, fp16_scale_window=None, joined_dictionary=False, log_format=None, log_interval=100, lr_scheduler='fixed', memory_efficient_bf16=False, memory_efficient_fp16=False, min_loss_scale=0.0001, model_parallel_size=1, no_progress_bar=False, nwordssrc=-1, nwordstgt=-1, only_source=True, optimizer=None, padding_factor=8, plasma_path='/tmp/plasma', profile=False, quantization_config_path=None, reset_logging=False, scoring='bleu', seed=1, simul_type=None, slurm_job_id=None, slurm_job_name=None, source_lang=None, srcdict='/n/fs/nlp-awettig/data-pretraining/gpt2_bpe/dict.txt', suppress_crashes=False, tag=None, target_lang=None, task='translation', tensorboard_logdir=None, testpref=None, tgtdict=None, threshold_loss_scale=None, thresholdsrc=0, thresholdtgt=0, tokenizer=None, tpu=False, trainpref='/n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard10-8', use_plasma_view=False, user_dir=None, validpref='/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe,/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe', wandb_project=None, workers=12)
2
+ [None] Dictionary: 50264 types
3
+ [None] /n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard10-8: 18342620 sents, 520181911 tokens, 0.0% replaced by <unk>
4
+ [None] Dictionary: 50264 types
5
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe: 1801350 sents, 117517428 tokens, 0.0% replaced by <unk>
6
+ [None] Dictionary: 50264 types
7
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe: 3760 sents, 246403 tokens, 0.0% replaced by <unk>
8
+ Wrote preprocessed data to /n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard10-8
bin-shard10-8/train.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea5ea80bd19eafa08403203538d3f59ce304c75f7fb7c8dfa99081dc97e66617
3
+ size 1040363822
bin-shard10-8/train.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d2c24ed9804d173e3beecd807ee3ba0e5f1d42a2c4b062faefe7997db675bb0
3
+ size 220111466
bin-shard10-8/valid.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c590981bafb9ee39836011daf94fa23e429d3de88dd1c7a7d6c1699aa8147828
3
+ size 492806
bin-shard10-8/valid.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c730179c16607e30876530e1886e62ef9cdd0daf228f8b86de8d5eb752a91c
3
+ size 45146
bin-shard11-8/dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
bin-shard11-8/preprocess.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Namespace(align_suffix=None, alignfile=None, all_gather_list_size=16384, azureml_logging=False, bf16=False, bpe=None, cpu=False, criterion='cross_entropy', dataset_impl='mmap', destdir='/n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard11-8', empty_cache_freq=0, fp16=False, fp16_init_scale=128, fp16_no_flatten_grads=False, fp16_scale_tolerance=0.0, fp16_scale_window=None, joined_dictionary=False, log_format=None, log_interval=100, lr_scheduler='fixed', memory_efficient_bf16=False, memory_efficient_fp16=False, min_loss_scale=0.0001, model_parallel_size=1, no_progress_bar=False, nwordssrc=-1, nwordstgt=-1, only_source=True, optimizer=None, padding_factor=8, plasma_path='/tmp/plasma', profile=False, quantization_config_path=None, reset_logging=False, scoring='bleu', seed=1, simul_type=None, slurm_job_id=None, slurm_job_name=None, source_lang=None, srcdict='/n/fs/nlp-awettig/data-pretraining/gpt2_bpe/dict.txt', suppress_crashes=False, tag=None, target_lang=None, task='translation', tensorboard_logdir=None, testpref=None, tgtdict=None, threshold_loss_scale=None, thresholdsrc=0, thresholdtgt=0, tokenizer=None, tpu=False, trainpref='/n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard11-8', use_plasma_view=False, user_dir=None, validpref='/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe,/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe', wandb_project=None, workers=12)
2
+ [None] Dictionary: 50264 types
3
+ [None] /n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard11-8: 18925132 sents, 553472931 tokens, 0.0% replaced by <unk>
4
+ [None] Dictionary: 50264 types
5
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe: 1801350 sents, 117517428 tokens, 0.0% replaced by <unk>
6
+ [None] Dictionary: 50264 types
7
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe: 3760 sents, 246403 tokens, 0.0% replaced by <unk>
8
+ Wrote preprocessed data to /n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard11-8
bin-shard11-8/train.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7eb97278a692c3f8d79b8a9486600dcab7cb83e86bcca8a600397c51bf03746f
3
+ size 1106945862
bin-shard11-8/train.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f0c141555c85ef3a2d4e589a1c83d481b268b6e4287dc070f7c3cbd94ddffc0
3
+ size 227101610
bin-shard11-8/valid.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c590981bafb9ee39836011daf94fa23e429d3de88dd1c7a7d6c1699aa8147828
3
+ size 492806
bin-shard11-8/valid.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c730179c16607e30876530e1886e62ef9cdd0daf228f8b86de8d5eb752a91c
3
+ size 45146
bin-shard12-8/dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
bin-shard12-8/preprocess.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Namespace(align_suffix=None, alignfile=None, all_gather_list_size=16384, azureml_logging=False, bf16=False, bpe=None, cpu=False, criterion='cross_entropy', dataset_impl='mmap', destdir='/n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard12-8', empty_cache_freq=0, fp16=False, fp16_init_scale=128, fp16_no_flatten_grads=False, fp16_scale_tolerance=0.0, fp16_scale_window=None, joined_dictionary=False, log_format=None, log_interval=100, lr_scheduler='fixed', memory_efficient_bf16=False, memory_efficient_fp16=False, min_loss_scale=0.0001, model_parallel_size=1, no_progress_bar=False, nwordssrc=-1, nwordstgt=-1, only_source=True, optimizer=None, padding_factor=8, plasma_path='/tmp/plasma', profile=False, quantization_config_path=None, reset_logging=False, scoring='bleu', seed=1, simul_type=None, slurm_job_id=None, slurm_job_name=None, source_lang=None, srcdict='/n/fs/nlp-awettig/data-pretraining/gpt2_bpe/dict.txt', suppress_crashes=False, tag=None, target_lang=None, task='translation', tensorboard_logdir=None, testpref=None, tgtdict=None, threshold_loss_scale=None, thresholdsrc=0, thresholdtgt=0, tokenizer=None, tpu=False, trainpref='/n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard12-8', use_plasma_view=False, user_dir=None, validpref='/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe,/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe', wandb_project=None, workers=12)
2
+ [None] Dictionary: 50264 types
3
+ [None] /n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard12-8: 18138571 sents, 509042782 tokens, 0.0% replaced by <unk>
4
+ [None] Dictionary: 50264 types
5
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe: 1801350 sents, 117517428 tokens, 0.0% replaced by <unk>
6
+ [None] Dictionary: 50264 types
7
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe: 3760 sents, 246403 tokens, 0.0% replaced by <unk>
8
+ Wrote preprocessed data to /n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard12-8
bin-shard12-8/train.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06721237c0310148440da207cfe1ba969c7d07e0e934898473a3cf4691e395b6
3
+ size 1018085564
bin-shard12-8/train.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d375676403816952b37b2fbba65fffeb34ef5b2e3895cb212b9af27b959123d
3
+ size 217662878
bin-shard12-8/valid.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c590981bafb9ee39836011daf94fa23e429d3de88dd1c7a7d6c1699aa8147828
3
+ size 492806
bin-shard12-8/valid.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c730179c16607e30876530e1886e62ef9cdd0daf228f8b86de8d5eb752a91c
3
+ size 45146
bin-shard13-8/dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
bin-shard13-8/preprocess.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Namespace(align_suffix=None, alignfile=None, all_gather_list_size=16384, azureml_logging=False, bf16=False, bpe=None, cpu=False, criterion='cross_entropy', dataset_impl='mmap', destdir='/n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard13-8', empty_cache_freq=0, fp16=False, fp16_init_scale=128, fp16_no_flatten_grads=False, fp16_scale_tolerance=0.0, fp16_scale_window=None, joined_dictionary=False, log_format=None, log_interval=100, lr_scheduler='fixed', memory_efficient_bf16=False, memory_efficient_fp16=False, min_loss_scale=0.0001, model_parallel_size=1, no_progress_bar=False, nwordssrc=-1, nwordstgt=-1, only_source=True, optimizer=None, padding_factor=8, plasma_path='/tmp/plasma', profile=False, quantization_config_path=None, reset_logging=False, scoring='bleu', seed=1, simul_type=None, slurm_job_id=None, slurm_job_name=None, source_lang=None, srcdict='/n/fs/nlp-awettig/data-pretraining/gpt2_bpe/dict.txt', suppress_crashes=False, tag=None, target_lang=None, task='translation', tensorboard_logdir=None, testpref=None, tgtdict=None, threshold_loss_scale=None, thresholdsrc=0, thresholdtgt=0, tokenizer=None, tpu=False, trainpref='/n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard13-8', use_plasma_view=False, user_dir=None, validpref='/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe,/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe', wandb_project=None, workers=12)
2
+ [None] Dictionary: 50264 types
3
+ [None] /n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard13-8: 18108750 sents, 500893800 tokens, 0.0% replaced by <unk>
4
+ [None] Dictionary: 50264 types
5
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe: 1801350 sents, 117517428 tokens, 0.0% replaced by <unk>
6
+ [None] Dictionary: 50264 types
7
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe: 3760 sents, 246403 tokens, 0.0% replaced by <unk>
8
+ Wrote preprocessed data to /n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard13-8
bin-shard13-8/train.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27a2bcd4d79e13964136189389509325ff2ade944ea46cb73b615d7d53d44525
3
+ size 1001787600
bin-shard13-8/train.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84ac14ac9553d64bd83c2542282817ee6459a805a99655fb0234fb1a81fdec53
3
+ size 217305026
bin-shard13-8/valid.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c590981bafb9ee39836011daf94fa23e429d3de88dd1c7a7d6c1699aa8147828
3
+ size 492806
bin-shard13-8/valid.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c730179c16607e30876530e1886e62ef9cdd0daf228f8b86de8d5eb752a91c
3
+ size 45146
bin-shard14-8/dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
bin-shard14-8/preprocess.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Namespace(align_suffix=None, alignfile=None, all_gather_list_size=16384, azureml_logging=False, bf16=False, bpe=None, cpu=False, criterion='cross_entropy', dataset_impl='mmap', destdir='/n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard14-8', empty_cache_freq=0, fp16=False, fp16_init_scale=128, fp16_no_flatten_grads=False, fp16_scale_tolerance=0.0, fp16_scale_window=None, joined_dictionary=False, log_format=None, log_interval=100, lr_scheduler='fixed', memory_efficient_bf16=False, memory_efficient_fp16=False, min_loss_scale=0.0001, model_parallel_size=1, no_progress_bar=False, nwordssrc=-1, nwordstgt=-1, only_source=True, optimizer=None, padding_factor=8, plasma_path='/tmp/plasma', profile=False, quantization_config_path=None, reset_logging=False, scoring='bleu', seed=1, simul_type=None, slurm_job_id=None, slurm_job_name=None, source_lang=None, srcdict='/n/fs/nlp-awettig/data-pretraining/gpt2_bpe/dict.txt', suppress_crashes=False, tag=None, target_lang=None, task='translation', tensorboard_logdir=None, testpref=None, tgtdict=None, threshold_loss_scale=None, thresholdsrc=0, thresholdtgt=0, tokenizer=None, tpu=False, trainpref='/n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard14-8', use_plasma_view=False, user_dir=None, validpref='/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe,/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe', wandb_project=None, workers=12)
2
+ [None] Dictionary: 50264 types
3
+ [None] /n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard14-8: 18280671 sents, 501536992 tokens, 0.0% replaced by <unk>
4
+ [None] Dictionary: 50264 types
5
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe: 1801350 sents, 117517428 tokens, 0.0% replaced by <unk>
6
+ [None] Dictionary: 50264 types
7
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe: 3760 sents, 246403 tokens, 0.0% replaced by <unk>
8
+ Wrote preprocessed data to /n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard14-8
bin-shard14-8/train.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8d11535116b05893b11e731c0eab3f56b87b5871bbeea8aa896c50af3a45480
3
+ size 1003073984
bin-shard14-8/train.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48e742485297c133457e184fba924276c3e48231560b9be426b507039264fe1a
3
+ size 219368078
bin-shard14-8/valid.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c590981bafb9ee39836011daf94fa23e429d3de88dd1c7a7d6c1699aa8147828
3
+ size 492806
bin-shard14-8/valid.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c730179c16607e30876530e1886e62ef9cdd0daf228f8b86de8d5eb752a91c
3
+ size 45146
bin-shard15-8/dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
bin-shard15-8/preprocess.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Namespace(align_suffix=None, alignfile=None, all_gather_list_size=16384, azureml_logging=False, bf16=False, bpe=None, cpu=False, criterion='cross_entropy', dataset_impl='mmap', destdir='/n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard15-8', empty_cache_freq=0, fp16=False, fp16_init_scale=128, fp16_no_flatten_grads=False, fp16_scale_tolerance=0.0, fp16_scale_window=None, joined_dictionary=False, log_format=None, log_interval=100, lr_scheduler='fixed', memory_efficient_bf16=False, memory_efficient_fp16=False, min_loss_scale=0.0001, model_parallel_size=1, no_progress_bar=False, nwordssrc=-1, nwordstgt=-1, only_source=True, optimizer=None, padding_factor=8, plasma_path='/tmp/plasma', profile=False, quantization_config_path=None, reset_logging=False, scoring='bleu', seed=1, simul_type=None, slurm_job_id=None, slurm_job_name=None, source_lang=None, srcdict='/n/fs/nlp-awettig/data-pretraining/gpt2_bpe/dict.txt', suppress_crashes=False, tag=None, target_lang=None, task='translation', tensorboard_logdir=None, testpref=None, tgtdict=None, threshold_loss_scale=None, thresholdsrc=0, thresholdtgt=0, tokenizer=None, tpu=False, trainpref='/n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard15-8', use_plasma_view=False, user_dir=None, validpref='/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe,/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe', wandb_project=None, workers=12)
2
+ [None] Dictionary: 50264 types
3
+ [None] /n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard15-8: 19039993 sents, 531767013 tokens, 0.0% replaced by <unk>
4
+ [None] Dictionary: 50264 types
5
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe: 1801350 sents, 117517428 tokens, 0.0% replaced by <unk>
6
+ [None] Dictionary: 50264 types
7
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe: 3760 sents, 246403 tokens, 0.0% replaced by <unk>
8
+ Wrote preprocessed data to /n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard15-8
bin-shard15-8/train.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f22529f99f3d2357b69cb7080fe71704bf01ba47a400d0a2fe47d0a01a5ec7a
3
+ size 1063534026
bin-shard15-8/train.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e30b4390ca1d000ebfe160c56235b9475c2f7165df9a51b96aeade233d0ec95
3
+ size 228479942
bin-shard15-8/valid.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c590981bafb9ee39836011daf94fa23e429d3de88dd1c7a7d6c1699aa8147828
3
+ size 492806
bin-shard15-8/valid.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c730179c16607e30876530e1886e62ef9cdd0daf228f8b86de8d5eb752a91c
3
+ size 45146
bin-shard2-8/dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
bin-shard2-8/preprocess.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Namespace(align_suffix=None, alignfile=None, all_gather_list_size=16384, azureml_logging=False, bf16=False, bpe=None, cpu=False, criterion='cross_entropy', dataset_impl='mmap', destdir='/n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard2-8', empty_cache_freq=0, fp16=False, fp16_init_scale=128, fp16_no_flatten_grads=False, fp16_scale_tolerance=0.0, fp16_scale_window=None, joined_dictionary=False, log_format=None, log_interval=100, lr_scheduler='fixed', memory_efficient_bf16=False, memory_efficient_fp16=False, min_loss_scale=0.0001, model_parallel_size=1, no_progress_bar=False, nwordssrc=-1, nwordstgt=-1, only_source=True, optimizer=None, padding_factor=8, plasma_path='/tmp/plasma', profile=False, quantization_config_path=None, reset_logging=False, scoring='bleu', seed=1, simul_type=None, slurm_job_id=None, slurm_job_name=None, source_lang=None, srcdict='/n/fs/nlp-awettig/data-pretraining/gpt2_bpe/dict.txt', suppress_crashes=False, tag=None, target_lang=None, task='translation', tensorboard_logdir=None, testpref=None, tgtdict=None, threshold_loss_scale=None, thresholdsrc=0, thresholdtgt=0, tokenizer=None, tpu=False, trainpref='/n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard2-8', use_plasma_view=False, user_dir=None, validpref='/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe,/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe', wandb_project=None, workers=12)
2
+ [None] Dictionary: 50264 types
3
+ [None] /n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard2-8: 18055318 sents, 508219784 tokens, 0.0% replaced by <unk>
4
+ [None] Dictionary: 50264 types
5
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe: 1801350 sents, 117517428 tokens, 0.0% replaced by <unk>
6
+ [None] Dictionary: 50264 types
7
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe: 3760 sents, 246403 tokens, 0.0% replaced by <unk>
8
+ Wrote preprocessed data to /n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard2-8
bin-shard2-8/train.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25e4adc3245eb9cc943ffaa28296b81f84c41e16b4568c63d97038f046b8b48d
3
+ size 1016439568
bin-shard2-8/train.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d978c7563a6447c6559e966a872ebc98aface5ab354623be96a0e39cfe2ae594
3
+ size 216663842
bin-shard2-8/valid.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c590981bafb9ee39836011daf94fa23e429d3de88dd1c7a7d6c1699aa8147828
3
+ size 492806
bin-shard2-8/valid.idx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c730179c16607e30876530e1886e62ef9cdd0daf228f8b86de8d5eb752a91c
3
+ size 45146
bin-shard3-8/dict.txt ADDED
The diff for this file is too large to render. See raw diff
 
bin-shard3-8/preprocess.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ Namespace(align_suffix=None, alignfile=None, all_gather_list_size=16384, azureml_logging=False, bf16=False, bpe=None, cpu=False, criterion='cross_entropy', dataset_impl='mmap', destdir='/n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard3-8', empty_cache_freq=0, fp16=False, fp16_init_scale=128, fp16_no_flatten_grads=False, fp16_scale_tolerance=0.0, fp16_scale_window=None, joined_dictionary=False, log_format=None, log_interval=100, lr_scheduler='fixed', memory_efficient_bf16=False, memory_efficient_fp16=False, min_loss_scale=0.0001, model_parallel_size=1, no_progress_bar=False, nwordssrc=-1, nwordstgt=-1, only_source=True, optimizer=None, padding_factor=8, plasma_path='/tmp/plasma', profile=False, quantization_config_path=None, reset_logging=False, scoring='bleu', seed=1, simul_type=None, slurm_job_id=None, slurm_job_name=None, source_lang=None, srcdict='/n/fs/nlp-awettig/data-pretraining/gpt2_bpe/dict.txt', suppress_crashes=False, tag=None, target_lang=None, task='translation', tensorboard_logdir=None, testpref=None, tgtdict=None, threshold_loss_scale=None, thresholdsrc=0, thresholdtgt=0, tokenizer=None, tpu=False, trainpref='/n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard3-8', use_plasma_view=False, user_dir=None, validpref='/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe,/n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe', wandb_project=None, workers=12)
2
+ [None] Dictionary: 50264 types
3
+ [None] /n/fs/nlp-awettig/data-pretraining/wikibooks/wikibooks.bpe.shard3-8: 18928307 sents, 550719501 tokens, 0.0% replaced by <unk>
4
+ [None] Dictionary: 50264 types
5
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.train.bpe: 1801350 sents, 117517428 tokens, 0.0% replaced by <unk>
6
+ [None] Dictionary: 50264 types
7
+ [None] /n/fs/nlp-awettig/data-pretraining/wikitext-103/wiki.valid.bpe: 3760 sents, 246403 tokens, 0.0% replaced by <unk>
8
+ Wrote preprocessed data to /n/fs/nlp-awettig/data-pretraining/wikibooks/bin-shard3-8