sanchit-gandhi HF staff commited on
Commit
fc53a59
1 Parent(s): 622e8b5

Training in progress, step 500

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. pytorch_model.bin +1 -1
  3. training_args.bin +1 -1
  4. wandb/debug-internal.log +1 -1
  5. wandb/debug.log +1 -1
  6. wandb/latest-run +1 -1
  7. wandb/run-20220302_041332-j5suzd56/files/config.yaml +7 -0
  8. wandb/run-20220302_041332-j5suzd56/files/output.log +7 -0
  9. wandb/run-20220302_041332-j5suzd56/files/wandb-summary.json +0 -0
  10. wandb/run-20220302_041332-j5suzd56/logs/debug-internal.log +114 -0
  11. wandb/run-20220302_041332-j5suzd56/logs/debug.log +114 -0
  12. wandb/run-20220302_041332-j5suzd56/run-j5suzd56.wandb +2 -2
  13. wandb/run-20220302_055556-ymuc7hv0/files/config.yaml +0 -0
  14. wandb/run-20220302_055556-ymuc7hv0/files/output.log +0 -0
  15. wandb/run-20220302_055556-ymuc7hv0/files/requirements.txt +184 -0
  16. wandb/run-20220302_055556-ymuc7hv0/files/wandb-metadata.json +60 -0
  17. wandb/run-20220302_055556-ymuc7hv0/files/wandb-summary.json +0 -0
  18. wandb/run-20220302_055556-ymuc7hv0/logs/debug-internal.log +0 -0
  19. wandb/run-20220302_055556-ymuc7hv0/logs/debug.log +125 -0
  20. wandb/run-20220302_055556-ymuc7hv0/run-ymuc7hv0.wandb +3 -0
  21. wandb/run-20220302_063647-bmivw6vv/files/config.yaml +0 -0
  22. wandb/run-20220302_063647-bmivw6vv/files/output.log +0 -0
  23. wandb/run-20220302_063647-bmivw6vv/files/requirements.txt +184 -0
  24. wandb/run-20220302_063647-bmivw6vv/files/wandb-metadata.json +60 -0
  25. wandb/run-20220302_063647-bmivw6vv/files/wandb-summary.json +0 -0
  26. wandb/run-20220302_063647-bmivw6vv/logs/debug-internal.log +0 -0
  27. wandb/run-20220302_063647-bmivw6vv/logs/debug.log +141 -0
  28. wandb/run-20220302_063647-bmivw6vv/run-bmivw6vv.wandb +3 -0
  29. wandb/run-20220302_074637-35y19oi2/files/config.yaml +0 -0
  30. wandb/run-20220302_074637-35y19oi2/files/output.log +0 -0
  31. wandb/run-20220302_074637-35y19oi2/files/requirements.txt +184 -0
  32. wandb/run-20220302_074637-35y19oi2/files/wandb-metadata.json +60 -0
  33. wandb/run-20220302_074637-35y19oi2/files/wandb-summary.json +0 -0
  34. wandb/run-20220302_074637-35y19oi2/logs/debug-internal.log +0 -0
  35. wandb/run-20220302_074637-35y19oi2/logs/debug.log +125 -0
  36. wandb/run-20220302_074637-35y19oi2/run-35y19oi2.wandb +3 -0
  37. wandb/run-20220302_084833-2m6jtwtj/files/config.yaml +713 -0
  38. wandb/run-20220302_084833-2m6jtwtj/files/output.log +37 -0
  39. wandb/run-20220302_084833-2m6jtwtj/files/requirements.txt +184 -0
  40. wandb/run-20220302_084833-2m6jtwtj/files/wandb-metadata.json +60 -0
  41. wandb/run-20220302_084833-2m6jtwtj/files/wandb-summary.json +1 -0
  42. wandb/run-20220302_084833-2m6jtwtj/logs/debug-internal.log +137 -0
  43. wandb/run-20220302_084833-2m6jtwtj/logs/debug.log +107 -0
  44. wandb/run-20220302_084833-2m6jtwtj/run-2m6jtwtj.wandb +0 -0
  45. wandb/run-20220302_085000-3jbwcde6/files/config.yaml +713 -0
  46. wandb/run-20220302_085000-3jbwcde6/files/output.log +37 -0
  47. wandb/run-20220302_085000-3jbwcde6/files/requirements.txt +184 -0
  48. wandb/run-20220302_085000-3jbwcde6/files/wandb-metadata.json +60 -0
  49. wandb/run-20220302_085000-3jbwcde6/files/wandb-summary.json +1 -0
  50. wandb/run-20220302_085000-3jbwcde6/logs/debug-internal.log +141 -0
.gitattributes CHANGED
@@ -49,3 +49,7 @@ wandb/run-20220301_221232-283qa50u/run-283qa50u.wandb filter=lfs diff=lfs merge=
49
  wandb/run-20220302_000300-2ebwk6gp/run-2ebwk6gp.wandb filter=lfs diff=lfs merge=lfs -text
50
  wandb/run-20220302_021624-vszekdxg/run-vszekdxg.wandb filter=lfs diff=lfs merge=lfs -text
51
  wandb/run-20220302_041332-j5suzd56/run-j5suzd56.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
49
  wandb/run-20220302_000300-2ebwk6gp/run-2ebwk6gp.wandb filter=lfs diff=lfs merge=lfs -text
50
  wandb/run-20220302_021624-vszekdxg/run-vszekdxg.wandb filter=lfs diff=lfs merge=lfs -text
51
  wandb/run-20220302_041332-j5suzd56/run-j5suzd56.wandb filter=lfs diff=lfs merge=lfs -text
52
+ wandb/run-20220302_055556-ymuc7hv0/run-ymuc7hv0.wandb filter=lfs diff=lfs merge=lfs -text
53
+ wandb/run-20220302_063647-bmivw6vv/run-bmivw6vv.wandb filter=lfs diff=lfs merge=lfs -text
54
+ wandb/run-20220302_074637-35y19oi2/run-35y19oi2.wandb filter=lfs diff=lfs merge=lfs -text
55
+ wandb/run-20220302_085255-16llzpbl/run-16llzpbl.wandb filter=lfs diff=lfs merge=lfs -text
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab594e46112a0b3b6c97a2155d0861551020f73b9f2d0f1ce90b232f7eda2e16
3
  size 3210531882
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0231e55487797a42f7a98c2d31b1c557e54ded159d58fb491feed454bb814029
3
  size 3210531882
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e5296214d28c5e8d2d279dbf83f78ee85351aa888d0629a8d317f53222766206
3
  size 3119
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cda31eca71861bfef1fdefa9727c0c3604b9d1cb42ab0a0a66a5b41235b287f
3
  size 3119
wandb/debug-internal.log CHANGED
@@ -1 +1 @@
1
- run-20220302_041332-j5suzd56/logs/debug-internal.log
1
+ run-20220302_085255-16llzpbl/logs/debug-internal.log
wandb/debug.log CHANGED
@@ -1 +1 @@
1
- run-20220302_041332-j5suzd56/logs/debug.log
1
+ run-20220302_085255-16llzpbl/logs/debug.log
wandb/latest-run CHANGED
@@ -1 +1 @@
1
- run-20220302_041332-j5suzd56
1
+ run-20220302_085255-16llzpbl
wandb/run-20220302_041332-j5suzd56/files/config.yaml CHANGED
@@ -10673,7 +10673,14 @@ _wandb:
10673
  - 1
10674
  - 5
10675
  - 11
 
 
 
 
 
10676
  3:
 
 
10677
  - 13
10678
  4: 3.9.5
10679
  5: 0.12.10
10673
  - 1
10674
  - 5
10675
  - 11
10676
+ 2:
10677
+ - 1
10678
+ - 5
10679
+ - 11
10680
+ - 12
10681
  3:
10682
+ - 1
10683
+ - 7
10684
  - 13
10685
  4: 3.9.5
10686
  5: 0.12.10
wandb/run-20220302_041332-j5suzd56/files/output.log CHANGED
@@ -2787,3 +2787,10 @@ Upload file wandb/run-20220302_041332-j5suzd56/run-j5suzd56.wandb: 42%|██
2787
  eval_samples_per_second = 2.812
2788
  eval_steps_per_second = 0.235
2789
  [INFO|modeling_utils.py:1081] 2022-03-02 05:53:57,022 >> Model weights saved in ./pytorch_model.bin:06<08:59, 2.47s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
 
 
 
 
 
 
 
2787
  eval_samples_per_second = 2.812
2788
  eval_steps_per_second = 0.235
2789
  [INFO|modeling_utils.py:1081] 2022-03-02 05:53:57,022 >> Model weights saved in ./pytorch_model.bin:06<08:59, 2.47s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2790
+ Upload file wandb/run-20220302_041332-j5suzd56/run-j5suzd56.wandb: 50%|█████▍ | 17.2M/34.7M [00:01<00:01, 18.0MB/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2791
+ Upload file wandb/run-20220302_041332-j5suzd56/run-j5suzd56.wandb: 100%|███████████| 34.7M/34.7M [00:03<00:00, 12.1MB/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2792
+ Upload file wandb/run-20220302_041332-j5suzd56/run-j5suzd56.wandb: 100%|███████████| 34.7M/34.7M [00:03<00:00, 12.1MB/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2793
+ 03/02/2022 05:54:26 - WARNING - huggingface_hub.repository - To https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search
2794
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2795
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2796
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
wandb/run-20220302_041332-j5suzd56/files/wandb-summary.json CHANGED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_041332-j5suzd56/logs/debug-internal.log CHANGED
@@ -7502,3 +7502,117 @@
7502
  2022-03-02 05:54:18,049 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: stop_status
7503
  2022-03-02 05:54:18,049 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: stop_status
7504
  2022-03-02 05:54:18,804 DEBUG SenderThread:252447 [sender.py:send():235] send: stats
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7502
  2022-03-02 05:54:18,049 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: stop_status
7503
  2022-03-02 05:54:18,049 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: stop_status
7504
  2022-03-02 05:54:18,804 DEBUG SenderThread:252447 [sender.py:send():235] send: stats
7505
+ 2022-03-02 05:54:26,927 INFO Thread-8 :252447 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/output.log
7506
+ 2022-03-02 05:54:28,928 INFO Thread-8 :252447 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/output.log
7507
+ 2022-03-02 05:54:32,833 DEBUG SenderThread:252447 [sender.py:send():235] send: telemetry
7508
+ 2022-03-02 05:54:32,834 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7509
+ 2022-03-02 05:54:32,834 DEBUG SenderThread:252447 [sender.py:send():235] send: exit
7510
+ 2022-03-02 05:54:32,834 INFO SenderThread:252447 [sender.py:send_exit():371] handling exit code: 1
7511
+ 2022-03-02 05:54:32,834 INFO SenderThread:252447 [sender.py:send_exit():373] handling runtime: 6059
7512
+ 2022-03-02 05:54:32,887 INFO SenderThread:252447 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
7513
+ 2022-03-02 05:54:32,888 INFO SenderThread:252447 [sender.py:send_exit():379] send defer
7514
+ 2022-03-02 05:54:32,888 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7515
+ 2022-03-02 05:54:32,888 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7516
+ 2022-03-02 05:54:32,889 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 0
7517
+ 2022-03-02 05:54:32,889 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7518
+ 2022-03-02 05:54:32,889 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 0
7519
+ 2022-03-02 05:54:32,889 INFO SenderThread:252447 [sender.py:transition_state():392] send defer: 1
7520
+ 2022-03-02 05:54:32,889 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7521
+ 2022-03-02 05:54:32,889 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 1
7522
+ 2022-03-02 05:54:32,929 INFO Thread-8 :252447 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/wandb-summary.json
7523
+ 2022-03-02 05:54:32,929 INFO Thread-8 :252447 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/output.log
7524
+ 2022-03-02 05:54:33,004 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7525
+ 2022-03-02 05:54:33,005 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7526
+ 2022-03-02 05:54:33,005 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 1
7527
+ 2022-03-02 05:54:33,005 INFO SenderThread:252447 [sender.py:transition_state():392] send defer: 2
7528
+ 2022-03-02 05:54:33,005 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7529
+ 2022-03-02 05:54:33,005 DEBUG SenderThread:252447 [sender.py:send():235] send: stats
7530
+ 2022-03-02 05:54:33,006 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7531
+ 2022-03-02 05:54:33,006 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 2
7532
+ 2022-03-02 05:54:33,007 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7533
+ 2022-03-02 05:54:33,007 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 2
7534
+ 2022-03-02 05:54:33,007 INFO SenderThread:252447 [sender.py:transition_state():392] send defer: 3
7535
+ 2022-03-02 05:54:33,007 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7536
+ 2022-03-02 05:54:33,007 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 3
7537
+ 2022-03-02 05:54:33,067 DEBUG SenderThread:252447 [sender.py:send():235] send: summary
7538
+ 2022-03-02 05:54:33,133 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7539
+ 2022-03-02 05:54:33,195 INFO SenderThread:252447 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
7540
+ 2022-03-02 05:54:33,195 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7541
+ 2022-03-02 05:54:33,195 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 3
7542
+ 2022-03-02 05:54:33,195 INFO SenderThread:252447 [sender.py:transition_state():392] send defer: 4
7543
+ 2022-03-02 05:54:33,195 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7544
+ 2022-03-02 05:54:33,196 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7545
+ 2022-03-02 05:54:33,196 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 4
7546
+ 2022-03-02 05:54:33,196 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7547
+ 2022-03-02 05:54:33,196 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 4
7548
+ 2022-03-02 05:54:33,297 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7549
+ 2022-03-02 05:54:33,982 INFO Thread-8 :252447 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/wandb-summary.json
7550
+ 2022-03-02 05:54:33,983 INFO SenderThread:252447 [sender.py:transition_state():392] send defer: 5
7551
+ 2022-03-02 05:54:33,983 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7552
+ 2022-03-02 05:54:33,984 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7553
+ 2022-03-02 05:54:33,984 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 5
7554
+ 2022-03-02 05:54:33,984 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7555
+ 2022-03-02 05:54:33,984 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 5
7556
+ 2022-03-02 05:54:33,984 INFO SenderThread:252447 [dir_watcher.py:finish():283] shutting down directory watcher
7557
+ 2022-03-02 05:54:34,085 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7558
+ 2022-03-02 05:54:34,982 INFO Thread-8 :252447 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/config.yaml
7559
+ 2022-03-02 05:54:34,983 INFO SenderThread:252447 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files
7560
+ 2022-03-02 05:54:34,983 INFO SenderThread:252447 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/wandb-metadata.json wandb-metadata.json
7561
+ 2022-03-02 05:54:34,983 INFO SenderThread:252447 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/output.log output.log
7562
+ 2022-03-02 05:54:34,983 INFO SenderThread:252447 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/wandb-summary.json wandb-summary.json
7563
+ 2022-03-02 05:54:34,984 INFO SenderThread:252447 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/requirements.txt requirements.txt
7564
+ 2022-03-02 05:54:34,984 INFO SenderThread:252447 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/config.yaml config.yaml
7565
+ 2022-03-02 05:54:34,984 INFO SenderThread:252447 [sender.py:transition_state():392] send defer: 6
7566
+ 2022-03-02 05:54:34,984 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7567
+ 2022-03-02 05:54:34,990 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7568
+ 2022-03-02 05:54:34,991 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 6
7569
+ 2022-03-02 05:54:34,994 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7570
+ 2022-03-02 05:54:34,994 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 6
7571
+ 2022-03-02 05:54:34,994 INFO SenderThread:252447 [file_pusher.py:finish():177] shutting down file pusher
7572
+ 2022-03-02 05:54:35,093 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7573
+ 2022-03-02 05:54:35,095 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7574
+ 2022-03-02 05:54:35,196 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7575
+ 2022-03-02 05:54:35,196 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7576
+ 2022-03-02 05:54:35,288 INFO Thread-12 :252447 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/output.log
7577
+ 2022-03-02 05:54:35,298 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7578
+ 2022-03-02 05:54:35,298 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7579
+ 2022-03-02 05:54:35,328 INFO Thread-14 :252447 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/requirements.txt
7580
+ 2022-03-02 05:54:35,336 INFO Thread-13 :252447 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/wandb-summary.json
7581
+ 2022-03-02 05:54:35,399 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7582
+ 2022-03-02 05:54:35,399 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7583
+ 2022-03-02 05:54:35,416 INFO Thread-15 :252447 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/files/config.yaml
7584
+ 2022-03-02 05:54:35,501 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7585
+ 2022-03-02 05:54:35,501 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7586
+ 2022-03-02 05:54:35,602 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7587
+ 2022-03-02 05:54:35,602 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7588
+ 2022-03-02 05:54:35,617 INFO Thread-7 :252447 [sender.py:transition_state():392] send defer: 7
7589
+ 2022-03-02 05:54:35,617 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7590
+ 2022-03-02 05:54:35,617 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 7
7591
+ 2022-03-02 05:54:35,618 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7592
+ 2022-03-02 05:54:35,618 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 7
7593
+ 2022-03-02 05:54:35,703 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7594
+ 2022-03-02 05:54:36,916 INFO SenderThread:252447 [sender.py:transition_state():392] send defer: 8
7595
+ 2022-03-02 05:54:36,916 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7596
+ 2022-03-02 05:54:36,917 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7597
+ 2022-03-02 05:54:36,917 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 8
7598
+ 2022-03-02 05:54:36,917 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7599
+ 2022-03-02 05:54:36,917 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 8
7600
+ 2022-03-02 05:54:36,917 INFO SenderThread:252447 [sender.py:transition_state():392] send defer: 9
7601
+ 2022-03-02 05:54:36,918 DEBUG SenderThread:252447 [sender.py:send():235] send: final
7602
+ 2022-03-02 05:54:36,919 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: defer
7603
+ 2022-03-02 05:54:36,920 DEBUG SenderThread:252447 [sender.py:send():235] send: footer
7604
+ 2022-03-02 05:54:36,920 INFO HandlerThread:252447 [handler.py:handle_request_defer():154] handle defer: 9
7605
+ 2022-03-02 05:54:36,920 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: defer
7606
+ 2022-03-02 05:54:36,920 INFO SenderThread:252447 [sender.py:send_request_defer():388] handle sender defer: 9
7607
+ 2022-03-02 05:54:37,018 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: poll_exit
7608
+ 2022-03-02 05:54:37,018 DEBUG SenderThread:252447 [sender.py:send_request():249] send_request: poll_exit
7609
+ 2022-03-02 05:54:37,018 INFO SenderThread:252447 [file_pusher.py:join():182] waiting for file pusher
7610
+ 2022-03-02 05:54:37,081 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: get_summary
7611
+ 2022-03-02 05:54:37,184 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: sampled_history
7612
+ 2022-03-02 05:54:37,187 DEBUG HandlerThread:252447 [handler.py:handle_request():131] handle_request: shutdown
7613
+ 2022-03-02 05:54:37,187 INFO HandlerThread:252447 [handler.py:finish():739] shutting down handler
7614
+ 2022-03-02 05:54:37,919 INFO WriterThread:252447 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_041332-j5suzd56/run-j5suzd56.wandb
7615
+ 2022-03-02 05:54:38,080 INFO SenderThread:252447 [sender.py:finish():1075] shutting down sender
7616
+ 2022-03-02 05:54:38,080 INFO SenderThread:252447 [file_pusher.py:finish():177] shutting down file pusher
7617
+ 2022-03-02 05:54:38,080 INFO SenderThread:252447 [file_pusher.py:join():182] waiting for file pusher
7618
+ 2022-03-02 05:54:38,087 INFO MainThread:252447 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220302_041332-j5suzd56/logs/debug.log CHANGED
@@ -25,3 +25,117 @@ config: {}
25
  2022-03-02 04:13:33,866 INFO MainThread:252348 [wandb_init.py:init():651] run started, returning control to user process
26
  2022-03-02 04:13:33,868 INFO MainThread:252348 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_04-12-50_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 12, 'eval_batch_size': 12}
27
  2022-03-02 04:13:33,872 INFO MainThread:252348 [wandb_watch.py:watch():43] Watching
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  2022-03-02 04:13:33,866 INFO MainThread:252348 [wandb_init.py:init():651] run started, returning control to user process
26
  2022-03-02 04:13:33,868 INFO MainThread:252348 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_04-12-50_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 12, 'eval_batch_size': 12}
27
  2022-03-02 04:13:33,872 INFO MainThread:252348 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-02 05:54:30,324 INFO MainThread:252348 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-02 05:54:30,327 INFO MainThread:252348 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-02 05:54:32,889 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-02 05:54:33,006 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-02 05:54:33,196 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 2095
52
+ }
53
+
54
+ 2022-03-02 05:54:33,984 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 1
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2095
59
+ total_bytes: 2095
60
+ }
61
+
62
+ 2022-03-02 05:54:34,991 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 3
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 2095
67
+ total_bytes: 2391419
68
+ }
69
+
70
+ 2022-03-02 05:54:35,095 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 914810
75
+ total_bytes: 2692823
76
+ }
77
+
78
+ 2022-03-02 05:54:35,197 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 2692823
83
+ total_bytes: 2692823
84
+ }
85
+
86
+ 2022-03-02 05:54:35,298 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 2692823
91
+ total_bytes: 2692823
92
+ }
93
+
94
+ 2022-03-02 05:54:35,400 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 2692823
99
+ total_bytes: 2692823
100
+ }
101
+
102
+ 2022-03-02 05:54:35,501 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 2692823
107
+ total_bytes: 2692823
108
+ }
109
+
110
+ 2022-03-02 05:54:35,603 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
111
+ wandb_count: 5
112
+ }
113
+ pusher_stats {
114
+ uploaded_bytes: 2692823
115
+ total_bytes: 2692823
116
+ }
117
+
118
+ 2022-03-02 05:54:36,917 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
119
+ wandb_count: 5
120
+ }
121
+ pusher_stats {
122
+ uploaded_bytes: 2692823
123
+ total_bytes: 2692823
124
+ }
125
+
126
+ 2022-03-02 05:54:37,080 INFO MainThread:252348 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
127
+ exit_result {
128
+ }
129
+ file_counts {
130
+ wandb_count: 5
131
+ }
132
+ pusher_stats {
133
+ uploaded_bytes: 2692823
134
+ total_bytes: 2692823
135
+ }
136
+ local_info {
137
+ }
138
+
139
+ 2022-03-02 05:54:38,224 INFO MainThread:252348 [wandb_run.py:_append_history():2144] rendering history
140
+ 2022-03-02 05:54:38,225 INFO MainThread:252348 [wandb_run.py:_append_summary():2102] rendering summary
141
+ 2022-03-02 05:54:38,225 INFO MainThread:252348 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220302_041332-j5suzd56/run-j5suzd56.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:229bc5195c54ed39017b8e9aa44f1db2dd40c9060bf0579751030f07deb5b4d1
3
- size 36421026
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d110957484b68080ee103fb92c345f9656f34b2deb75d1df2b764632ebaf12a
3
+ size 37940125
wandb/run-20220302_055556-ymuc7hv0/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_055556-ymuc7hv0/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_055556-ymuc7hv0/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220302_055556-ymuc7hv0/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-02T05:55:57.281359",
5
+ "startedAt": "2022-03-02T05:55:56.178653",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=2",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "622e8b594e8af8169be8cfe538228e49ae08c59d"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220302_055556-ymuc7hv0/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_055556-ymuc7hv0/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_055556-ymuc7hv0/logs/debug.log ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-02 05:55:56,180 INFO MainThread:253209 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-02 05:55:56,180 INFO MainThread:253209 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-02 05:55:56,180 INFO MainThread:253209 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-02 05:55:56,180 INFO MainThread:253209 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-02 05:55:56,180 INFO MainThread:253209 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_055556-ymuc7hv0/logs/debug.log
6
+ 2022-03-02 05:55:56,180 INFO MainThread:253209 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_055556-ymuc7hv0/logs/debug-internal.log
7
+ 2022-03-02 05:55:56,180 INFO MainThread:253209 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-02 05:55:56,181 INFO MainThread:253209 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-02 05:55:56,181 INFO MainThread:253209 [wandb_init.py:init():471] starting backend
11
+ 2022-03-02 05:55:56,181 INFO MainThread:253209 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-02 05:55:56,237 INFO MainThread:253209 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-02 05:55:56,292 INFO MainThread:253209 [backend.py:ensure_launched():224] started backend process with pid: 253308
14
+ 2022-03-02 05:55:56,294 INFO MainThread:253209 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-02 05:55:56,304 INFO MainThread:253209 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-02 05:55:56,436 INFO MainThread:253209 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-02 05:55:57,175 INFO MainThread:253209 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.11 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
18
+
19
+ 2022-03-02 05:55:57,176 INFO MainThread:253209 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
20
+ 2022-03-02 05:55:57,276 INFO MainThread:253209 [wandb_init.py:init():624] starting run threads in backend
21
+ 2022-03-02 05:55:57,386 INFO MainThread:253209 [wandb_run.py:_console_start():1827] atexit reg
22
+ 2022-03-02 05:55:57,386 INFO MainThread:253209 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
23
+ 2022-03-02 05:55:57,387 INFO MainThread:253209 [wandb_run.py:_redirect():1706] Redirecting console.
24
+ 2022-03-02 05:55:57,388 INFO MainThread:253209 [wandb_run.py:_redirect():1762] Redirects installed.
25
+ 2022-03-02 05:55:57,389 INFO MainThread:253209 [wandb_init.py:init():651] run started, returning control to user process
26
+ 2022-03-02 05:55:57,391 INFO MainThread:253209 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_05-55-14_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
27
+ 2022-03-02 05:55:57,394 INFO MainThread:253209 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-02 06:35:23,000 INFO MainThread:253209 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-02 06:35:23,001 INFO MainThread:253209 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-02 06:35:25,887 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-02 06:35:26,060 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-02 06:35:26,883 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 2095
52
+ }
53
+
54
+ 2022-03-02 06:35:27,673 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 4
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2095
59
+ total_bytes: 1796625
60
+ }
61
+
62
+ 2022-03-02 06:35:27,776 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 360250
67
+ total_bytes: 2094199
68
+ }
69
+
70
+ 2022-03-02 06:35:27,878 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 2094199
75
+ total_bytes: 2094199
76
+ }
77
+
78
+ 2022-03-02 06:35:27,980 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 2094199
83
+ total_bytes: 2094199
84
+ }
85
+
86
+ 2022-03-02 06:35:28,081 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 2094199
91
+ total_bytes: 2094199
92
+ }
93
+
94
+ 2022-03-02 06:35:28,183 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 2094199
99
+ total_bytes: 2094199
100
+ }
101
+
102
+ 2022-03-02 06:35:29,490 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 2094199
107
+ total_bytes: 2094199
108
+ }
109
+
110
+ 2022-03-02 06:35:29,647 INFO MainThread:253209 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
111
+ exit_result {
112
+ }
113
+ file_counts {
114
+ wandb_count: 5
115
+ }
116
+ pusher_stats {
117
+ uploaded_bytes: 2094199
118
+ total_bytes: 2094199
119
+ }
120
+ local_info {
121
+ }
122
+
123
+ 2022-03-02 06:35:30,792 INFO MainThread:253209 [wandb_run.py:_append_history():2144] rendering history
124
+ 2022-03-02 06:35:30,792 INFO MainThread:253209 [wandb_run.py:_append_summary():2102] rendering summary
125
+ 2022-03-02 06:35:30,793 INFO MainThread:253209 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220302_055556-ymuc7hv0/run-ymuc7hv0.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7d9238c321667ddae0d1b24fc1a82ac5c8fc5729064808fc88ea560fe53eb29
3
+ size 17128618
wandb/run-20220302_063647-bmivw6vv/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_063647-bmivw6vv/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_063647-bmivw6vv/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220302_063647-bmivw6vv/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-02T06:36:48.207194",
5
+ "startedAt": "2022-03-02T06:36:47.139630",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=4",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "622e8b594e8af8169be8cfe538228e49ae08c59d"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220302_063647-bmivw6vv/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_063647-bmivw6vv/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_063647-bmivw6vv/logs/debug.log ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_063647-bmivw6vv/logs/debug.log
6
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_063647-bmivw6vv/logs/debug-internal.log
7
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [wandb_init.py:init():471] starting backend
11
+ 2022-03-02 06:36:47,141 INFO MainThread:253446 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-02 06:36:47,197 INFO MainThread:253446 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-02 06:36:47,251 INFO MainThread:253446 [backend.py:ensure_launched():224] started backend process with pid: 253545
14
+ 2022-03-02 06:36:47,254 INFO MainThread:253446 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-02 06:36:47,263 INFO MainThread:253446 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-02 06:36:47,391 INFO MainThread:253446 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-02 06:36:48,104 INFO MainThread:253446 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.11 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
18
+
19
+ 2022-03-02 06:36:48,104 INFO MainThread:253446 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
20
+ 2022-03-02 06:36:48,202 INFO MainThread:253446 [wandb_init.py:init():624] starting run threads in backend
21
+ 2022-03-02 06:36:48,309 INFO MainThread:253446 [wandb_run.py:_console_start():1827] atexit reg
22
+ 2022-03-02 06:36:48,309 INFO MainThread:253446 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
23
+ 2022-03-02 06:36:48,310 INFO MainThread:253446 [wandb_run.py:_redirect():1706] Redirecting console.
24
+ 2022-03-02 06:36:48,312 INFO MainThread:253446 [wandb_run.py:_redirect():1762] Redirects installed.
25
+ 2022-03-02 06:36:48,312 INFO MainThread:253446 [wandb_init.py:init():651] run started, returning control to user process
26
+ 2022-03-02 06:36:48,314 INFO MainThread:253446 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_06-36-06_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
27
+ 2022-03-02 06:36:48,317 INFO MainThread:253446 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-02 07:45:11,103 INFO MainThread:253446 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-02 07:45:11,105 INFO MainThread:253446 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-02 07:45:13,531 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-02 07:45:13,648 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-02 07:45:13,791 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 2095
52
+ }
53
+
54
+ 2022-03-02 07:45:15,313 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 1
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2095
59
+ total_bytes: 2095
60
+ }
61
+
62
+ 2022-03-02 07:45:16,230 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 2095
67
+ total_bytes: 2345838
68
+ }
69
+
70
+ 2022-03-02 07:45:16,332 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 889444
75
+ total_bytes: 2345838
76
+ }
77
+
78
+ 2022-03-02 07:45:16,433 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 2345838
83
+ total_bytes: 2345838
84
+ }
85
+
86
+ 2022-03-02 07:45:16,535 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 2345838
91
+ total_bytes: 2345838
92
+ }
93
+
94
+ 2022-03-02 07:45:16,636 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 2345838
99
+ total_bytes: 2345838
100
+ }
101
+
102
+ 2022-03-02 07:45:16,738 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 2345838
107
+ total_bytes: 2345838
108
+ }
109
+
110
+ 2022-03-02 07:45:16,839 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
111
+ wandb_count: 5
112
+ }
113
+ pusher_stats {
114
+ uploaded_bytes: 2345838
115
+ total_bytes: 2345838
116
+ }
117
+
118
+ 2022-03-02 07:45:18,394 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
119
+ wandb_count: 5
120
+ }
121
+ pusher_stats {
122
+ uploaded_bytes: 2345838
123
+ total_bytes: 2345838
124
+ }
125
+
126
+ 2022-03-02 07:45:18,557 INFO MainThread:253446 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
127
+ exit_result {
128
+ }
129
+ file_counts {
130
+ wandb_count: 5
131
+ }
132
+ pusher_stats {
133
+ uploaded_bytes: 2345838
134
+ total_bytes: 2345838
135
+ }
136
+ local_info {
137
+ }
138
+
139
+ 2022-03-02 07:45:19,706 INFO MainThread:253446 [wandb_run.py:_append_history():2144] rendering history
140
+ 2022-03-02 07:45:19,707 INFO MainThread:253446 [wandb_run.py:_append_summary():2102] rendering summary
141
+ 2022-03-02 07:45:19,708 INFO MainThread:253446 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220302_063647-bmivw6vv/run-bmivw6vv.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95a27d78f42670f0237b947ed81744c2bc569e3ecc5ba6f04e6ff4143e1036d1
3
+ size 29743466
wandb/run-20220302_074637-35y19oi2/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_074637-35y19oi2/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_074637-35y19oi2/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220302_074637-35y19oi2/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-02T07:46:38.278448",
5
+ "startedAt": "2022-03-02T07:46:37.113392",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=8",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "622e8b594e8af8169be8cfe538228e49ae08c59d"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220302_074637-35y19oi2/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_074637-35y19oi2/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_074637-35y19oi2/logs/debug.log ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_074637-35y19oi2/logs/debug.log
6
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_074637-35y19oi2/logs/debug-internal.log
7
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [wandb_init.py:init():471] starting backend
11
+ 2022-03-02 07:46:37,115 INFO MainThread:253766 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-02 07:46:37,171 INFO MainThread:253766 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-02 07:46:37,226 INFO MainThread:253766 [backend.py:ensure_launched():224] started backend process with pid: 253865
14
+ 2022-03-02 07:46:37,228 INFO MainThread:253766 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-02 07:46:37,239 INFO MainThread:253766 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-02 07:46:37,370 INFO MainThread:253766 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-02 07:46:38,111 INFO MainThread:253766 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.11 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
18
+
19
+ 2022-03-02 07:46:38,112 INFO MainThread:253766 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
20
+ 2022-03-02 07:46:38,273 INFO MainThread:253766 [wandb_init.py:init():624] starting run threads in backend
21
+ 2022-03-02 07:46:38,384 INFO MainThread:253766 [wandb_run.py:_console_start():1827] atexit reg
22
+ 2022-03-02 07:46:38,385 INFO MainThread:253766 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
23
+ 2022-03-02 07:46:38,385 INFO MainThread:253766 [wandb_run.py:_redirect():1706] Redirecting console.
24
+ 2022-03-02 07:46:38,387 INFO MainThread:253766 [wandb_run.py:_redirect():1762] Redirects installed.
25
+ 2022-03-02 07:46:38,387 INFO MainThread:253766 [wandb_init.py:init():651] run started, returning control to user process
26
+ 2022-03-02 07:46:38,390 INFO MainThread:253766 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_07-45-55_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
27
+ 2022-03-02 07:46:38,393 INFO MainThread:253766 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-02 08:47:07,035 INFO MainThread:253766 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-02 08:47:07,036 INFO MainThread:253766 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-02 08:47:09,429 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-02 08:47:09,599 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-02 08:47:10,414 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 2095
52
+ }
53
+
54
+ 2022-03-02 08:47:11,074 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 3
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2095
59
+ total_bytes: 1985562
60
+ }
61
+
62
+ 2022-03-02 08:47:11,178 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 2095
67
+ total_bytes: 2286373
68
+ }
69
+
70
+ 2022-03-02 08:47:11,280 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 2286373
75
+ total_bytes: 2286373
76
+ }
77
+
78
+ 2022-03-02 08:47:11,381 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 2286373
83
+ total_bytes: 2286373
84
+ }
85
+
86
+ 2022-03-02 08:47:11,483 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 2286373
91
+ total_bytes: 2286373
92
+ }
93
+
94
+ 2022-03-02 08:47:11,584 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 2286373
99
+ total_bytes: 2286373
100
+ }
101
+
102
+ 2022-03-02 08:47:13,407 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 2286373
107
+ total_bytes: 2286373
108
+ }
109
+
110
+ 2022-03-02 08:47:13,567 INFO MainThread:253766 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
111
+ exit_result {
112
+ }
113
+ file_counts {
114
+ wandb_count: 5
115
+ }
116
+ pusher_stats {
117
+ uploaded_bytes: 2286373
118
+ total_bytes: 2286373
119
+ }
120
+ local_info {
121
+ }
122
+
123
+ 2022-03-02 08:47:14,706 INFO MainThread:253766 [wandb_run.py:_append_history():2144] rendering history
124
+ 2022-03-02 08:47:14,706 INFO MainThread:253766 [wandb_run.py:_append_summary():2102] rendering summary
125
+ 2022-03-02 08:47:14,708 INFO MainThread:253766 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220302_074637-35y19oi2/run-35y19oi2.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5216f35956d3c59ea78d23e16d6b60fe893790ba50a61224a8a63df7d9957bd
3
+ size 26782320
wandb/run-20220302_084833-2m6jtwtj/files/config.yaml ADDED
@@ -0,0 +1,713 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: ./
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.10
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.9.5
22
+ start_time: 1646210913
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.9.5
37
+ 5: 0.12.10
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ adafactor:
42
+ desc: null
43
+ value: false
44
+ adam_beta1:
45
+ desc: null
46
+ value: 0.9
47
+ adam_beta2:
48
+ desc: null
49
+ value: 0.999
50
+ adam_epsilon:
51
+ desc: null
52
+ value: 1.0e-08
53
+ add_cross_attention:
54
+ desc: null
55
+ value: false
56
+ architectures:
57
+ desc: null
58
+ value:
59
+ - SpeechEncoderDecoderModel
60
+ bad_words_ids:
61
+ desc: null
62
+ value: null
63
+ bf16:
64
+ desc: null
65
+ value: false
66
+ bf16_full_eval:
67
+ desc: null
68
+ value: false
69
+ bos_token_id:
70
+ desc: null
71
+ value: null
72
+ chunk_size_feed_forward:
73
+ desc: null
74
+ value: 0
75
+ cross_attention_hidden_size:
76
+ desc: null
77
+ value: null
78
+ dataloader_drop_last:
79
+ desc: null
80
+ value: false
81
+ dataloader_num_workers:
82
+ desc: null
83
+ value: 0
84
+ dataloader_pin_memory:
85
+ desc: null
86
+ value: true
87
+ ddp_bucket_cap_mb:
88
+ desc: null
89
+ value: None
90
+ ddp_find_unused_parameters:
91
+ desc: null
92
+ value: None
93
+ debug:
94
+ desc: null
95
+ value: '[]'
96
+ decoder:
97
+ desc: null
98
+ value:
99
+ _name_or_path: gpt2-medium
100
+ activation_function: gelu_new
101
+ add_cross_attention: true
102
+ architectures:
103
+ - GPT2LMHeadModel
104
+ attn_pdrop: 0.0
105
+ bad_words_ids: null
106
+ bos_token_id: 50256
107
+ chunk_size_feed_forward: 0
108
+ cross_attention_hidden_size: null
109
+ decoder_start_token_id: null
110
+ diversity_penalty: 0.0
111
+ do_sample: false
112
+ early_stopping: false
113
+ embd_pdrop: 0.0
114
+ encoder_no_repeat_ngram_size: 0
115
+ eos_token_id: 50256
116
+ finetuning_task: null
117
+ forced_bos_token_id: null
118
+ forced_eos_token_id: null
119
+ id2label:
120
+ '0': LABEL_0
121
+ '1': LABEL_1
122
+ initializer_range: 0.02
123
+ is_decoder: true
124
+ is_encoder_decoder: false
125
+ label2id:
126
+ LABEL_0: 0
127
+ LABEL_1: 1
128
+ layer_norm_epsilon: 1.0e-05
129
+ length_penalty: 1.0
130
+ max_length: 20
131
+ min_length: 0
132
+ model_type: gpt2
133
+ n_ctx: 1024
134
+ n_embd: 1024
135
+ n_head: 16
136
+ n_inner: null
137
+ n_layer: 24
138
+ n_positions: 1024
139
+ n_special: 0
140
+ no_repeat_ngram_size: 0
141
+ num_beam_groups: 1
142
+ num_beams: 1
143
+ num_return_sequences: 1
144
+ output_attentions: false
145
+ output_hidden_states: false
146
+ output_scores: false
147
+ pad_token_id: null
148
+ predict_special_tokens: true
149
+ prefix: null
150
+ problem_type: null
151
+ pruned_heads: {}
152
+ remove_invalid_values: false
153
+ reorder_and_upcast_attn: false
154
+ repetition_penalty: 1.0
155
+ resid_pdrop: 0.0
156
+ return_dict: true
157
+ return_dict_in_generate: false
158
+ scale_attn_by_inverse_layer_idx: false
159
+ scale_attn_weights: true
160
+ sep_token_id: null
161
+ summary_activation: null
162
+ summary_first_dropout: 0.0
163
+ summary_proj_to_labels: true
164
+ summary_type: cls_index
165
+ summary_use_proj: true
166
+ task_specific_params:
167
+ text-generation:
168
+ do_sample: true
169
+ max_length: 50
170
+ temperature: 1.0
171
+ tie_encoder_decoder: false
172
+ tie_word_embeddings: true
173
+ tokenizer_class: null
174
+ top_k: 50
175
+ top_p: 1.0
176
+ torch_dtype: null
177
+ torchscript: false
178
+ transformers_version: 4.17.0.dev0
179
+ use_bfloat16: false
180
+ use_cache: false
181
+ vocab_size: 50257
182
+ decoder_start_token_id:
183
+ desc: null
184
+ value: 50256
185
+ deepspeed:
186
+ desc: null
187
+ value: None
188
+ disable_tqdm:
189
+ desc: null
190
+ value: false
191
+ diversity_penalty:
192
+ desc: null
193
+ value: 0.0
194
+ do_eval:
195
+ desc: null
196
+ value: true
197
+ do_predict:
198
+ desc: null
199
+ value: false
200
+ do_sample:
201
+ desc: null
202
+ value: false
203
+ do_train:
204
+ desc: null
205
+ value: true
206
+ early_stopping:
207
+ desc: null
208
+ value: false
209
+ encoder:
210
+ desc: null
211
+ value:
212
+ _name_or_path: facebook/wav2vec2-large-lv60
213
+ activation_dropout: 0.0
214
+ adapter_kernel_size: 3
215
+ adapter_stride: 2
216
+ add_adapter: true
217
+ add_cross_attention: false
218
+ apply_spec_augment: false
219
+ architectures:
220
+ - Wav2Vec2ForPreTraining
221
+ attention_dropout: 0.0
222
+ bad_words_ids: null
223
+ bos_token_id: 1
224
+ chunk_size_feed_forward: 0
225
+ classifier_proj_size: 256
226
+ codevector_dim: 768
227
+ contrastive_logits_temperature: 0.1
228
+ conv_bias: true
229
+ conv_dim:
230
+ - 512
231
+ - 512
232
+ - 512
233
+ - 512
234
+ - 512
235
+ - 512
236
+ - 512
237
+ conv_kernel:
238
+ - 10
239
+ - 3
240
+ - 3
241
+ - 3
242
+ - 3
243
+ - 2
244
+ - 2
245
+ conv_stride:
246
+ - 5
247
+ - 2
248
+ - 2
249
+ - 2
250
+ - 2
251
+ - 2
252
+ - 2
253
+ cross_attention_hidden_size: null
254
+ ctc_loss_reduction: sum
255
+ ctc_zero_infinity: false
256
+ decoder_start_token_id: null
257
+ diversity_loss_weight: 0.1
258
+ diversity_penalty: 0.0
259
+ do_sample: false
260
+ do_stable_layer_norm: true
261
+ early_stopping: false
262
+ encoder_no_repeat_ngram_size: 0
263
+ eos_token_id: 2
264
+ feat_extract_activation: gelu
265
+ feat_extract_dropout: 0.0
266
+ feat_extract_norm: layer
267
+ feat_proj_dropout: 0.0
268
+ feat_quantizer_dropout: 0.0
269
+ final_dropout: 0.0
270
+ finetuning_task: null
271
+ forced_bos_token_id: null
272
+ forced_eos_token_id: null
273
+ gradient_checkpointing: false
274
+ hidden_act: gelu
275
+ hidden_dropout: 0.0
276
+ hidden_dropout_prob: 0.0
277
+ hidden_size: 1024
278
+ id2label:
279
+ '0': LABEL_0
280
+ '1': LABEL_1
281
+ initializer_range: 0.02
282
+ intermediate_size: 4096
283
+ is_decoder: false
284
+ is_encoder_decoder: false
285
+ label2id:
286
+ LABEL_0: 0
287
+ LABEL_1: 1
288
+ layer_norm_eps: 1.0e-05
289
+ layerdrop: 0.0
290
+ length_penalty: 1.0
291
+ mask_feature_length: 10
292
+ mask_feature_min_masks: 0
293
+ mask_feature_prob: 0.0
294
+ mask_time_length: 10
295
+ mask_time_min_masks: 2
296
+ mask_time_prob: 0.0
297
+ max_length: 20
298
+ min_length: 0
299
+ model_type: wav2vec2
300
+ no_repeat_ngram_size: 0
301
+ num_adapter_layers: 3
302
+ num_attention_heads: 16
303
+ num_beam_groups: 1
304
+ num_beams: 1
305
+ num_codevector_groups: 2
306
+ num_codevectors_per_group: 320
307
+ num_conv_pos_embedding_groups: 16
308
+ num_conv_pos_embeddings: 128
309
+ num_feat_extract_layers: 7
310
+ num_hidden_layers: 24
311
+ num_negatives: 100
312
+ num_return_sequences: 1
313
+ output_attentions: false
314
+ output_hidden_size: 1024
315
+ output_hidden_states: false
316
+ output_scores: false
317
+ pad_token_id: 0
318
+ prefix: null
319
+ problem_type: null
320
+ proj_codevector_dim: 768
321
+ pruned_heads: {}
322
+ remove_invalid_values: false
323
+ repetition_penalty: 1.0
324
+ return_dict: true
325
+ return_dict_in_generate: false
326
+ sep_token_id: null
327
+ task_specific_params: null
328
+ tdnn_dilation:
329
+ - 1
330
+ - 2
331
+ - 3
332
+ - 1
333
+ - 1
334
+ tdnn_dim:
335
+ - 512
336
+ - 512
337
+ - 512
338
+ - 512
339
+ - 1500
340
+ tdnn_kernel:
341
+ - 5
342
+ - 3
343
+ - 3
344
+ - 1
345
+ - 1
346
+ temperature: 1.0
347
+ tie_encoder_decoder: false
348
+ tie_word_embeddings: true
349
+ tokenizer_class: null
350
+ top_k: 50
351
+ top_p: 1.0
352
+ torch_dtype: null
353
+ torchscript: false
354
+ transformers_version: 4.17.0.dev0
355
+ use_bfloat16: false
356
+ use_weighted_layer_sum: false
357
+ vocab_size: 32
358
+ xvector_output_dim: 512
359
+ encoder_no_repeat_ngram_size:
360
+ desc: null
361
+ value: 0
362
+ eos_token_id:
363
+ desc: null
364
+ value: 50256
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: None
368
+ eval_batch_size:
369
+ desc: null
370
+ value: 16
371
+ eval_steps:
372
+ desc: null
373
+ value: 500
374
+ evaluation_strategy:
375
+ desc: null
376
+ value: steps
377
+ finetuning_task:
378
+ desc: null
379
+ value: null
380
+ forced_bos_token_id:
381
+ desc: null
382
+ value: null
383
+ forced_eos_token_id:
384
+ desc: null
385
+ value: null
386
+ fp16:
387
+ desc: null
388
+ value: true
389
+ fp16_backend:
390
+ desc: null
391
+ value: auto
392
+ fp16_full_eval:
393
+ desc: null
394
+ value: false
395
+ fp16_opt_level:
396
+ desc: null
397
+ value: O1
398
+ generation_max_length:
399
+ desc: null
400
+ value: 40
401
+ generation_num_beams:
402
+ desc: null
403
+ value: 1
404
+ gradient_accumulation_steps:
405
+ desc: null
406
+ value: 2
407
+ gradient_checkpointing:
408
+ desc: null
409
+ value: true
410
+ greater_is_better:
411
+ desc: null
412
+ value: None
413
+ group_by_length:
414
+ desc: null
415
+ value: true
416
+ half_precision_backend:
417
+ desc: null
418
+ value: amp
419
+ hub_model_id:
420
+ desc: null
421
+ value: None
422
+ hub_strategy:
423
+ desc: null
424
+ value: every_save
425
+ hub_token:
426
+ desc: null
427
+ value: <HUB_TOKEN>
428
+ id2label:
429
+ desc: null
430
+ value:
431
+ '0': LABEL_0
432
+ '1': LABEL_1
433
+ ignore_data_skip:
434
+ desc: null
435
+ value: false
436
+ is_decoder:
437
+ desc: null
438
+ value: false
439
+ is_encoder_decoder:
440
+ desc: null
441
+ value: true
442
+ label2id:
443
+ desc: null
444
+ value:
445
+ LABEL_0: 0
446
+ LABEL_1: 1
447
+ label_names:
448
+ desc: null
449
+ value: None
450
+ label_smoothing_factor:
451
+ desc: null
452
+ value: 0.0
453
+ learning_rate:
454
+ desc: null
455
+ value: 0.0001
456
+ length_column_name:
457
+ desc: null
458
+ value: input_length
459
+ length_penalty:
460
+ desc: null
461
+ value: 1.0
462
+ load_best_model_at_end:
463
+ desc: null
464
+ value: false
465
+ local_rank:
466
+ desc: null
467
+ value: -1
468
+ log_level:
469
+ desc: null
470
+ value: -1
471
+ log_level_replica:
472
+ desc: null
473
+ value: -1
474
+ log_on_each_node:
475
+ desc: null
476
+ value: true
477
+ logging_dir:
478
+ desc: null
479
+ value: ./runs/Mar02_08-47-50_sanchit--v100
480
+ logging_first_step:
481
+ desc: null
482
+ value: false
483
+ logging_nan_inf_filter:
484
+ desc: null
485
+ value: true
486
+ logging_steps:
487
+ desc: null
488
+ value: 1
489
+ logging_strategy:
490
+ desc: null
491
+ value: steps
492
+ lr_scheduler_type:
493
+ desc: null
494
+ value: linear
495
+ max_grad_norm:
496
+ desc: null
497
+ value: 1.0
498
+ max_length:
499
+ desc: null
500
+ value: 50
501
+ max_steps:
502
+ desc: null
503
+ value: -1
504
+ metric_for_best_model:
505
+ desc: null
506
+ value: None
507
+ min_length:
508
+ desc: null
509
+ value: 0
510
+ model_type:
511
+ desc: null
512
+ value: speech-encoder-decoder
513
+ mp_parameters:
514
+ desc: null
515
+ value: ''
516
+ no_cuda:
517
+ desc: null
518
+ value: false
519
+ no_repeat_ngram_size:
520
+ desc: null
521
+ value: 0
522
+ num_beam_groups:
523
+ desc: null
524
+ value: 1
525
+ num_beams:
526
+ desc: null
527
+ value: 1
528
+ num_return_sequences:
529
+ desc: null
530
+ value: 1
531
+ num_train_epochs:
532
+ desc: null
533
+ value: 1.0
534
+ optim:
535
+ desc: null
536
+ value: adamw_hf
537
+ output_attentions:
538
+ desc: null
539
+ value: false
540
+ output_dir:
541
+ desc: null
542
+ value: ./
543
+ output_hidden_states:
544
+ desc: null
545
+ value: false
546
+ output_scores:
547
+ desc: null
548
+ value: false
549
+ overwrite_output_dir:
550
+ desc: null
551
+ value: true
552
+ pad_token_id:
553
+ desc: null
554
+ value: 50256
555
+ past_index:
556
+ desc: null
557
+ value: -1
558
+ per_device_eval_batch_size:
559
+ desc: null
560
+ value: 16
561
+ per_device_train_batch_size:
562
+ desc: null
563
+ value: 16
564
+ per_gpu_eval_batch_size:
565
+ desc: null
566
+ value: None
567
+ per_gpu_train_batch_size:
568
+ desc: null
569
+ value: None
570
+ predict_with_generate:
571
+ desc: null
572
+ value: true
573
+ prediction_loss_only:
574
+ desc: null
575
+ value: false
576
+ prefix:
577
+ desc: null
578
+ value: null
579
+ problem_type:
580
+ desc: null
581
+ value: null
582
+ processor_class:
583
+ desc: null
584
+ value: Wav2Vec2Processor
585
+ pruned_heads:
586
+ desc: null
587
+ value: {}
588
+ push_to_hub:
589
+ desc: null
590
+ value: true
591
+ push_to_hub_model_id:
592
+ desc: null
593
+ value: None
594
+ push_to_hub_organization:
595
+ desc: null
596
+ value: None
597
+ push_to_hub_token:
598
+ desc: null
599
+ value: <PUSH_TO_HUB_TOKEN>
600
+ remove_invalid_values:
601
+ desc: null
602
+ value: false
603
+ remove_unused_columns:
604
+ desc: null
605
+ value: true
606
+ repetition_penalty:
607
+ desc: null
608
+ value: 1.0
609
+ report_to:
610
+ desc: null
611
+ value: '[''wandb'']'
612
+ resume_from_checkpoint:
613
+ desc: null
614
+ value: None
615
+ return_dict:
616
+ desc: null
617
+ value: true
618
+ return_dict_in_generate:
619
+ desc: null
620
+ value: false
621
+ run_name:
622
+ desc: null
623
+ value: ./
624
+ save_on_each_node:
625
+ desc: null
626
+ value: false
627
+ save_steps:
628
+ desc: null
629
+ value: 500
630
+ save_strategy:
631
+ desc: null
632
+ value: steps
633
+ save_total_limit:
634
+ desc: null
635
+ value: 1
636
+ seed:
637
+ desc: null
638
+ value: 42
639
+ sep_token_id:
640
+ desc: null
641
+ value: null
642
+ sharded_ddp:
643
+ desc: null
644
+ value: '[]'
645
+ skip_memory_metrics:
646
+ desc: null
647
+ value: true
648
+ sortish_sampler:
649
+ desc: null
650
+ value: false
651
+ task_specific_params:
652
+ desc: null
653
+ value: null
654
+ temperature:
655
+ desc: null
656
+ value: 1.0
657
+ tf32:
658
+ desc: null
659
+ value: None
660
+ tie_encoder_decoder:
661
+ desc: null
662
+ value: false
663
+ tie_word_embeddings:
664
+ desc: null
665
+ value: false
666
+ tokenizer_class:
667
+ desc: null
668
+ value: null
669
+ top_k:
670
+ desc: null
671
+ value: 50
672
+ top_p:
673
+ desc: null
674
+ value: 1.0
675
+ torch_dtype:
676
+ desc: null
677
+ value: torch.float32
678
+ torchscript:
679
+ desc: null
680
+ value: false
681
+ tpu_metrics_debug:
682
+ desc: null
683
+ value: false
684
+ tpu_num_cores:
685
+ desc: null
686
+ value: None
687
+ train_batch_size:
688
+ desc: null
689
+ value: 16
690
+ transformers_version:
691
+ desc: null
692
+ value: null
693
+ use_bfloat16:
694
+ desc: null
695
+ value: false
696
+ use_cache:
697
+ desc: null
698
+ value: false
699
+ use_legacy_prediction_loop:
700
+ desc: null
701
+ value: false
702
+ warmup_ratio:
703
+ desc: null
704
+ value: 0.0
705
+ warmup_steps:
706
+ desc: null
707
+ value: 500
708
+ weight_decay:
709
+ desc: null
710
+ value: 0.0
711
+ xpu_backend:
712
+ desc: null
713
+ value: None
wandb/run-20220302_084833-2m6jtwtj/files/output.log ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ 0%| | 0/892 [00:00<?, ?it/s][WARNING|modeling_utils.py:388] 2022-03-02 08:48:38,520 >> Could not estimate the number of tokens of the input, floating-point operations will not be computed
4
+ Traceback (most recent call last):
5
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 539, in <module>
6
+ main()
7
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 491, in main
8
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
9
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1384, in train
10
+ tr_loss_step = self.training_step(model, inputs)
11
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1959, in training_step
12
+ loss = self.compute_loss(model, inputs)
13
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1991, in compute_loss
14
+ outputs = model(**inputs)
15
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
16
+ return forward_call(*input, **kwargs)
17
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py", line 503, in forward
18
+ encoder_outputs = self.encoder(
19
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
20
+ return forward_call(*input, **kwargs)
21
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1346, in forward
22
+ extract_features = self.feature_extractor(input_values)
23
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
24
+ return forward_call(*input, **kwargs)
25
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 514, in forward
26
+ hidden_states = conv_layer(hidden_states)
27
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
28
+ return forward_call(*input, **kwargs)
29
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 389, in forward
30
+ hidden_states = self.layer_norm(hidden_states)
31
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
32
+ return forward_call(*input, **kwargs)
33
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/normalization.py", line 189, in forward
34
+ return F.layer_norm(
35
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/functional.py", line 2347, in layer_norm
36
+ return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
37
+ RuntimeError: CUDA out of memory. Tried to allocate 1.63 GiB (GPU 0; 15.78 GiB total capacity; 10.07 GiB already allocated; 707.31 MiB free; 13.39 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
wandb/run-20220302_084833-2m6jtwtj/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220302_084833-2m6jtwtj/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-02T08:48:34.227833",
5
+ "startedAt": "2022-03-02T08:48:33.147557",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=16",
24
+ "--per_device_eval_batch_size=16",
25
+ "--gradient_accumulation_steps=2",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "622e8b594e8af8169be8cfe538228e49ae08c59d"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220302_084833-2m6jtwtj/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
1
+ {"_wandb": {"runtime": 8}}
wandb/run-20220302_084833-2m6jtwtj/logs/debug-internal.log ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-02 08:48:34,041 INFO MainThread:254137 [internal.py:wandb_internal():89] W&B internal server running at pid: 254137, started at: 2022-03-02 08:48:34.041581
2
+ 2022-03-02 08:48:34,043 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: check_version
3
+ 2022-03-02 08:48:34,044 INFO WriterThread:254137 [datastore.py:open_for_write():77] open: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/run-2m6jtwtj.wandb
4
+ 2022-03-02 08:48:34,045 DEBUG SenderThread:254137 [sender.py:send():235] send: header
5
+ 2022-03-02 08:48:34,046 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: check_version
6
+ 2022-03-02 08:48:34,116 DEBUG SenderThread:254137 [sender.py:send():235] send: run
7
+ 2022-03-02 08:48:34,222 INFO SenderThread:254137 [dir_watcher.py:__init__():169] watching files in: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files
8
+ 2022-03-02 08:48:34,222 INFO SenderThread:254137 [sender.py:_start_run_threads():809] run started: 2m6jtwtj with start time 1646210913
9
+ 2022-03-02 08:48:34,222 DEBUG SenderThread:254137 [sender.py:send():235] send: summary
10
+ 2022-03-02 08:48:34,222 INFO SenderThread:254137 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
11
+ 2022-03-02 08:48:34,223 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: run_start
12
+ 2022-03-02 08:48:34,227 DEBUG HandlerThread:254137 [meta.py:__init__():36] meta init
13
+ 2022-03-02 08:48:34,227 DEBUG HandlerThread:254137 [meta.py:__init__():50] meta init done
14
+ 2022-03-02 08:48:34,227 DEBUG HandlerThread:254137 [meta.py:probe():210] probe
15
+ 2022-03-02 08:48:34,234 DEBUG HandlerThread:254137 [meta.py:_setup_git():200] setup git
16
+ 2022-03-02 08:48:34,248 DEBUG HandlerThread:254137 [meta.py:_setup_git():207] setup git done
17
+ 2022-03-02 08:48:34,248 DEBUG HandlerThread:254137 [meta.py:_save_pip():54] save pip
18
+ 2022-03-02 08:48:34,249 DEBUG HandlerThread:254137 [meta.py:_save_pip():68] save pip done
19
+ 2022-03-02 08:48:34,249 DEBUG HandlerThread:254137 [meta.py:probe():248] probe done
20
+ 2022-03-02 08:48:34,326 DEBUG SenderThread:254137 [sender.py:send():235] send: files
21
+ 2022-03-02 08:48:34,327 INFO SenderThread:254137 [sender.py:_save_file():944] saving file wandb-metadata.json with policy now
22
+ 2022-03-02 08:48:34,331 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: stop_status
23
+ 2022-03-02 08:48:34,331 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: stop_status
24
+ 2022-03-02 08:48:34,417 DEBUG SenderThread:254137 [sender.py:send():235] send: config
25
+ 2022-03-02 08:48:34,418 DEBUG SenderThread:254137 [sender.py:send():235] send: metric
26
+ 2022-03-02 08:48:34,418 DEBUG SenderThread:254137 [sender.py:send():235] send: metric
27
+ 2022-03-02 08:48:34,418 WARNING SenderThread:254137 [sender.py:send_metric():902] Seen metric with glob (shouldnt happen)
28
+ 2022-03-02 08:48:34,653 INFO Thread-11 :254137 [upload_job.py:push():137] Uploaded file /tmp/tmpvy5l0hy6wandb/q5wa9252-wandb-metadata.json
29
+ 2022-03-02 08:48:35,224 INFO Thread-8 :254137 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/requirements.txt
30
+ 2022-03-02 08:48:35,224 INFO Thread-8 :254137 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/wandb-summary.json
31
+ 2022-03-02 08:48:35,224 INFO Thread-8 :254137 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/output.log
32
+ 2022-03-02 08:48:35,224 INFO Thread-8 :254137 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/wandb-metadata.json
33
+ 2022-03-02 08:48:37,223 INFO Thread-8 :254137 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/output.log
34
+ 2022-03-02 08:48:41,224 INFO Thread-8 :254137 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/output.log
35
+ 2022-03-02 08:48:42,671 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: poll_exit
36
+ 2022-03-02 08:48:42,671 DEBUG SenderThread:254137 [sender.py:send():235] send: telemetry
37
+ 2022-03-02 08:48:42,671 DEBUG SenderThread:254137 [sender.py:send():235] send: exit
38
+ 2022-03-02 08:48:42,671 INFO SenderThread:254137 [sender.py:send_exit():371] handling exit code: 1
39
+ 2022-03-02 08:48:42,671 INFO SenderThread:254137 [sender.py:send_exit():373] handling runtime: 8
40
+ 2022-03-02 08:48:42,672 INFO SenderThread:254137 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
41
+ 2022-03-02 08:48:42,672 INFO SenderThread:254137 [sender.py:send_exit():379] send defer
42
+ 2022-03-02 08:48:42,672 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: poll_exit
43
+ 2022-03-02 08:48:42,672 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
44
+ 2022-03-02 08:48:42,673 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 0
45
+ 2022-03-02 08:48:42,673 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
46
+ 2022-03-02 08:48:42,673 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 0
47
+ 2022-03-02 08:48:42,673 INFO SenderThread:254137 [sender.py:transition_state():392] send defer: 1
48
+ 2022-03-02 08:48:42,673 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
49
+ 2022-03-02 08:48:42,673 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 1
50
+ 2022-03-02 08:48:42,753 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
51
+ 2022-03-02 08:48:42,753 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 1
52
+ 2022-03-02 08:48:42,753 INFO SenderThread:254137 [sender.py:transition_state():392] send defer: 2
53
+ 2022-03-02 08:48:42,754 DEBUG SenderThread:254137 [sender.py:send():235] send: stats
54
+ 2022-03-02 08:48:42,754 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
55
+ 2022-03-02 08:48:42,754 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 2
56
+ 2022-03-02 08:48:42,755 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
57
+ 2022-03-02 08:48:42,755 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 2
58
+ 2022-03-02 08:48:42,755 INFO SenderThread:254137 [sender.py:transition_state():392] send defer: 3
59
+ 2022-03-02 08:48:42,755 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
60
+ 2022-03-02 08:48:42,755 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 3
61
+ 2022-03-02 08:48:42,755 DEBUG SenderThread:254137 [sender.py:send():235] send: summary
62
+ 2022-03-02 08:48:42,756 INFO SenderThread:254137 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
63
+ 2022-03-02 08:48:42,756 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
64
+ 2022-03-02 08:48:42,756 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 3
65
+ 2022-03-02 08:48:42,756 INFO SenderThread:254137 [sender.py:transition_state():392] send defer: 4
66
+ 2022-03-02 08:48:42,756 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
67
+ 2022-03-02 08:48:42,756 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 4
68
+ 2022-03-02 08:48:42,756 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
69
+ 2022-03-02 08:48:42,756 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 4
70
+ 2022-03-02 08:48:42,774 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: poll_exit
71
+ 2022-03-02 08:48:42,847 INFO SenderThread:254137 [sender.py:transition_state():392] send defer: 5
72
+ 2022-03-02 08:48:42,848 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: poll_exit
73
+ 2022-03-02 08:48:42,848 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
74
+ 2022-03-02 08:48:42,848 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 5
75
+ 2022-03-02 08:48:42,848 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
76
+ 2022-03-02 08:48:42,849 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 5
77
+ 2022-03-02 08:48:42,849 INFO SenderThread:254137 [dir_watcher.py:finish():283] shutting down directory watcher
78
+ 2022-03-02 08:48:42,949 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: poll_exit
79
+ 2022-03-02 08:48:43,225 INFO Thread-8 :254137 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/wandb-summary.json
80
+ 2022-03-02 08:48:43,226 INFO SenderThread:254137 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/output.log
81
+ 2022-03-02 08:48:43,227 INFO SenderThread:254137 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/config.yaml
82
+ 2022-03-02 08:48:43,227 INFO SenderThread:254137 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files
83
+ 2022-03-02 08:48:43,227 INFO SenderThread:254137 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/wandb-metadata.json wandb-metadata.json
84
+ 2022-03-02 08:48:43,227 INFO SenderThread:254137 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/output.log output.log
85
+ 2022-03-02 08:48:43,227 INFO SenderThread:254137 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/wandb-summary.json wandb-summary.json
86
+ 2022-03-02 08:48:43,227 INFO SenderThread:254137 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/requirements.txt requirements.txt
87
+ 2022-03-02 08:48:43,228 INFO SenderThread:254137 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/config.yaml config.yaml
88
+ 2022-03-02 08:48:43,230 INFO SenderThread:254137 [sender.py:transition_state():392] send defer: 6
89
+ 2022-03-02 08:48:43,231 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: poll_exit
90
+ 2022-03-02 08:48:43,231 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
91
+ 2022-03-02 08:48:43,231 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 6
92
+ 2022-03-02 08:48:43,234 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
93
+ 2022-03-02 08:48:43,234 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 6
94
+ 2022-03-02 08:48:43,234 INFO SenderThread:254137 [file_pusher.py:finish():177] shutting down file pusher
95
+ 2022-03-02 08:48:43,339 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: poll_exit
96
+ 2022-03-02 08:48:43,339 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: poll_exit
97
+ 2022-03-02 08:48:43,440 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: poll_exit
98
+ 2022-03-02 08:48:43,441 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: poll_exit
99
+ 2022-03-02 08:48:43,497 INFO Thread-13 :254137 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/wandb-summary.json
100
+ 2022-03-02 08:48:43,498 INFO Thread-14 :254137 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/requirements.txt
101
+ 2022-03-02 08:48:43,502 INFO Thread-12 :254137 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/output.log
102
+ 2022-03-02 08:48:43,533 INFO Thread-15 :254137 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/files/config.yaml
103
+ 2022-03-02 08:48:43,542 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: poll_exit
104
+ 2022-03-02 08:48:43,542 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: poll_exit
105
+ 2022-03-02 08:48:43,644 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: poll_exit
106
+ 2022-03-02 08:48:43,644 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: poll_exit
107
+ 2022-03-02 08:48:43,733 INFO Thread-7 :254137 [sender.py:transition_state():392] send defer: 7
108
+ 2022-03-02 08:48:43,734 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
109
+ 2022-03-02 08:48:43,734 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 7
110
+ 2022-03-02 08:48:43,734 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
111
+ 2022-03-02 08:48:43,734 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 7
112
+ 2022-03-02 08:48:43,745 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: poll_exit
113
+ 2022-03-02 08:48:43,905 INFO SenderThread:254137 [sender.py:transition_state():392] send defer: 8
114
+ 2022-03-02 08:48:43,905 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: poll_exit
115
+ 2022-03-02 08:48:43,906 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
116
+ 2022-03-02 08:48:43,906 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 8
117
+ 2022-03-02 08:48:43,906 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
118
+ 2022-03-02 08:48:43,906 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 8
119
+ 2022-03-02 08:48:43,906 INFO SenderThread:254137 [sender.py:transition_state():392] send defer: 9
120
+ 2022-03-02 08:48:43,907 DEBUG SenderThread:254137 [sender.py:send():235] send: final
121
+ 2022-03-02 08:48:43,907 DEBUG SenderThread:254137 [sender.py:send():235] send: footer
122
+ 2022-03-02 08:48:43,907 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: defer
123
+ 2022-03-02 08:48:43,907 INFO HandlerThread:254137 [handler.py:handle_request_defer():154] handle defer: 9
124
+ 2022-03-02 08:48:43,907 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: defer
125
+ 2022-03-02 08:48:43,908 INFO SenderThread:254137 [sender.py:send_request_defer():388] handle sender defer: 9
126
+ 2022-03-02 08:48:44,007 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: poll_exit
127
+ 2022-03-02 08:48:44,007 DEBUG SenderThread:254137 [sender.py:send_request():249] send_request: poll_exit
128
+ 2022-03-02 08:48:44,007 INFO SenderThread:254137 [file_pusher.py:join():182] waiting for file pusher
129
+ 2022-03-02 08:48:44,067 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: get_summary
130
+ 2022-03-02 08:48:44,068 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: sampled_history
131
+ 2022-03-02 08:48:44,068 DEBUG HandlerThread:254137 [handler.py:handle_request():131] handle_request: shutdown
132
+ 2022-03-02 08:48:44,068 INFO HandlerThread:254137 [handler.py:finish():739] shutting down handler
133
+ 2022-03-02 08:48:44,907 INFO WriterThread:254137 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/run-2m6jtwtj.wandb
134
+ 2022-03-02 08:48:45,066 INFO SenderThread:254137 [sender.py:finish():1075] shutting down sender
135
+ 2022-03-02 08:48:45,066 INFO SenderThread:254137 [file_pusher.py:finish():177] shutting down file pusher
136
+ 2022-03-02 08:48:45,066 INFO SenderThread:254137 [file_pusher.py:join():182] waiting for file pusher
137
+ 2022-03-02 08:48:45,068 INFO MainThread:254137 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220302_084833-2m6jtwtj/logs/debug.log ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/logs/debug.log
6
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_084833-2m6jtwtj/logs/debug-internal.log
7
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [wandb_init.py:init():471] starting backend
11
+ 2022-03-02 08:48:33,149 INFO MainThread:254038 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-02 08:48:33,204 INFO MainThread:254038 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-02 08:48:33,259 INFO MainThread:254038 [backend.py:ensure_launched():224] started backend process with pid: 254137
14
+ 2022-03-02 08:48:33,261 INFO MainThread:254038 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-02 08:48:33,270 INFO MainThread:254038 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-02 08:48:33,399 INFO MainThread:254038 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-02 08:48:34,114 INFO MainThread:254038 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.11 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
18
+
19
+ 2022-03-02 08:48:34,115 INFO MainThread:254038 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
20
+ 2022-03-02 08:48:34,222 INFO MainThread:254038 [wandb_init.py:init():624] starting run threads in backend
21
+ 2022-03-02 08:48:34,331 INFO MainThread:254038 [wandb_run.py:_console_start():1827] atexit reg
22
+ 2022-03-02 08:48:34,331 INFO MainThread:254038 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
23
+ 2022-03-02 08:48:34,332 INFO MainThread:254038 [wandb_run.py:_redirect():1706] Redirecting console.
24
+ 2022-03-02 08:48:34,333 INFO MainThread:254038 [wandb_run.py:_redirect():1762] Redirects installed.
25
+ 2022-03-02 08:48:34,334 INFO MainThread:254038 [wandb_init.py:init():651] run started, returning control to user process
26
+ 2022-03-02 08:48:34,337 INFO MainThread:254038 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0001, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_08-47-50_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 16, 'eval_batch_size': 16}
27
+ 2022-03-02 08:48:34,342 INFO MainThread:254038 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-02 08:48:40,411 INFO MainThread:254038 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-02 08:48:40,413 INFO MainThread:254038 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-02 08:48:42,673 INFO MainThread:254038 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-02 08:48:42,848 INFO MainThread:254038 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-02 08:48:43,238 INFO MainThread:254038 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 4
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 8718
52
+ }
53
+
54
+ 2022-03-02 08:48:43,340 INFO MainThread:254038 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 5
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 21066
59
+ total_bytes: 21066
60
+ }
61
+
62
+ 2022-03-02 08:48:43,441 INFO MainThread:254038 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 21066
67
+ total_bytes: 21066
68
+ }
69
+
70
+ 2022-03-02 08:48:43,543 INFO MainThread:254038 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 21066
75
+ total_bytes: 21066
76
+ }
77
+
78
+ 2022-03-02 08:48:43,644 INFO MainThread:254038 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 21066
83
+ total_bytes: 21066
84
+ }
85
+
86
+ 2022-03-02 08:48:43,906 INFO MainThread:254038 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 21066
91
+ total_bytes: 21066
92
+ }
93
+
94
+ 2022-03-02 08:48:44,066 INFO MainThread:254038 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
95
+ exit_result {
96
+ }
97
+ file_counts {
98
+ wandb_count: 5
99
+ }
100
+ pusher_stats {
101
+ uploaded_bytes: 21066
102
+ total_bytes: 21066
103
+ }
104
+ local_info {
105
+ }
106
+
107
+ 2022-03-02 08:48:45,198 INFO MainThread:254038 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220302_084833-2m6jtwtj/run-2m6jtwtj.wandb ADDED
Binary file (13.7 kB). View file
wandb/run-20220302_085000-3jbwcde6/files/config.yaml ADDED
@@ -0,0 +1,713 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: ./
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.10
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.9.5
22
+ start_time: 1646211000
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.9.5
37
+ 5: 0.12.10
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ adafactor:
42
+ desc: null
43
+ value: false
44
+ adam_beta1:
45
+ desc: null
46
+ value: 0.9
47
+ adam_beta2:
48
+ desc: null
49
+ value: 0.999
50
+ adam_epsilon:
51
+ desc: null
52
+ value: 1.0e-08
53
+ add_cross_attention:
54
+ desc: null
55
+ value: false
56
+ architectures:
57
+ desc: null
58
+ value:
59
+ - SpeechEncoderDecoderModel
60
+ bad_words_ids:
61
+ desc: null
62
+ value: null
63
+ bf16:
64
+ desc: null
65
+ value: false
66
+ bf16_full_eval:
67
+ desc: null
68
+ value: false
69
+ bos_token_id:
70
+ desc: null
71
+ value: null
72
+ chunk_size_feed_forward:
73
+ desc: null
74
+ value: 0
75
+ cross_attention_hidden_size:
76
+ desc: null
77
+ value: null
78
+ dataloader_drop_last:
79
+ desc: null
80
+ value: false
81
+ dataloader_num_workers:
82
+ desc: null
83
+ value: 0
84
+ dataloader_pin_memory:
85
+ desc: null
86
+ value: true
87
+ ddp_bucket_cap_mb:
88
+ desc: null
89
+ value: None
90
+ ddp_find_unused_parameters:
91
+ desc: null
92
+ value: None
93
+ debug:
94
+ desc: null
95
+ value: '[]'
96
+ decoder:
97
+ desc: null
98
+ value:
99
+ _name_or_path: gpt2-medium
100
+ activation_function: gelu_new
101
+ add_cross_attention: true
102
+ architectures:
103
+ - GPT2LMHeadModel
104
+ attn_pdrop: 0.0
105
+ bad_words_ids: null
106
+ bos_token_id: 50256
107
+ chunk_size_feed_forward: 0
108
+ cross_attention_hidden_size: null
109
+ decoder_start_token_id: null
110
+ diversity_penalty: 0.0
111
+ do_sample: false
112
+ early_stopping: false
113
+ embd_pdrop: 0.0
114
+ encoder_no_repeat_ngram_size: 0
115
+ eos_token_id: 50256
116
+ finetuning_task: null
117
+ forced_bos_token_id: null
118
+ forced_eos_token_id: null
119
+ id2label:
120
+ '0': LABEL_0
121
+ '1': LABEL_1
122
+ initializer_range: 0.02
123
+ is_decoder: true
124
+ is_encoder_decoder: false
125
+ label2id:
126
+ LABEL_0: 0
127
+ LABEL_1: 1
128
+ layer_norm_epsilon: 1.0e-05
129
+ length_penalty: 1.0
130
+ max_length: 20
131
+ min_length: 0
132
+ model_type: gpt2
133
+ n_ctx: 1024
134
+ n_embd: 1024
135
+ n_head: 16
136
+ n_inner: null
137
+ n_layer: 24
138
+ n_positions: 1024
139
+ n_special: 0
140
+ no_repeat_ngram_size: 0
141
+ num_beam_groups: 1
142
+ num_beams: 1
143
+ num_return_sequences: 1
144
+ output_attentions: false
145
+ output_hidden_states: false
146
+ output_scores: false
147
+ pad_token_id: null
148
+ predict_special_tokens: true
149
+ prefix: null
150
+ problem_type: null
151
+ pruned_heads: {}
152
+ remove_invalid_values: false
153
+ reorder_and_upcast_attn: false
154
+ repetition_penalty: 1.0
155
+ resid_pdrop: 0.0
156
+ return_dict: true
157
+ return_dict_in_generate: false
158
+ scale_attn_by_inverse_layer_idx: false
159
+ scale_attn_weights: true
160
+ sep_token_id: null
161
+ summary_activation: null
162
+ summary_first_dropout: 0.0
163
+ summary_proj_to_labels: true
164
+ summary_type: cls_index
165
+ summary_use_proj: true
166
+ task_specific_params:
167
+ text-generation:
168
+ do_sample: true
169
+ max_length: 50
170
+ temperature: 1.0
171
+ tie_encoder_decoder: false
172
+ tie_word_embeddings: true
173
+ tokenizer_class: null
174
+ top_k: 50
175
+ top_p: 1.0
176
+ torch_dtype: null
177
+ torchscript: false
178
+ transformers_version: 4.17.0.dev0
179
+ use_bfloat16: false
180
+ use_cache: false
181
+ vocab_size: 50257
182
+ decoder_start_token_id:
183
+ desc: null
184
+ value: 50256
185
+ deepspeed:
186
+ desc: null
187
+ value: None
188
+ disable_tqdm:
189
+ desc: null
190
+ value: false
191
+ diversity_penalty:
192
+ desc: null
193
+ value: 0.0
194
+ do_eval:
195
+ desc: null
196
+ value: true
197
+ do_predict:
198
+ desc: null
199
+ value: false
200
+ do_sample:
201
+ desc: null
202
+ value: false
203
+ do_train:
204
+ desc: null
205
+ value: true
206
+ early_stopping:
207
+ desc: null
208
+ value: false
209
+ encoder:
210
+ desc: null
211
+ value:
212
+ _name_or_path: facebook/wav2vec2-large-lv60
213
+ activation_dropout: 0.0
214
+ adapter_kernel_size: 3
215
+ adapter_stride: 2
216
+ add_adapter: true
217
+ add_cross_attention: false
218
+ apply_spec_augment: false
219
+ architectures:
220
+ - Wav2Vec2ForPreTraining
221
+ attention_dropout: 0.0
222
+ bad_words_ids: null
223
+ bos_token_id: 1
224
+ chunk_size_feed_forward: 0
225
+ classifier_proj_size: 256
226
+ codevector_dim: 768
227
+ contrastive_logits_temperature: 0.1
228
+ conv_bias: true
229
+ conv_dim:
230
+ - 512
231
+ - 512
232
+ - 512
233
+ - 512
234
+ - 512
235
+ - 512
236
+ - 512
237
+ conv_kernel:
238
+ - 10
239
+ - 3
240
+ - 3
241
+ - 3
242
+ - 3
243
+ - 2
244
+ - 2
245
+ conv_stride:
246
+ - 5
247
+ - 2
248
+ - 2
249
+ - 2
250
+ - 2
251
+ - 2
252
+ - 2
253
+ cross_attention_hidden_size: null
254
+ ctc_loss_reduction: sum
255
+ ctc_zero_infinity: false
256
+ decoder_start_token_id: null
257
+ diversity_loss_weight: 0.1
258
+ diversity_penalty: 0.0
259
+ do_sample: false
260
+ do_stable_layer_norm: true
261
+ early_stopping: false
262
+ encoder_no_repeat_ngram_size: 0
263
+ eos_token_id: 2
264
+ feat_extract_activation: gelu
265
+ feat_extract_dropout: 0.0
266
+ feat_extract_norm: layer
267
+ feat_proj_dropout: 0.0
268
+ feat_quantizer_dropout: 0.0
269
+ final_dropout: 0.0
270
+ finetuning_task: null
271
+ forced_bos_token_id: null
272
+ forced_eos_token_id: null
273
+ gradient_checkpointing: false
274
+ hidden_act: gelu
275
+ hidden_dropout: 0.0
276
+ hidden_dropout_prob: 0.0
277
+ hidden_size: 1024
278
+ id2label:
279
+ '0': LABEL_0
280
+ '1': LABEL_1
281
+ initializer_range: 0.02
282
+ intermediate_size: 4096
283
+ is_decoder: false
284
+ is_encoder_decoder: false
285
+ label2id:
286
+ LABEL_0: 0
287
+ LABEL_1: 1
288
+ layer_norm_eps: 1.0e-05
289
+ layerdrop: 0.0
290
+ length_penalty: 1.0
291
+ mask_feature_length: 10
292
+ mask_feature_min_masks: 0
293
+ mask_feature_prob: 0.0
294
+ mask_time_length: 10
295
+ mask_time_min_masks: 2
296
+ mask_time_prob: 0.0
297
+ max_length: 20
298
+ min_length: 0
299
+ model_type: wav2vec2
300
+ no_repeat_ngram_size: 0
301
+ num_adapter_layers: 3
302
+ num_attention_heads: 16
303
+ num_beam_groups: 1
304
+ num_beams: 1
305
+ num_codevector_groups: 2
306
+ num_codevectors_per_group: 320
307
+ num_conv_pos_embedding_groups: 16
308
+ num_conv_pos_embeddings: 128
309
+ num_feat_extract_layers: 7
310
+ num_hidden_layers: 24
311
+ num_negatives: 100
312
+ num_return_sequences: 1
313
+ output_attentions: false
314
+ output_hidden_size: 1024
315
+ output_hidden_states: false
316
+ output_scores: false
317
+ pad_token_id: 0
318
+ prefix: null
319
+ problem_type: null
320
+ proj_codevector_dim: 768
321
+ pruned_heads: {}
322
+ remove_invalid_values: false
323
+ repetition_penalty: 1.0
324
+ return_dict: true
325
+ return_dict_in_generate: false
326
+ sep_token_id: null
327
+ task_specific_params: null
328
+ tdnn_dilation:
329
+ - 1
330
+ - 2
331
+ - 3
332
+ - 1
333
+ - 1
334
+ tdnn_dim:
335
+ - 512
336
+ - 512
337
+ - 512
338
+ - 512
339
+ - 1500
340
+ tdnn_kernel:
341
+ - 5
342
+ - 3
343
+ - 3
344
+ - 1
345
+ - 1
346
+ temperature: 1.0
347
+ tie_encoder_decoder: false
348
+ tie_word_embeddings: true
349
+ tokenizer_class: null
350
+ top_k: 50
351
+ top_p: 1.0
352
+ torch_dtype: null
353
+ torchscript: false
354
+ transformers_version: 4.17.0.dev0
355
+ use_bfloat16: false
356
+ use_weighted_layer_sum: false
357
+ vocab_size: 32
358
+ xvector_output_dim: 512
359
+ encoder_no_repeat_ngram_size:
360
+ desc: null
361
+ value: 0
362
+ eos_token_id:
363
+ desc: null
364
+ value: 50256
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: None
368
+ eval_batch_size:
369
+ desc: null
370
+ value: 16
371
+ eval_steps:
372
+ desc: null
373
+ value: 500
374
+ evaluation_strategy:
375
+ desc: null
376
+ value: steps
377
+ finetuning_task:
378
+ desc: null
379
+ value: null
380
+ forced_bos_token_id:
381
+ desc: null
382
+ value: null
383
+ forced_eos_token_id:
384
+ desc: null
385
+ value: null
386
+ fp16:
387
+ desc: null
388
+ value: true
389
+ fp16_backend:
390
+ desc: null
391
+ value: auto
392
+ fp16_full_eval:
393
+ desc: null
394
+ value: false
395
+ fp16_opt_level:
396
+ desc: null
397
+ value: O1
398
+ generation_max_length:
399
+ desc: null
400
+ value: 40
401
+ generation_num_beams:
402
+ desc: null
403
+ value: 1
404
+ gradient_accumulation_steps:
405
+ desc: null
406
+ value: 4
407
+ gradient_checkpointing:
408
+ desc: null
409
+ value: true
410
+ greater_is_better:
411
+ desc: null
412
+ value: None
413
+ group_by_length:
414
+ desc: null
415
+ value: true
416
+ half_precision_backend:
417
+ desc: null
418
+ value: amp
419
+ hub_model_id:
420
+ desc: null
421
+ value: None
422
+ hub_strategy:
423
+ desc: null
424
+ value: every_save
425
+ hub_token:
426
+ desc: null
427
+ value: <HUB_TOKEN>
428
+ id2label:
429
+ desc: null
430
+ value:
431
+ '0': LABEL_0
432
+ '1': LABEL_1
433
+ ignore_data_skip:
434
+ desc: null
435
+ value: false
436
+ is_decoder:
437
+ desc: null
438
+ value: false
439
+ is_encoder_decoder:
440
+ desc: null
441
+ value: true
442
+ label2id:
443
+ desc: null
444
+ value:
445
+ LABEL_0: 0
446
+ LABEL_1: 1
447
+ label_names:
448
+ desc: null
449
+ value: None
450
+ label_smoothing_factor:
451
+ desc: null
452
+ value: 0.0
453
+ learning_rate:
454
+ desc: null
455
+ value: 0.0001
456
+ length_column_name:
457
+ desc: null
458
+ value: input_length
459
+ length_penalty:
460
+ desc: null
461
+ value: 1.0
462
+ load_best_model_at_end:
463
+ desc: null
464
+ value: false
465
+ local_rank:
466
+ desc: null
467
+ value: -1
468
+ log_level:
469
+ desc: null
470
+ value: -1
471
+ log_level_replica:
472
+ desc: null
473
+ value: -1
474
+ log_on_each_node:
475
+ desc: null
476
+ value: true
477
+ logging_dir:
478
+ desc: null
479
+ value: ./runs/Mar02_08-49-20_sanchit--v100
480
+ logging_first_step:
481
+ desc: null
482
+ value: false
483
+ logging_nan_inf_filter:
484
+ desc: null
485
+ value: true
486
+ logging_steps:
487
+ desc: null
488
+ value: 1
489
+ logging_strategy:
490
+ desc: null
491
+ value: steps
492
+ lr_scheduler_type:
493
+ desc: null
494
+ value: linear
495
+ max_grad_norm:
496
+ desc: null
497
+ value: 1.0
498
+ max_length:
499
+ desc: null
500
+ value: 50
501
+ max_steps:
502
+ desc: null
503
+ value: -1
504
+ metric_for_best_model:
505
+ desc: null
506
+ value: None
507
+ min_length:
508
+ desc: null
509
+ value: 0
510
+ model_type:
511
+ desc: null
512
+ value: speech-encoder-decoder
513
+ mp_parameters:
514
+ desc: null
515
+ value: ''
516
+ no_cuda:
517
+ desc: null
518
+ value: false
519
+ no_repeat_ngram_size:
520
+ desc: null
521
+ value: 0
522
+ num_beam_groups:
523
+ desc: null
524
+ value: 1
525
+ num_beams:
526
+ desc: null
527
+ value: 1
528
+ num_return_sequences:
529
+ desc: null
530
+ value: 1
531
+ num_train_epochs:
532
+ desc: null
533
+ value: 1.0
534
+ optim:
535
+ desc: null
536
+ value: adamw_hf
537
+ output_attentions:
538
+ desc: null
539
+ value: false
540
+ output_dir:
541
+ desc: null
542
+ value: ./
543
+ output_hidden_states:
544
+ desc: null
545
+ value: false
546
+ output_scores:
547
+ desc: null
548
+ value: false
549
+ overwrite_output_dir:
550
+ desc: null
551
+ value: true
552
+ pad_token_id:
553
+ desc: null
554
+ value: 50256
555
+ past_index:
556
+ desc: null
557
+ value: -1
558
+ per_device_eval_batch_size:
559
+ desc: null
560
+ value: 16
561
+ per_device_train_batch_size:
562
+ desc: null
563
+ value: 16
564
+ per_gpu_eval_batch_size:
565
+ desc: null
566
+ value: None
567
+ per_gpu_train_batch_size:
568
+ desc: null
569
+ value: None
570
+ predict_with_generate:
571
+ desc: null
572
+ value: true
573
+ prediction_loss_only:
574
+ desc: null
575
+ value: false
576
+ prefix:
577
+ desc: null
578
+ value: null
579
+ problem_type:
580
+ desc: null
581
+ value: null
582
+ processor_class:
583
+ desc: null
584
+ value: Wav2Vec2Processor
585
+ pruned_heads:
586
+ desc: null
587
+ value: {}
588
+ push_to_hub:
589
+ desc: null
590
+ value: true
591
+ push_to_hub_model_id:
592
+ desc: null
593
+ value: None
594
+ push_to_hub_organization:
595
+ desc: null
596
+ value: None
597
+ push_to_hub_token:
598
+ desc: null
599
+ value: <PUSH_TO_HUB_TOKEN>
600
+ remove_invalid_values:
601
+ desc: null
602
+ value: false
603
+ remove_unused_columns:
604
+ desc: null
605
+ value: true
606
+ repetition_penalty:
607
+ desc: null
608
+ value: 1.0
609
+ report_to:
610
+ desc: null
611
+ value: '[''wandb'']'
612
+ resume_from_checkpoint:
613
+ desc: null
614
+ value: None
615
+ return_dict:
616
+ desc: null
617
+ value: true
618
+ return_dict_in_generate:
619
+ desc: null
620
+ value: false
621
+ run_name:
622
+ desc: null
623
+ value: ./
624
+ save_on_each_node:
625
+ desc: null
626
+ value: false
627
+ save_steps:
628
+ desc: null
629
+ value: 500
630
+ save_strategy:
631
+ desc: null
632
+ value: steps
633
+ save_total_limit:
634
+ desc: null
635
+ value: 1
636
+ seed:
637
+ desc: null
638
+ value: 42
639
+ sep_token_id:
640
+ desc: null
641
+ value: null
642
+ sharded_ddp:
643
+ desc: null
644
+ value: '[]'
645
+ skip_memory_metrics:
646
+ desc: null
647
+ value: true
648
+ sortish_sampler:
649
+ desc: null
650
+ value: false
651
+ task_specific_params:
652
+ desc: null
653
+ value: null
654
+ temperature:
655
+ desc: null
656
+ value: 1.0
657
+ tf32:
658
+ desc: null
659
+ value: None
660
+ tie_encoder_decoder:
661
+ desc: null
662
+ value: false
663
+ tie_word_embeddings:
664
+ desc: null
665
+ value: false
666
+ tokenizer_class:
667
+ desc: null
668
+ value: null
669
+ top_k:
670
+ desc: null
671
+ value: 50
672
+ top_p:
673
+ desc: null
674
+ value: 1.0
675
+ torch_dtype:
676
+ desc: null
677
+ value: torch.float32
678
+ torchscript:
679
+ desc: null
680
+ value: false
681
+ tpu_metrics_debug:
682
+ desc: null
683
+ value: false
684
+ tpu_num_cores:
685
+ desc: null
686
+ value: None
687
+ train_batch_size:
688
+ desc: null
689
+ value: 16
690
+ transformers_version:
691
+ desc: null
692
+ value: null
693
+ use_bfloat16:
694
+ desc: null
695
+ value: false
696
+ use_cache:
697
+ desc: null
698
+ value: false
699
+ use_legacy_prediction_loop:
700
+ desc: null
701
+ value: false
702
+ warmup_ratio:
703
+ desc: null
704
+ value: 0.0
705
+ warmup_steps:
706
+ desc: null
707
+ value: 500
708
+ weight_decay:
709
+ desc: null
710
+ value: 0.0
711
+ xpu_backend:
712
+ desc: null
713
+ value: None
wandb/run-20220302_085000-3jbwcde6/files/output.log ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ 0%| | 0/446 [00:00<?, ?it/s][WARNING|modeling_utils.py:388] 2022-03-02 08:50:05,594 >> Could not estimate the number of tokens of the input, floating-point operations will not be computed
4
+ Traceback (most recent call last):
5
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 539, in <module>
6
+ main()
7
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 491, in main
8
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
9
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1384, in train
10
+ tr_loss_step = self.training_step(model, inputs)
11
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1959, in training_step
12
+ loss = self.compute_loss(model, inputs)
13
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1991, in compute_loss
14
+ outputs = model(**inputs)
15
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
16
+ return forward_call(*input, **kwargs)
17
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py", line 503, in forward
18
+ encoder_outputs = self.encoder(
19
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
20
+ return forward_call(*input, **kwargs)
21
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1346, in forward
22
+ extract_features = self.feature_extractor(input_values)
23
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
24
+ return forward_call(*input, **kwargs)
25
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 514, in forward
26
+ hidden_states = conv_layer(hidden_states)
27
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
28
+ return forward_call(*input, **kwargs)
29
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 389, in forward
30
+ hidden_states = self.layer_norm(hidden_states)
31
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
32
+ return forward_call(*input, **kwargs)
33
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/normalization.py", line 189, in forward
34
+ return F.layer_norm(
35
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/functional.py", line 2347, in layer_norm
36
+ return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
37
+ RuntimeError: CUDA out of memory. Tried to allocate 1.64 GiB (GPU 0; 15.78 GiB total capacity; 10.11 GiB already allocated; 707.31 MiB free; 13.39 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
wandb/run-20220302_085000-3jbwcde6/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220302_085000-3jbwcde6/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-02T08:50:01.367126",
5
+ "startedAt": "2022-03-02T08:50:00.316296",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=16",
24
+ "--per_device_eval_batch_size=16",
25
+ "--gradient_accumulation_steps=4",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "622e8b594e8af8169be8cfe538228e49ae08c59d"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220302_085000-3jbwcde6/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
1
+ {"_wandb": {"runtime": 8}}
wandb/run-20220302_085000-3jbwcde6/logs/debug-internal.log ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-02 08:50:01,203 INFO MainThread:254314 [internal.py:wandb_internal():89] W&B internal server running at pid: 254314, started at: 2022-03-02 08:50:01.203327
2
+ 2022-03-02 08:50:01,205 INFO WriterThread:254314 [datastore.py:open_for_write():77] open: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/run-3jbwcde6.wandb
3
+ 2022-03-02 08:50:01,206 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: check_version
4
+ 2022-03-02 08:50:01,207 DEBUG SenderThread:254314 [sender.py:send():235] send: header
5
+ 2022-03-02 08:50:01,207 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: check_version
6
+ 2022-03-02 08:50:01,273 DEBUG SenderThread:254314 [sender.py:send():235] send: run
7
+ 2022-03-02 08:50:01,361 INFO SenderThread:254314 [dir_watcher.py:__init__():169] watching files in: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files
8
+ 2022-03-02 08:50:01,361 INFO SenderThread:254314 [sender.py:_start_run_threads():809] run started: 3jbwcde6 with start time 1646211000
9
+ 2022-03-02 08:50:01,361 DEBUG SenderThread:254314 [sender.py:send():235] send: summary
10
+ 2022-03-02 08:50:01,361 INFO SenderThread:254314 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
11
+ 2022-03-02 08:50:01,362 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: run_start
12
+ 2022-03-02 08:50:01,366 DEBUG HandlerThread:254314 [meta.py:__init__():36] meta init
13
+ 2022-03-02 08:50:01,367 DEBUG HandlerThread:254314 [meta.py:__init__():50] meta init done
14
+ 2022-03-02 08:50:01,367 DEBUG HandlerThread:254314 [meta.py:probe():210] probe
15
+ 2022-03-02 08:50:01,373 DEBUG HandlerThread:254314 [meta.py:_setup_git():200] setup git
16
+ 2022-03-02 08:50:01,387 DEBUG HandlerThread:254314 [meta.py:_setup_git():207] setup git done
17
+ 2022-03-02 08:50:01,388 DEBUG HandlerThread:254314 [meta.py:_save_pip():54] save pip
18
+ 2022-03-02 08:50:01,388 DEBUG HandlerThread:254314 [meta.py:_save_pip():68] save pip done
19
+ 2022-03-02 08:50:01,388 DEBUG HandlerThread:254314 [meta.py:probe():248] probe done
20
+ 2022-03-02 08:50:01,468 DEBUG SenderThread:254314 [sender.py:send():235] send: files
21
+ 2022-03-02 08:50:01,468 INFO SenderThread:254314 [sender.py:_save_file():944] saving file wandb-metadata.json with policy now
22
+ 2022-03-02 08:50:01,473 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: stop_status
23
+ 2022-03-02 08:50:01,473 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: stop_status
24
+ 2022-03-02 08:50:01,520 DEBUG SenderThread:254314 [sender.py:send():235] send: config
25
+ 2022-03-02 08:50:01,521 DEBUG SenderThread:254314 [sender.py:send():235] send: metric
26
+ 2022-03-02 08:50:01,522 DEBUG SenderThread:254314 [sender.py:send():235] send: metric
27
+ 2022-03-02 08:50:01,522 WARNING SenderThread:254314 [sender.py:send_metric():902] Seen metric with glob (shouldnt happen)
28
+ 2022-03-02 08:50:01,741 INFO Thread-11 :254314 [upload_job.py:push():137] Uploaded file /tmp/tmpfcch1qofwandb/3oinsdv0-wandb-metadata.json
29
+ 2022-03-02 08:50:02,363 INFO Thread-8 :254314 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/wandb-metadata.json
30
+ 2022-03-02 08:50:02,363 INFO Thread-8 :254314 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/wandb-summary.json
31
+ 2022-03-02 08:50:02,363 INFO Thread-8 :254314 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/output.log
32
+ 2022-03-02 08:50:02,363 INFO Thread-8 :254314 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/requirements.txt
33
+ 2022-03-02 08:50:04,362 INFO Thread-8 :254314 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/output.log
34
+ 2022-03-02 08:50:08,364 INFO Thread-8 :254314 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/output.log
35
+ 2022-03-02 08:50:09,886 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
36
+ 2022-03-02 08:50:09,886 DEBUG SenderThread:254314 [sender.py:send():235] send: telemetry
37
+ 2022-03-02 08:50:09,887 DEBUG SenderThread:254314 [sender.py:send():235] send: exit
38
+ 2022-03-02 08:50:09,887 INFO SenderThread:254314 [sender.py:send_exit():371] handling exit code: 1
39
+ 2022-03-02 08:50:09,887 INFO SenderThread:254314 [sender.py:send_exit():373] handling runtime: 8
40
+ 2022-03-02 08:50:09,887 INFO SenderThread:254314 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
41
+ 2022-03-02 08:50:09,887 INFO SenderThread:254314 [sender.py:send_exit():379] send defer
42
+ 2022-03-02 08:50:09,888 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
43
+ 2022-03-02 08:50:09,888 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
44
+ 2022-03-02 08:50:09,888 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 0
45
+ 2022-03-02 08:50:09,888 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
46
+ 2022-03-02 08:50:09,888 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 0
47
+ 2022-03-02 08:50:09,889 INFO SenderThread:254314 [sender.py:transition_state():392] send defer: 1
48
+ 2022-03-02 08:50:09,889 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
49
+ 2022-03-02 08:50:09,889 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 1
50
+ 2022-03-02 08:50:09,992 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
51
+ 2022-03-02 08:50:09,992 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
52
+ 2022-03-02 08:50:09,992 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 1
53
+ 2022-03-02 08:50:09,993 INFO SenderThread:254314 [sender.py:transition_state():392] send defer: 2
54
+ 2022-03-02 08:50:09,993 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
55
+ 2022-03-02 08:50:09,993 DEBUG SenderThread:254314 [sender.py:send():235] send: stats
56
+ 2022-03-02 08:50:09,994 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
57
+ 2022-03-02 08:50:09,994 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 2
58
+ 2022-03-02 08:50:09,994 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
59
+ 2022-03-02 08:50:09,995 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 2
60
+ 2022-03-02 08:50:09,995 INFO SenderThread:254314 [sender.py:transition_state():392] send defer: 3
61
+ 2022-03-02 08:50:09,995 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
62
+ 2022-03-02 08:50:09,995 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 3
63
+ 2022-03-02 08:50:09,995 DEBUG SenderThread:254314 [sender.py:send():235] send: summary
64
+ 2022-03-02 08:50:09,995 INFO SenderThread:254314 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
65
+ 2022-03-02 08:50:09,996 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
66
+ 2022-03-02 08:50:09,996 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 3
67
+ 2022-03-02 08:50:09,996 INFO SenderThread:254314 [sender.py:transition_state():392] send defer: 4
68
+ 2022-03-02 08:50:09,996 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
69
+ 2022-03-02 08:50:09,996 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 4
70
+ 2022-03-02 08:50:09,996 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
71
+ 2022-03-02 08:50:09,996 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 4
72
+ 2022-03-02 08:50:10,100 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
73
+ 2022-03-02 08:50:10,126 INFO SenderThread:254314 [sender.py:transition_state():392] send defer: 5
74
+ 2022-03-02 08:50:10,126 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
75
+ 2022-03-02 08:50:10,127 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
76
+ 2022-03-02 08:50:10,127 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 5
77
+ 2022-03-02 08:50:10,127 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
78
+ 2022-03-02 08:50:10,127 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 5
79
+ 2022-03-02 08:50:10,127 INFO SenderThread:254314 [dir_watcher.py:finish():283] shutting down directory watcher
80
+ 2022-03-02 08:50:10,228 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
81
+ 2022-03-02 08:50:10,364 INFO Thread-8 :254314 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/wandb-summary.json
82
+ 2022-03-02 08:50:10,365 INFO SenderThread:254314 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/output.log
83
+ 2022-03-02 08:50:10,366 INFO SenderThread:254314 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/config.yaml
84
+ 2022-03-02 08:50:10,366 INFO SenderThread:254314 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files
85
+ 2022-03-02 08:50:10,366 INFO SenderThread:254314 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/wandb-metadata.json wandb-metadata.json
86
+ 2022-03-02 08:50:10,366 INFO SenderThread:254314 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/output.log output.log
87
+ 2022-03-02 08:50:10,366 INFO SenderThread:254314 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/wandb-summary.json wandb-summary.json
88
+ 2022-03-02 08:50:10,367 INFO SenderThread:254314 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/requirements.txt requirements.txt
89
+ 2022-03-02 08:50:10,372 INFO SenderThread:254314 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/config.yaml config.yaml
90
+ 2022-03-02 08:50:10,375 INFO SenderThread:254314 [sender.py:transition_state():392] send defer: 6
91
+ 2022-03-02 08:50:10,376 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
92
+ 2022-03-02 08:50:10,376 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
93
+ 2022-03-02 08:50:10,377 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 6
94
+ 2022-03-02 08:50:10,379 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
95
+ 2022-03-02 08:50:10,379 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 6
96
+ 2022-03-02 08:50:10,379 INFO SenderThread:254314 [file_pusher.py:finish():177] shutting down file pusher
97
+ 2022-03-02 08:50:10,477 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
98
+ 2022-03-02 08:50:10,478 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
99
+ 2022-03-02 08:50:10,579 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
100
+ 2022-03-02 08:50:10,579 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
101
+ 2022-03-02 08:50:10,641 INFO Thread-12 :254314 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/output.log
102
+ 2022-03-02 08:50:10,642 INFO Thread-14 :254314 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/requirements.txt
103
+ 2022-03-02 08:50:10,656 INFO Thread-15 :254314 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/config.yaml
104
+ 2022-03-02 08:50:10,681 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
105
+ 2022-03-02 08:50:10,681 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
106
+ 2022-03-02 08:50:10,700 INFO Thread-13 :254314 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/files/wandb-summary.json
107
+ 2022-03-02 08:50:10,782 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
108
+ 2022-03-02 08:50:10,782 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
109
+ 2022-03-02 08:50:10,884 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
110
+ 2022-03-02 08:50:10,884 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
111
+ 2022-03-02 08:50:10,901 INFO Thread-7 :254314 [sender.py:transition_state():392] send defer: 7
112
+ 2022-03-02 08:50:10,901 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
113
+ 2022-03-02 08:50:10,901 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 7
114
+ 2022-03-02 08:50:10,902 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
115
+ 2022-03-02 08:50:10,902 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 7
116
+ 2022-03-02 08:50:10,985 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
117
+ 2022-03-02 08:50:11,046 INFO SenderThread:254314 [sender.py:transition_state():392] send defer: 8
118
+ 2022-03-02 08:50:11,047 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
119
+ 2022-03-02 08:50:11,047 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
120
+ 2022-03-02 08:50:11,047 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 8
121
+ 2022-03-02 08:50:11,047 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
122
+ 2022-03-02 08:50:11,047 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 8
123
+ 2022-03-02 08:50:11,047 INFO SenderThread:254314 [sender.py:transition_state():392] send defer: 9
124
+ 2022-03-02 08:50:11,048 DEBUG SenderThread:254314 [sender.py:send():235] send: final
125
+ 2022-03-02 08:50:11,048 DEBUG SenderThread:254314 [sender.py:send():235] send: footer
126
+ 2022-03-02 08:50:11,048 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: defer
127
+ 2022-03-02 08:50:11,048 INFO HandlerThread:254314 [handler.py:handle_request_defer():154] handle defer: 9
128
+ 2022-03-02 08:50:11,049 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: defer
129
+ 2022-03-02 08:50:11,049 INFO SenderThread:254314 [sender.py:send_request_defer():388] handle sender defer: 9
130
+ 2022-03-02 08:50:11,148 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: poll_exit
131
+ 2022-03-02 08:50:11,148 DEBUG SenderThread:254314 [sender.py:send_request():249] send_request: poll_exit
132
+ 2022-03-02 08:50:11,148 INFO SenderThread:254314 [file_pusher.py:join():182] waiting for file pusher
133
+ 2022-03-02 08:50:11,233 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: get_summary
134
+ 2022-03-02 08:50:11,233 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: sampled_history
135
+ 2022-03-02 08:50:11,234 DEBUG HandlerThread:254314 [handler.py:handle_request():131] handle_request: shutdown
136
+ 2022-03-02 08:50:11,234 INFO HandlerThread:254314 [handler.py:finish():739] shutting down handler
137
+ 2022-03-02 08:50:12,049 INFO WriterThread:254314 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_085000-3jbwcde6/run-3jbwcde6.wandb
138
+ 2022-03-02 08:50:12,232 INFO SenderThread:254314 [sender.py:finish():1075] shutting down sender
139
+ 2022-03-02 08:50:12,232 INFO SenderThread:254314 [file_pusher.py:finish():177] shutting down file pusher
140
+ 2022-03-02 08:50:12,232 INFO SenderThread:254314 [file_pusher.py:join():182] waiting for file pusher
141
+ 2022-03-02 08:50:12,234 INFO MainThread:254314 [internal.py:handle_exit():79] Internal process exited