sanchit-gandhi HF staff commited on
Commit
a5afdd5
1 Parent(s): 7dec219

Training in progress, step 500

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. pytorch_model.bin +1 -1
  3. training_args.bin +1 -1
  4. wandb/debug-internal.log +1 -1
  5. wandb/debug.log +1 -1
  6. wandb/latest-run +1 -1
  7. wandb/run-20220228_204859-8xn2plkx/files/config.yaml +7 -0
  8. wandb/run-20220228_204859-8xn2plkx/files/output.log +7 -0
  9. wandb/run-20220228_204859-8xn2plkx/files/wandb-summary.json +0 -0
  10. wandb/run-20220228_204859-8xn2plkx/logs/debug-internal.log +111 -0
  11. wandb/run-20220228_204859-8xn2plkx/logs/debug.log +98 -0
  12. wandb/run-20220228_204859-8xn2plkx/run-8xn2plkx.wandb +2 -2
  13. wandb/run-20220228_223243-2ay2wvge/files/config.yaml +0 -0
  14. wandb/run-20220228_223243-2ay2wvge/files/output.log +0 -0
  15. wandb/run-20220228_223243-2ay2wvge/files/requirements.txt +184 -0
  16. wandb/run-20220228_223243-2ay2wvge/files/wandb-metadata.json +60 -0
  17. wandb/run-20220228_223243-2ay2wvge/files/wandb-summary.json +0 -0
  18. wandb/run-20220228_223243-2ay2wvge/logs/debug-internal.log +0 -0
  19. wandb/run-20220228_223243-2ay2wvge/logs/debug.log +124 -0
  20. wandb/run-20220228_223243-2ay2wvge/run-2ay2wvge.wandb +3 -0
  21. wandb/run-20220228_231357-3lq2qpez/files/config.yaml +0 -0
  22. wandb/run-20220228_231357-3lq2qpez/files/output.log +0 -0
  23. wandb/run-20220228_231357-3lq2qpez/files/requirements.txt +184 -0
  24. wandb/run-20220228_231357-3lq2qpez/files/wandb-metadata.json +60 -0
  25. wandb/run-20220228_231357-3lq2qpez/files/wandb-summary.json +0 -0
  26. wandb/run-20220228_231357-3lq2qpez/logs/debug-internal.log +0 -0
  27. wandb/run-20220228_231357-3lq2qpez/logs/debug.log +132 -0
  28. wandb/run-20220228_231357-3lq2qpez/run-3lq2qpez.wandb +3 -0
  29. wandb/run-20220301_002446-2vmlu6y4/files/config.yaml +0 -0
  30. wandb/run-20220301_002446-2vmlu6y4/files/output.log +0 -0
  31. wandb/run-20220301_002446-2vmlu6y4/files/requirements.txt +184 -0
  32. wandb/run-20220301_002446-2vmlu6y4/files/wandb-metadata.json +60 -0
  33. wandb/run-20220301_002446-2vmlu6y4/files/wandb-summary.json +0 -0
  34. wandb/run-20220301_002446-2vmlu6y4/logs/debug-internal.log +0 -0
  35. wandb/run-20220301_002446-2vmlu6y4/logs/debug.log +124 -0
  36. wandb/run-20220301_002446-2vmlu6y4/run-2vmlu6y4.wandb +3 -0
  37. wandb/run-20220301_012648-nzue7jk0/files/config.yaml +713 -0
  38. wandb/run-20220301_012648-nzue7jk0/files/output.log +37 -0
  39. wandb/run-20220301_012648-nzue7jk0/files/requirements.txt +184 -0
  40. wandb/run-20220301_012648-nzue7jk0/files/wandb-metadata.json +60 -0
  41. wandb/run-20220301_012648-nzue7jk0/files/wandb-summary.json +1 -0
  42. wandb/run-20220301_012648-nzue7jk0/logs/debug-internal.log +137 -0
  43. wandb/run-20220301_012648-nzue7jk0/logs/debug.log +106 -0
  44. wandb/run-20220301_012648-nzue7jk0/run-nzue7jk0.wandb +0 -0
  45. wandb/run-20220301_012816-3bh80nzn/files/config.yaml +713 -0
  46. wandb/run-20220301_012816-3bh80nzn/files/output.log +37 -0
  47. wandb/run-20220301_012816-3bh80nzn/files/requirements.txt +184 -0
  48. wandb/run-20220301_012816-3bh80nzn/files/wandb-metadata.json +60 -0
  49. wandb/run-20220301_012816-3bh80nzn/files/wandb-summary.json +1 -0
  50. wandb/run-20220301_012816-3bh80nzn/logs/debug-internal.log +139 -0
.gitattributes CHANGED
@@ -31,3 +31,7 @@ wandb/run-20220228_144113-18osrjm3/run-18osrjm3.wandb filter=lfs diff=lfs merge=
31
  wandb/run-20220228_163348-20lja62k/run-20lja62k.wandb filter=lfs diff=lfs merge=lfs -text
32
  wandb/run-20220228_185039-2l3jouo4/run-2l3jouo4.wandb filter=lfs diff=lfs merge=lfs -text
33
  wandb/run-20220228_204859-8xn2plkx/run-8xn2plkx.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
31
  wandb/run-20220228_163348-20lja62k/run-20lja62k.wandb filter=lfs diff=lfs merge=lfs -text
32
  wandb/run-20220228_185039-2l3jouo4/run-2l3jouo4.wandb filter=lfs diff=lfs merge=lfs -text
33
  wandb/run-20220228_204859-8xn2plkx/run-8xn2plkx.wandb filter=lfs diff=lfs merge=lfs -text
34
+ wandb/run-20220228_223243-2ay2wvge/run-2ay2wvge.wandb filter=lfs diff=lfs merge=lfs -text
35
+ wandb/run-20220228_231357-3lq2qpez/run-3lq2qpez.wandb filter=lfs diff=lfs merge=lfs -text
36
+ wandb/run-20220301_002446-2vmlu6y4/run-2vmlu6y4.wandb filter=lfs diff=lfs merge=lfs -text
37
+ wandb/run-20220301_013112-3e2necnj/run-3e2necnj.wandb filter=lfs diff=lfs merge=lfs -text
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32f0558147dab98c8bda239de13130a757f5440c4ed9bcb29975f2168158fe49
3
  size 3210531882
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b2cae6eb7b8fa6ba6b2fd41ab009d75a9e528102a7b684e368b46fa34267212
3
  size 3210531882
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af69507712faa8324ea7d50db5588b4a257ef8dd1b66c1e247d9ba899810b9da
3
  size 3119
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d030c260875dd4bb2687a9e66b5015be5f62ce53102733c365f4bf114fba4378
3
  size 3119
wandb/debug-internal.log CHANGED
@@ -1 +1 @@
1
- run-20220228_204859-8xn2plkx/logs/debug-internal.log
1
+ run-20220301_013112-3e2necnj/logs/debug-internal.log
wandb/debug.log CHANGED
@@ -1 +1 @@
1
- run-20220228_204859-8xn2plkx/logs/debug.log
1
+ run-20220301_013112-3e2necnj/logs/debug.log
wandb/latest-run CHANGED
@@ -1 +1 @@
1
- run-20220228_204859-8xn2plkx
1
+ run-20220301_013112-3e2necnj
wandb/run-20220228_204859-8xn2plkx/files/config.yaml CHANGED
@@ -10673,7 +10673,14 @@ _wandb:
10673
  - 1
10674
  - 5
10675
  - 11
 
 
 
 
 
10676
  3:
 
 
10677
  - 13
10678
  4: 3.9.5
10679
  5: 0.12.10
10673
  - 1
10674
  - 5
10675
  - 11
10676
+ 2:
10677
+ - 1
10678
+ - 5
10679
+ - 11
10680
+ - 12
10681
  3:
10682
+ - 1
10683
+ - 7
10684
  - 13
10685
  4: 3.9.5
10686
  5: 0.12.10
wandb/run-20220228_204859-8xn2plkx/files/output.log CHANGED
@@ -2807,3 +2807,10 @@ Upload file wandb/run-20220228_204859-8xn2plkx/run-8xn2plkx.wandb: 43%|██
2807
  eval_samples_per_second = 2.791
2808
  eval_steps_per_second = 0.233
2809
  [INFO|modeling_utils.py:1081] 2022-02-28 22:30:33,779 >> Model weights saved in ./pytorch_model.bin:11<11:26, 3.16s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
 
 
 
 
 
 
 
2807
  eval_samples_per_second = 2.791
2808
  eval_steps_per_second = 0.233
2809
  [INFO|modeling_utils.py:1081] 2022-02-28 22:30:33,779 >> Model weights saved in ./pytorch_model.bin:11<11:26, 3.16s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2810
+ Upload file wandb/run-20220228_204859-8xn2plkx/run-8xn2plkx.wandb: 45%|█████ | 15.9M/35.0M [00:01<00:01, 16.6MB/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2811
+ Upload file wandb/run-20220228_204859-8xn2plkx/run-8xn2plkx.wandb: 100%|███████████| 35.0M/35.0M [00:02<00:00, 18.3MB/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2812
+ Upload file wandb/run-20220228_204859-8xn2plkx/run-8xn2plkx.wandb: 100%|███████████| 35.0M/35.0M [00:02<00:00, 18.3MB/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2813
+ 02/28/2022 22:31:02 - WARNING - huggingface_hub.repository - To https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search
2814
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2815
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2816
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
wandb/run-20220228_204859-8xn2plkx/files/wandb-summary.json CHANGED
The diff for this file is too large to render. See raw diff
wandb/run-20220228_204859-8xn2plkx/logs/debug-internal.log CHANGED
@@ -7540,3 +7540,114 @@
7540
  2022-02-28 22:30:53,346 DEBUG SenderThread:233254 [sender.py:send():235] send: stats
7541
  2022-02-28 22:30:54,327 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: stop_status
7542
  2022-02-28 22:30:54,327 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: stop_status
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7540
  2022-02-28 22:30:53,346 DEBUG SenderThread:233254 [sender.py:send():235] send: stats
7541
  2022-02-28 22:30:54,327 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: stop_status
7542
  2022-02-28 22:30:54,327 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: stop_status
7543
+ 2022-02-28 22:31:02,806 INFO Thread-8 :233254 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/output.log
7544
+ 2022-02-28 22:31:04,807 INFO Thread-8 :233254 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/output.log
7545
+ 2022-02-28 22:31:05,807 INFO Thread-8 :233254 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/output.log
7546
+ 2022-02-28 22:31:08,808 INFO Thread-8 :233254 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/output.log
7547
+ 2022-02-28 22:31:09,327 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7548
+ 2022-02-28 22:31:09,327 DEBUG SenderThread:233254 [sender.py:send():235] send: telemetry
7549
+ 2022-02-28 22:31:09,327 DEBUG SenderThread:233254 [sender.py:send():235] send: exit
7550
+ 2022-02-28 22:31:09,327 INFO SenderThread:233254 [sender.py:send_exit():371] handling exit code: 1
7551
+ 2022-02-28 22:31:09,327 INFO SenderThread:233254 [sender.py:send_exit():373] handling runtime: 6128
7552
+ 2022-02-28 22:31:09,383 INFO SenderThread:233254 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
7553
+ 2022-02-28 22:31:09,383 INFO SenderThread:233254 [sender.py:send_exit():379] send defer
7554
+ 2022-02-28 22:31:09,383 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7555
+ 2022-02-28 22:31:09,384 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7556
+ 2022-02-28 22:31:09,384 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 0
7557
+ 2022-02-28 22:31:09,384 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7558
+ 2022-02-28 22:31:09,384 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 0
7559
+ 2022-02-28 22:31:09,384 INFO SenderThread:233254 [sender.py:transition_state():392] send defer: 1
7560
+ 2022-02-28 22:31:09,384 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7561
+ 2022-02-28 22:31:09,384 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 1
7562
+ 2022-02-28 22:31:09,439 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7563
+ 2022-02-28 22:31:09,439 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 1
7564
+ 2022-02-28 22:31:09,439 INFO SenderThread:233254 [sender.py:transition_state():392] send defer: 2
7565
+ 2022-02-28 22:31:09,439 DEBUG SenderThread:233254 [sender.py:send():235] send: stats
7566
+ 2022-02-28 22:31:09,440 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7567
+ 2022-02-28 22:31:09,440 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 2
7568
+ 2022-02-28 22:31:09,440 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7569
+ 2022-02-28 22:31:09,440 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 2
7570
+ 2022-02-28 22:31:09,440 INFO SenderThread:233254 [sender.py:transition_state():392] send defer: 3
7571
+ 2022-02-28 22:31:09,441 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7572
+ 2022-02-28 22:31:09,441 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 3
7573
+ 2022-02-28 22:31:09,502 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7574
+ 2022-02-28 22:31:09,504 DEBUG SenderThread:233254 [sender.py:send():235] send: summary
7575
+ 2022-02-28 22:31:09,594 INFO SenderThread:233254 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
7576
+ 2022-02-28 22:31:09,594 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7577
+ 2022-02-28 22:31:09,595 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 3
7578
+ 2022-02-28 22:31:09,595 INFO SenderThread:233254 [sender.py:transition_state():392] send defer: 4
7579
+ 2022-02-28 22:31:09,595 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7580
+ 2022-02-28 22:31:09,595 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7581
+ 2022-02-28 22:31:09,596 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 4
7582
+ 2022-02-28 22:31:09,596 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7583
+ 2022-02-28 22:31:09,596 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 4
7584
+ 2022-02-28 22:31:09,697 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7585
+ 2022-02-28 22:31:09,874 INFO Thread-8 :233254 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/output.log
7586
+ 2022-02-28 22:31:09,879 INFO Thread-8 :233254 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/wandb-summary.json
7587
+ 2022-02-28 22:31:10,442 INFO SenderThread:233254 [sender.py:transition_state():392] send defer: 5
7588
+ 2022-02-28 22:31:10,442 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7589
+ 2022-02-28 22:31:10,443 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7590
+ 2022-02-28 22:31:10,443 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 5
7591
+ 2022-02-28 22:31:10,443 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7592
+ 2022-02-28 22:31:10,443 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 5
7593
+ 2022-02-28 22:31:10,443 INFO SenderThread:233254 [dir_watcher.py:finish():283] shutting down directory watcher
7594
+ 2022-02-28 22:31:10,544 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7595
+ 2022-02-28 22:31:10,864 INFO Thread-8 :233254 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/config.yaml
7596
+ 2022-02-28 22:31:10,864 INFO SenderThread:233254 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files
7597
+ 2022-02-28 22:31:10,865 INFO SenderThread:233254 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/wandb-metadata.json wandb-metadata.json
7598
+ 2022-02-28 22:31:10,865 INFO SenderThread:233254 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/output.log output.log
7599
+ 2022-02-28 22:31:10,865 INFO SenderThread:233254 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/wandb-summary.json wandb-summary.json
7600
+ 2022-02-28 22:31:10,868 INFO SenderThread:233254 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/requirements.txt requirements.txt
7601
+ 2022-02-28 22:31:10,868 INFO SenderThread:233254 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/config.yaml config.yaml
7602
+ 2022-02-28 22:31:10,868 INFO SenderThread:233254 [sender.py:transition_state():392] send defer: 6
7603
+ 2022-02-28 22:31:10,868 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7604
+ 2022-02-28 22:31:10,877 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7605
+ 2022-02-28 22:31:10,877 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 6
7606
+ 2022-02-28 22:31:10,877 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7607
+ 2022-02-28 22:31:10,877 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 6
7608
+ 2022-02-28 22:31:10,877 INFO SenderThread:233254 [file_pusher.py:finish():177] shutting down file pusher
7609
+ 2022-02-28 22:31:10,970 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7610
+ 2022-02-28 22:31:10,972 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7611
+ 2022-02-28 22:31:11,074 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7612
+ 2022-02-28 22:31:11,074 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7613
+ 2022-02-28 22:31:11,176 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7614
+ 2022-02-28 22:31:11,176 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7615
+ 2022-02-28 22:31:11,182 INFO Thread-12 :233254 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/output.log
7616
+ 2022-02-28 22:31:11,212 INFO Thread-15 :233254 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/config.yaml
7617
+ 2022-02-28 22:31:11,252 INFO Thread-13 :233254 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/wandb-summary.json
7618
+ 2022-02-28 22:31:11,263 INFO Thread-14 :233254 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/files/requirements.txt
7619
+ 2022-02-28 22:31:11,277 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7620
+ 2022-02-28 22:31:11,278 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7621
+ 2022-02-28 22:31:11,379 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7622
+ 2022-02-28 22:31:11,379 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7623
+ 2022-02-28 22:31:11,463 INFO Thread-7 :233254 [sender.py:transition_state():392] send defer: 7
7624
+ 2022-02-28 22:31:11,464 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7625
+ 2022-02-28 22:31:11,464 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 7
7626
+ 2022-02-28 22:31:11,464 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7627
+ 2022-02-28 22:31:11,464 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 7
7628
+ 2022-02-28 22:31:11,481 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7629
+ 2022-02-28 22:31:13,480 INFO SenderThread:233254 [sender.py:transition_state():392] send defer: 8
7630
+ 2022-02-28 22:31:13,480 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7631
+ 2022-02-28 22:31:13,481 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7632
+ 2022-02-28 22:31:13,481 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 8
7633
+ 2022-02-28 22:31:13,481 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7634
+ 2022-02-28 22:31:13,481 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 8
7635
+ 2022-02-28 22:31:13,481 INFO SenderThread:233254 [sender.py:transition_state():392] send defer: 9
7636
+ 2022-02-28 22:31:13,482 DEBUG SenderThread:233254 [sender.py:send():235] send: final
7637
+ 2022-02-28 22:31:13,483 DEBUG SenderThread:233254 [sender.py:send():235] send: footer
7638
+ 2022-02-28 22:31:13,483 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: defer
7639
+ 2022-02-28 22:31:13,484 INFO HandlerThread:233254 [handler.py:handle_request_defer():154] handle defer: 9
7640
+ 2022-02-28 22:31:13,484 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: defer
7641
+ 2022-02-28 22:31:13,484 INFO SenderThread:233254 [sender.py:send_request_defer():388] handle sender defer: 9
7642
+ 2022-02-28 22:31:13,582 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: poll_exit
7643
+ 2022-02-28 22:31:13,582 DEBUG SenderThread:233254 [sender.py:send_request():249] send_request: poll_exit
7644
+ 2022-02-28 22:31:13,582 INFO SenderThread:233254 [file_pusher.py:join():182] waiting for file pusher
7645
+ 2022-02-28 22:31:13,635 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: get_summary
7646
+ 2022-02-28 22:31:13,738 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: sampled_history
7647
+ 2022-02-28 22:31:13,741 DEBUG HandlerThread:233254 [handler.py:handle_request():131] handle_request: shutdown
7648
+ 2022-02-28 22:31:13,741 INFO HandlerThread:233254 [handler.py:finish():739] shutting down handler
7649
+ 2022-02-28 22:31:14,484 INFO WriterThread:233254 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_204859-8xn2plkx/run-8xn2plkx.wandb
7650
+ 2022-02-28 22:31:14,634 INFO SenderThread:233254 [sender.py:finish():1075] shutting down sender
7651
+ 2022-02-28 22:31:14,634 INFO SenderThread:233254 [file_pusher.py:finish():177] shutting down file pusher
7652
+ 2022-02-28 22:31:14,634 INFO SenderThread:233254 [file_pusher.py:join():182] waiting for file pusher
7653
+ 2022-02-28 22:31:14,641 INFO MainThread:233254 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220228_204859-8xn2plkx/logs/debug.log CHANGED
@@ -24,3 +24,101 @@ config: {}
24
  2022-02-28 20:49:00,664 INFO MainThread:233154 [wandb_init.py:init():651] run started, returning control to user process
25
  2022-02-28 20:49:00,666 INFO MainThread:233154 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 1e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Feb28_20-48-17_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 12, 'eval_batch_size': 12}
26
  2022-02-28 20:49:00,669 INFO MainThread:233154 [wandb_watch.py:watch():43] Watching
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  2022-02-28 20:49:00,664 INFO MainThread:233154 [wandb_init.py:init():651] run started, returning control to user process
25
  2022-02-28 20:49:00,666 INFO MainThread:233154 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 1e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Feb28_20-48-17_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 12, 'eval_batch_size': 12}
26
  2022-02-28 20:49:00,669 INFO MainThread:233154 [wandb_watch.py:watch():43] Watching
27
+ 2022-02-28 22:31:06,482 INFO MainThread:233154 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
28
+ 2022-02-28 22:31:06,485 INFO MainThread:233154 [wandb_run.py:_restore():1769] restore
29
+ 2022-02-28 22:31:09,384 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
30
+ wandb_count: 1
31
+ }
32
+ pusher_stats {
33
+ uploaded_bytes: 2095
34
+ total_bytes: 2095
35
+ }
36
+
37
+ 2022-02-28 22:31:09,596 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
38
+ wandb_count: 1
39
+ }
40
+ pusher_stats {
41
+ uploaded_bytes: 2095
42
+ total_bytes: 2095
43
+ }
44
+
45
+ 2022-02-28 22:31:10,443 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
46
+ wandb_count: 1
47
+ }
48
+ pusher_stats {
49
+ uploaded_bytes: 2095
50
+ total_bytes: 2095
51
+ }
52
+
53
+ 2022-02-28 22:31:10,869 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
54
+ wandb_count: 2
55
+ }
56
+ pusher_stats {
57
+ uploaded_bytes: 2095
58
+ total_bytes: 893027
59
+ }
60
+
61
+ 2022-02-28 22:31:10,973 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
62
+ wandb_count: 5
63
+ }
64
+ pusher_stats {
65
+ uploaded_bytes: 2095
66
+ total_bytes: 2749635
67
+ }
68
+
69
+ 2022-02-28 22:31:11,075 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
70
+ wandb_count: 5
71
+ }
72
+ pusher_stats {
73
+ uploaded_bytes: 2749635
74
+ total_bytes: 2749635
75
+ }
76
+
77
+ 2022-02-28 22:31:11,177 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
78
+ wandb_count: 5
79
+ }
80
+ pusher_stats {
81
+ uploaded_bytes: 2749635
82
+ total_bytes: 2749635
83
+ }
84
+
85
+ 2022-02-28 22:31:11,278 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
86
+ wandb_count: 5
87
+ }
88
+ pusher_stats {
89
+ uploaded_bytes: 2749635
90
+ total_bytes: 2749635
91
+ }
92
+
93
+ 2022-02-28 22:31:11,380 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
94
+ wandb_count: 5
95
+ }
96
+ pusher_stats {
97
+ uploaded_bytes: 2749635
98
+ total_bytes: 2749635
99
+ }
100
+
101
+ 2022-02-28 22:31:13,481 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
102
+ wandb_count: 5
103
+ }
104
+ pusher_stats {
105
+ uploaded_bytes: 2749635
106
+ total_bytes: 2749635
107
+ }
108
+
109
+ 2022-02-28 22:31:13,635 INFO MainThread:233154 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
110
+ exit_result {
111
+ }
112
+ file_counts {
113
+ wandb_count: 5
114
+ }
115
+ pusher_stats {
116
+ uploaded_bytes: 2749635
117
+ total_bytes: 2749635
118
+ }
119
+ local_info {
120
+ }
121
+
122
+ 2022-02-28 22:31:14,790 INFO MainThread:233154 [wandb_run.py:_append_history():2144] rendering history
123
+ 2022-02-28 22:31:14,791 INFO MainThread:233154 [wandb_run.py:_append_summary():2102] rendering summary
124
+ 2022-02-28 22:31:14,793 INFO MainThread:233154 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220228_204859-8xn2plkx/run-8xn2plkx.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:301e25b5082debd52df07c2c5e3e46ee673b775c33c5a7bef42d7885f9e4611a
3
- size 36752475
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:079286aac718257498de20c8410d4b385f1a29460677456b1dc94ec40ddad11c
3
+ size 38316785
wandb/run-20220228_223243-2ay2wvge/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220228_223243-2ay2wvge/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220228_223243-2ay2wvge/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220228_223243-2ay2wvge/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-02-28T22:32:44.226122",
5
+ "startedAt": "2022-02-28T22:32:43.089309",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=2",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "7dec2191123af092d2edc7951267a70d7948caba"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220228_223243-2ay2wvge/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220228_223243-2ay2wvge/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220228_223243-2ay2wvge/logs/debug.log ADDED
@@ -0,0 +1,124 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-02-28 22:32:43,095 INFO MainThread:234573 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-02-28 22:32:43,095 INFO MainThread:234573 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-02-28 22:32:43,095 INFO MainThread:234573 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-02-28 22:32:43,095 INFO MainThread:234573 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-02-28 22:32:43,095 INFO MainThread:234573 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_223243-2ay2wvge/logs/debug.log
6
+ 2022-02-28 22:32:43,095 INFO MainThread:234573 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_223243-2ay2wvge/logs/debug-internal.log
7
+ 2022-02-28 22:32:43,095 INFO MainThread:234573 [wandb_init.py:init():420] calling init triggers
8
+ 2022-02-28 22:32:43,095 INFO MainThread:234573 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-02-28 22:32:43,095 INFO MainThread:234573 [wandb_init.py:init():471] starting backend
11
+ 2022-02-28 22:32:43,096 INFO MainThread:234573 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-02-28 22:32:43,166 INFO MainThread:234573 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-02-28 22:32:43,235 INFO MainThread:234573 [backend.py:ensure_launched():224] started backend process with pid: 234672
14
+ 2022-02-28 22:32:43,237 INFO MainThread:234573 [wandb_init.py:init():480] backend started and connected
15
+ 2022-02-28 22:32:43,247 INFO MainThread:234573 [wandb_init.py:init():550] updated telemetry
16
+ 2022-02-28 22:32:43,413 INFO MainThread:234573 [wandb_init.py:init():581] communicating current version
17
+ 2022-02-28 22:32:44,124 INFO MainThread:234573 [wandb_init.py:init():586] got version response
18
+ 2022-02-28 22:32:44,124 INFO MainThread:234573 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
19
+ 2022-02-28 22:32:44,220 INFO MainThread:234573 [wandb_init.py:init():624] starting run threads in backend
20
+ 2022-02-28 22:32:44,355 INFO MainThread:234573 [wandb_run.py:_console_start():1827] atexit reg
21
+ 2022-02-28 22:32:44,355 INFO MainThread:234573 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
22
+ 2022-02-28 22:32:44,356 INFO MainThread:234573 [wandb_run.py:_redirect():1706] Redirecting console.
23
+ 2022-02-28 22:32:44,358 INFO MainThread:234573 [wandb_run.py:_redirect():1762] Redirects installed.
24
+ 2022-02-28 22:32:44,358 INFO MainThread:234573 [wandb_init.py:init():651] run started, returning control to user process
25
+ 2022-02-28 22:32:44,361 INFO MainThread:234573 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 1e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Feb28_22-32-00_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
26
+ 2022-02-28 22:32:44,363 INFO MainThread:234573 [wandb_watch.py:watch():43] Watching
27
+ 2022-02-28 23:12:33,105 INFO MainThread:234573 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
28
+ 2022-02-28 23:12:33,106 INFO MainThread:234573 [wandb_run.py:_restore():1769] restore
29
+ 2022-02-28 23:12:35,607 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
30
+ wandb_count: 1
31
+ }
32
+ pusher_stats {
33
+ uploaded_bytes: 2095
34
+ total_bytes: 2095
35
+ }
36
+
37
+ 2022-02-28 23:12:35,728 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
38
+ wandb_count: 1
39
+ }
40
+ pusher_stats {
41
+ uploaded_bytes: 2095
42
+ total_bytes: 2095
43
+ }
44
+
45
+ 2022-02-28 23:12:35,883 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
46
+ wandb_count: 1
47
+ }
48
+ pusher_stats {
49
+ uploaded_bytes: 2095
50
+ total_bytes: 2095
51
+ }
52
+
53
+ 2022-02-28 23:12:36,755 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
54
+ wandb_count: 1
55
+ }
56
+ pusher_stats {
57
+ uploaded_bytes: 2095
58
+ total_bytes: 2095
59
+ }
60
+
61
+ 2022-02-28 23:12:36,862 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
62
+ wandb_count: 5
63
+ }
64
+ pusher_stats {
65
+ uploaded_bytes: 2095
66
+ total_bytes: 2224381
67
+ }
68
+
69
+ 2022-02-28 23:12:36,964 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
70
+ wandb_count: 5
71
+ }
72
+ pusher_stats {
73
+ uploaded_bytes: 2224381
74
+ total_bytes: 2224381
75
+ }
76
+
77
+ 2022-02-28 23:12:37,066 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
78
+ wandb_count: 5
79
+ }
80
+ pusher_stats {
81
+ uploaded_bytes: 2224381
82
+ total_bytes: 2224381
83
+ }
84
+
85
+ 2022-02-28 23:12:37,168 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
86
+ wandb_count: 5
87
+ }
88
+ pusher_stats {
89
+ uploaded_bytes: 2224381
90
+ total_bytes: 2224381
91
+ }
92
+
93
+ 2022-02-28 23:12:37,270 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
94
+ wandb_count: 5
95
+ }
96
+ pusher_stats {
97
+ uploaded_bytes: 2224381
98
+ total_bytes: 2224381
99
+ }
100
+
101
+ 2022-02-28 23:12:38,781 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
102
+ wandb_count: 5
103
+ }
104
+ pusher_stats {
105
+ uploaded_bytes: 2224381
106
+ total_bytes: 2224381
107
+ }
108
+
109
+ 2022-02-28 23:12:38,939 INFO MainThread:234573 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
110
+ exit_result {
111
+ }
112
+ file_counts {
113
+ wandb_count: 5
114
+ }
115
+ pusher_stats {
116
+ uploaded_bytes: 2224381
117
+ total_bytes: 2224381
118
+ }
119
+ local_info {
120
+ }
121
+
122
+ 2022-02-28 23:12:40,092 INFO MainThread:234573 [wandb_run.py:_append_history():2144] rendering history
123
+ 2022-02-28 23:12:40,093 INFO MainThread:234573 [wandb_run.py:_append_summary():2102] rendering summary
124
+ 2022-02-28 23:12:40,094 INFO MainThread:234573 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220228_223243-2ay2wvge/run-2ay2wvge.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37c24c3ed991ccf81c406a83dcccf31e6858a245ee1b74b4bf6eaa756131c1a3
3
+ size 17454252
wandb/run-20220228_231357-3lq2qpez/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220228_231357-3lq2qpez/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220228_231357-3lq2qpez/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220228_231357-3lq2qpez/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-02-28T23:13:58.299210",
5
+ "startedAt": "2022-02-28T23:13:57.257935",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=4",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "7dec2191123af092d2edc7951267a70d7948caba"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220228_231357-3lq2qpez/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220228_231357-3lq2qpez/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220228_231357-3lq2qpez/logs/debug.log ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-02-28 23:13:57,263 INFO MainThread:234775 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-02-28 23:13:57,264 INFO MainThread:234775 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-02-28 23:13:57,264 INFO MainThread:234775 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-02-28 23:13:57,264 INFO MainThread:234775 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-02-28 23:13:57,264 INFO MainThread:234775 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_231357-3lq2qpez/logs/debug.log
6
+ 2022-02-28 23:13:57,264 INFO MainThread:234775 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220228_231357-3lq2qpez/logs/debug-internal.log
7
+ 2022-02-28 23:13:57,264 INFO MainThread:234775 [wandb_init.py:init():420] calling init triggers
8
+ 2022-02-28 23:13:57,264 INFO MainThread:234775 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-02-28 23:13:57,264 INFO MainThread:234775 [wandb_init.py:init():471] starting backend
11
+ 2022-02-28 23:13:57,264 INFO MainThread:234775 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-02-28 23:13:57,314 INFO MainThread:234775 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-02-28 23:13:57,363 INFO MainThread:234775 [backend.py:ensure_launched():224] started backend process with pid: 234874
14
+ 2022-02-28 23:13:57,365 INFO MainThread:234775 [wandb_init.py:init():480] backend started and connected
15
+ 2022-02-28 23:13:57,375 INFO MainThread:234775 [wandb_init.py:init():550] updated telemetry
16
+ 2022-02-28 23:13:57,495 INFO MainThread:234775 [wandb_init.py:init():581] communicating current version
17
+ 2022-02-28 23:13:58,203 INFO MainThread:234775 [wandb_init.py:init():586] got version response
18
+ 2022-02-28 23:13:58,203 INFO MainThread:234775 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
19
+ 2022-02-28 23:13:58,294 INFO MainThread:234775 [wandb_init.py:init():624] starting run threads in backend
20
+ 2022-02-28 23:13:58,405 INFO MainThread:234775 [wandb_run.py:_console_start():1827] atexit reg
21
+ 2022-02-28 23:13:58,406 INFO MainThread:234775 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
22
+ 2022-02-28 23:13:58,406 INFO MainThread:234775 [wandb_run.py:_redirect():1706] Redirecting console.
23
+ 2022-02-28 23:13:58,408 INFO MainThread:234775 [wandb_run.py:_redirect():1762] Redirects installed.
24
+ 2022-02-28 23:13:58,408 INFO MainThread:234775 [wandb_init.py:init():651] run started, returning control to user process
25
+ 2022-02-28 23:13:58,410 INFO MainThread:234775 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': 'None', 'learning_rate': 1e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Feb28_23-13-16_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
26
+ 2022-02-28 23:13:58,413 INFO MainThread:234775 [wandb_watch.py:watch():43] Watching
27
+ 2022-03-01 00:23:16,488 INFO MainThread:234775 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
28
+ 2022-03-01 00:23:16,490 INFO MainThread:234775 [wandb_run.py:_restore():1769] restore
29
+ 2022-03-01 00:23:18,612 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
30
+ wandb_count: 1
31
+ }
32
+ pusher_stats {
33
+ uploaded_bytes: 2095
34
+ total_bytes: 2095
35
+ }
36
+
37
+ 2022-03-01 00:23:18,814 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
38
+ wandb_count: 1
39
+ }
40
+ pusher_stats {
41
+ uploaded_bytes: 2095
42
+ total_bytes: 2095
43
+ }
44
+
45
+ 2022-03-01 00:23:19,629 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
46
+ wandb_count: 1
47
+ }
48
+ pusher_stats {
49
+ uploaded_bytes: 2095
50
+ total_bytes: 2095
51
+ }
52
+
53
+ 2022-03-01 00:23:19,937 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
54
+ wandb_count: 5
55
+ }
56
+ pusher_stats {
57
+ uploaded_bytes: 2095
58
+ total_bytes: 2505320
59
+ }
60
+
61
+ 2022-03-01 00:23:20,040 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
62
+ wandb_count: 5
63
+ }
64
+ pusher_stats {
65
+ uploaded_bytes: 646971
66
+ total_bytes: 2505320
67
+ }
68
+
69
+ 2022-03-01 00:23:20,141 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
70
+ wandb_count: 5
71
+ }
72
+ pusher_stats {
73
+ uploaded_bytes: 2505320
74
+ total_bytes: 2505320
75
+ }
76
+
77
+ 2022-03-01 00:23:20,242 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
78
+ wandb_count: 5
79
+ }
80
+ pusher_stats {
81
+ uploaded_bytes: 2505320
82
+ total_bytes: 2505320
83
+ }
84
+
85
+ 2022-03-01 00:23:20,344 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
86
+ wandb_count: 5
87
+ }
88
+ pusher_stats {
89
+ uploaded_bytes: 2505320
90
+ total_bytes: 2505320
91
+ }
92
+
93
+ 2022-03-01 00:23:20,446 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
94
+ wandb_count: 5
95
+ }
96
+ pusher_stats {
97
+ uploaded_bytes: 2505320
98
+ total_bytes: 2505320
99
+ }
100
+
101
+ 2022-03-01 00:23:20,548 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
102
+ wandb_count: 5
103
+ }
104
+ pusher_stats {
105
+ uploaded_bytes: 2505320
106
+ total_bytes: 2505320
107
+ }
108
+
109
+ 2022-03-01 00:23:22,347 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
110
+ wandb_count: 5
111
+ }
112
+ pusher_stats {
113
+ uploaded_bytes: 2505320
114
+ total_bytes: 2505320
115
+ }
116
+
117
+ 2022-03-01 00:23:22,511 INFO MainThread:234775 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
118
+ exit_result {
119
+ }
120
+ file_counts {
121
+ wandb_count: 5
122
+ }
123
+ pusher_stats {
124
+ uploaded_bytes: 2505320
125
+ total_bytes: 2505320
126
+ }
127
+ local_info {
128
+ }
129
+
130
+ 2022-03-01 00:23:23,669 INFO MainThread:234775 [wandb_run.py:_append_history():2144] rendering history
131
+ 2022-03-01 00:23:23,669 INFO MainThread:234775 [wandb_run.py:_append_summary():2102] rendering summary
132
+ 2022-03-01 00:23:23,670 INFO MainThread:234775 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220228_231357-3lq2qpez/run-3lq2qpez.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27e97ee4fdcfeff072679f67101808be953381985fa5fac8cf5d7cf2e1244dad
3
+ size 30431018
wandb/run-20220301_002446-2vmlu6y4/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_002446-2vmlu6y4/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_002446-2vmlu6y4/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220301_002446-2vmlu6y4/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-01T00:24:47.733307",
5
+ "startedAt": "2022-03-01T00:24:46.612755",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=8",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "7dec2191123af092d2edc7951267a70d7948caba"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220301_002446-2vmlu6y4/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_002446-2vmlu6y4/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_002446-2vmlu6y4/logs/debug.log ADDED
@@ -0,0 +1,124 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 00:24:46,617 INFO MainThread:236184 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-01 00:24:46,618 INFO MainThread:236184 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-01 00:24:46,618 INFO MainThread:236184 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-01 00:24:46,618 INFO MainThread:236184 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-01 00:24:46,618 INFO MainThread:236184 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_002446-2vmlu6y4/logs/debug.log
6
+ 2022-03-01 00:24:46,618 INFO MainThread:236184 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_002446-2vmlu6y4/logs/debug-internal.log
7
+ 2022-03-01 00:24:46,618 INFO MainThread:236184 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-01 00:24:46,618 INFO MainThread:236184 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-01 00:24:46,618 INFO MainThread:236184 [wandb_init.py:init():471] starting backend
11
+ 2022-03-01 00:24:46,618 INFO MainThread:236184 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-01 00:24:46,683 INFO MainThread:236184 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-01 00:24:46,745 INFO MainThread:236184 [backend.py:ensure_launched():224] started backend process with pid: 236283
14
+ 2022-03-01 00:24:46,748 INFO MainThread:236184 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-01 00:24:46,758 INFO MainThread:236184 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-01 00:24:46,910 INFO MainThread:236184 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-01 00:24:47,606 INFO MainThread:236184 [wandb_init.py:init():586] got version response
18
+ 2022-03-01 00:24:47,607 INFO MainThread:236184 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
19
+ 2022-03-01 00:24:47,728 INFO MainThread:236184 [wandb_init.py:init():624] starting run threads in backend
20
+ 2022-03-01 00:24:47,834 INFO MainThread:236184 [wandb_run.py:_console_start():1827] atexit reg
21
+ 2022-03-01 00:24:47,834 INFO MainThread:236184 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
22
+ 2022-03-01 00:24:47,835 INFO MainThread:236184 [wandb_run.py:_redirect():1706] Redirecting console.
23
+ 2022-03-01 00:24:47,836 INFO MainThread:236184 [wandb_run.py:_redirect():1762] Redirects installed.
24
+ 2022-03-01 00:24:47,836 INFO MainThread:236184 [wandb_init.py:init():651] run started, returning control to user process
25
+ 2022-03-01 00:24:47,839 INFO MainThread:236184 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 1e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar01_00-24-05_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
26
+ 2022-03-01 00:24:47,842 INFO MainThread:236184 [wandb_watch.py:watch():43] Watching
27
+ 2022-03-01 01:25:25,706 INFO MainThread:236184 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
28
+ 2022-03-01 01:25:25,708 INFO MainThread:236184 [wandb_run.py:_restore():1769] restore
29
+ 2022-03-01 01:25:27,881 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
30
+ wandb_count: 1
31
+ }
32
+ pusher_stats {
33
+ uploaded_bytes: 2095
34
+ total_bytes: 2095
35
+ }
36
+
37
+ 2022-03-01 01:25:28,060 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
38
+ wandb_count: 1
39
+ }
40
+ pusher_stats {
41
+ uploaded_bytes: 2095
42
+ total_bytes: 2095
43
+ }
44
+
45
+ 2022-03-01 01:25:28,846 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
46
+ wandb_count: 1
47
+ }
48
+ pusher_stats {
49
+ uploaded_bytes: 2095
50
+ total_bytes: 2095
51
+ }
52
+
53
+ 2022-03-01 01:25:29,109 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
54
+ wandb_count: 5
55
+ }
56
+ pusher_stats {
57
+ uploaded_bytes: 2095
58
+ total_bytes: 2335856
59
+ }
60
+
61
+ 2022-03-01 01:25:29,210 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
62
+ wandb_count: 5
63
+ }
64
+ pusher_stats {
65
+ uploaded_bytes: 522679
66
+ total_bytes: 2335856
67
+ }
68
+
69
+ 2022-03-01 01:25:29,312 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
70
+ wandb_count: 5
71
+ }
72
+ pusher_stats {
73
+ uploaded_bytes: 2335856
74
+ total_bytes: 2335856
75
+ }
76
+
77
+ 2022-03-01 01:25:29,414 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
78
+ wandb_count: 5
79
+ }
80
+ pusher_stats {
81
+ uploaded_bytes: 2335856
82
+ total_bytes: 2335856
83
+ }
84
+
85
+ 2022-03-01 01:25:29,515 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
86
+ wandb_count: 5
87
+ }
88
+ pusher_stats {
89
+ uploaded_bytes: 2335856
90
+ total_bytes: 2335856
91
+ }
92
+
93
+ 2022-03-01 01:25:29,617 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
94
+ wandb_count: 5
95
+ }
96
+ pusher_stats {
97
+ uploaded_bytes: 2335856
98
+ total_bytes: 2335856
99
+ }
100
+
101
+ 2022-03-01 01:25:31,119 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
102
+ wandb_count: 5
103
+ }
104
+ pusher_stats {
105
+ uploaded_bytes: 2335856
106
+ total_bytes: 2335856
107
+ }
108
+
109
+ 2022-03-01 01:25:31,278 INFO MainThread:236184 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
110
+ exit_result {
111
+ }
112
+ file_counts {
113
+ wandb_count: 5
114
+ }
115
+ pusher_stats {
116
+ uploaded_bytes: 2335856
117
+ total_bytes: 2335856
118
+ }
119
+ local_info {
120
+ }
121
+
122
+ 2022-03-01 01:25:32,421 INFO MainThread:236184 [wandb_run.py:_append_history():2144] rendering history
123
+ 2022-03-01 01:25:32,422 INFO MainThread:236184 [wandb_run.py:_append_summary():2102] rendering summary
124
+ 2022-03-01 01:25:32,423 INFO MainThread:236184 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220301_002446-2vmlu6y4/run-2vmlu6y4.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09662ce7cff1fa63f1309c6fc17704d0eb0b30e02d14a1dd4ad117d6c10f3d90
3
+ size 26865135
wandb/run-20220301_012648-nzue7jk0/files/config.yaml ADDED
@@ -0,0 +1,713 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: ./
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.10
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.9.5
22
+ start_time: 1646098009
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.9.5
37
+ 5: 0.12.10
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ adafactor:
42
+ desc: null
43
+ value: false
44
+ adam_beta1:
45
+ desc: null
46
+ value: 0.9
47
+ adam_beta2:
48
+ desc: null
49
+ value: 0.999
50
+ adam_epsilon:
51
+ desc: null
52
+ value: 1.0e-08
53
+ add_cross_attention:
54
+ desc: null
55
+ value: false
56
+ architectures:
57
+ desc: null
58
+ value:
59
+ - SpeechEncoderDecoderModel
60
+ bad_words_ids:
61
+ desc: null
62
+ value: null
63
+ bf16:
64
+ desc: null
65
+ value: false
66
+ bf16_full_eval:
67
+ desc: null
68
+ value: false
69
+ bos_token_id:
70
+ desc: null
71
+ value: null
72
+ chunk_size_feed_forward:
73
+ desc: null
74
+ value: 0
75
+ cross_attention_hidden_size:
76
+ desc: null
77
+ value: null
78
+ dataloader_drop_last:
79
+ desc: null
80
+ value: false
81
+ dataloader_num_workers:
82
+ desc: null
83
+ value: 0
84
+ dataloader_pin_memory:
85
+ desc: null
86
+ value: true
87
+ ddp_bucket_cap_mb:
88
+ desc: null
89
+ value: None
90
+ ddp_find_unused_parameters:
91
+ desc: null
92
+ value: None
93
+ debug:
94
+ desc: null
95
+ value: '[]'
96
+ decoder:
97
+ desc: null
98
+ value:
99
+ _name_or_path: gpt2-medium
100
+ activation_function: gelu_new
101
+ add_cross_attention: true
102
+ architectures:
103
+ - GPT2LMHeadModel
104
+ attn_pdrop: 0.0
105
+ bad_words_ids: null
106
+ bos_token_id: 50256
107
+ chunk_size_feed_forward: 0
108
+ cross_attention_hidden_size: null
109
+ decoder_start_token_id: null
110
+ diversity_penalty: 0.0
111
+ do_sample: false
112
+ early_stopping: false
113
+ embd_pdrop: 0.0
114
+ encoder_no_repeat_ngram_size: 0
115
+ eos_token_id: 50256
116
+ finetuning_task: null
117
+ forced_bos_token_id: null
118
+ forced_eos_token_id: null
119
+ id2label:
120
+ '0': LABEL_0
121
+ '1': LABEL_1
122
+ initializer_range: 0.02
123
+ is_decoder: true
124
+ is_encoder_decoder: false
125
+ label2id:
126
+ LABEL_0: 0
127
+ LABEL_1: 1
128
+ layer_norm_epsilon: 1.0e-05
129
+ length_penalty: 1.0
130
+ max_length: 20
131
+ min_length: 0
132
+ model_type: gpt2
133
+ n_ctx: 1024
134
+ n_embd: 1024
135
+ n_head: 16
136
+ n_inner: null
137
+ n_layer: 24
138
+ n_positions: 1024
139
+ n_special: 0
140
+ no_repeat_ngram_size: 0
141
+ num_beam_groups: 1
142
+ num_beams: 1
143
+ num_return_sequences: 1
144
+ output_attentions: false
145
+ output_hidden_states: false
146
+ output_scores: false
147
+ pad_token_id: null
148
+ predict_special_tokens: true
149
+ prefix: null
150
+ problem_type: null
151
+ pruned_heads: {}
152
+ remove_invalid_values: false
153
+ reorder_and_upcast_attn: false
154
+ repetition_penalty: 1.0
155
+ resid_pdrop: 0.0
156
+ return_dict: true
157
+ return_dict_in_generate: false
158
+ scale_attn_by_inverse_layer_idx: false
159
+ scale_attn_weights: true
160
+ sep_token_id: null
161
+ summary_activation: null
162
+ summary_first_dropout: 0.0
163
+ summary_proj_to_labels: true
164
+ summary_type: cls_index
165
+ summary_use_proj: true
166
+ task_specific_params:
167
+ text-generation:
168
+ do_sample: true
169
+ max_length: 50
170
+ temperature: 1.0
171
+ tie_encoder_decoder: false
172
+ tie_word_embeddings: true
173
+ tokenizer_class: null
174
+ top_k: 50
175
+ top_p: 1.0
176
+ torch_dtype: null
177
+ torchscript: false
178
+ transformers_version: 4.17.0.dev0
179
+ use_bfloat16: false
180
+ use_cache: false
181
+ vocab_size: 50257
182
+ decoder_start_token_id:
183
+ desc: null
184
+ value: 50256
185
+ deepspeed:
186
+ desc: null
187
+ value: None
188
+ disable_tqdm:
189
+ desc: null
190
+ value: false
191
+ diversity_penalty:
192
+ desc: null
193
+ value: 0.0
194
+ do_eval:
195
+ desc: null
196
+ value: true
197
+ do_predict:
198
+ desc: null
199
+ value: false
200
+ do_sample:
201
+ desc: null
202
+ value: false
203
+ do_train:
204
+ desc: null
205
+ value: true
206
+ early_stopping:
207
+ desc: null
208
+ value: false
209
+ encoder:
210
+ desc: null
211
+ value:
212
+ _name_or_path: facebook/wav2vec2-large-lv60
213
+ activation_dropout: 0.0
214
+ adapter_kernel_size: 3
215
+ adapter_stride: 2
216
+ add_adapter: true
217
+ add_cross_attention: false
218
+ apply_spec_augment: false
219
+ architectures:
220
+ - Wav2Vec2ForPreTraining
221
+ attention_dropout: 0.0
222
+ bad_words_ids: null
223
+ bos_token_id: 1
224
+ chunk_size_feed_forward: 0
225
+ classifier_proj_size: 256
226
+ codevector_dim: 768
227
+ contrastive_logits_temperature: 0.1
228
+ conv_bias: true
229
+ conv_dim:
230
+ - 512
231
+ - 512
232
+ - 512
233
+ - 512
234
+ - 512
235
+ - 512
236
+ - 512
237
+ conv_kernel:
238
+ - 10
239
+ - 3
240
+ - 3
241
+ - 3
242
+ - 3
243
+ - 2
244
+ - 2
245
+ conv_stride:
246
+ - 5
247
+ - 2
248
+ - 2
249
+ - 2
250
+ - 2
251
+ - 2
252
+ - 2
253
+ cross_attention_hidden_size: null
254
+ ctc_loss_reduction: sum
255
+ ctc_zero_infinity: false
256
+ decoder_start_token_id: null
257
+ diversity_loss_weight: 0.1
258
+ diversity_penalty: 0.0
259
+ do_sample: false
260
+ do_stable_layer_norm: true
261
+ early_stopping: false
262
+ encoder_no_repeat_ngram_size: 0
263
+ eos_token_id: 2
264
+ feat_extract_activation: gelu
265
+ feat_extract_dropout: 0.0
266
+ feat_extract_norm: layer
267
+ feat_proj_dropout: 0.0
268
+ feat_quantizer_dropout: 0.0
269
+ final_dropout: 0.0
270
+ finetuning_task: null
271
+ forced_bos_token_id: null
272
+ forced_eos_token_id: null
273
+ gradient_checkpointing: false
274
+ hidden_act: gelu
275
+ hidden_dropout: 0.0
276
+ hidden_dropout_prob: 0.0
277
+ hidden_size: 1024
278
+ id2label:
279
+ '0': LABEL_0
280
+ '1': LABEL_1
281
+ initializer_range: 0.02
282
+ intermediate_size: 4096
283
+ is_decoder: false
284
+ is_encoder_decoder: false
285
+ label2id:
286
+ LABEL_0: 0
287
+ LABEL_1: 1
288
+ layer_norm_eps: 1.0e-05
289
+ layerdrop: 0.0
290
+ length_penalty: 1.0
291
+ mask_feature_length: 10
292
+ mask_feature_min_masks: 0
293
+ mask_feature_prob: 0.0
294
+ mask_time_length: 10
295
+ mask_time_min_masks: 2
296
+ mask_time_prob: 0.0
297
+ max_length: 20
298
+ min_length: 0
299
+ model_type: wav2vec2
300
+ no_repeat_ngram_size: 0
301
+ num_adapter_layers: 3
302
+ num_attention_heads: 16
303
+ num_beam_groups: 1
304
+ num_beams: 1
305
+ num_codevector_groups: 2
306
+ num_codevectors_per_group: 320
307
+ num_conv_pos_embedding_groups: 16
308
+ num_conv_pos_embeddings: 128
309
+ num_feat_extract_layers: 7
310
+ num_hidden_layers: 24
311
+ num_negatives: 100
312
+ num_return_sequences: 1
313
+ output_attentions: false
314
+ output_hidden_size: 1024
315
+ output_hidden_states: false
316
+ output_scores: false
317
+ pad_token_id: 0
318
+ prefix: null
319
+ problem_type: null
320
+ proj_codevector_dim: 768
321
+ pruned_heads: {}
322
+ remove_invalid_values: false
323
+ repetition_penalty: 1.0
324
+ return_dict: true
325
+ return_dict_in_generate: false
326
+ sep_token_id: null
327
+ task_specific_params: null
328
+ tdnn_dilation:
329
+ - 1
330
+ - 2
331
+ - 3
332
+ - 1
333
+ - 1
334
+ tdnn_dim:
335
+ - 512
336
+ - 512
337
+ - 512
338
+ - 512
339
+ - 1500
340
+ tdnn_kernel:
341
+ - 5
342
+ - 3
343
+ - 3
344
+ - 1
345
+ - 1
346
+ temperature: 1.0
347
+ tie_encoder_decoder: false
348
+ tie_word_embeddings: true
349
+ tokenizer_class: null
350
+ top_k: 50
351
+ top_p: 1.0
352
+ torch_dtype: null
353
+ torchscript: false
354
+ transformers_version: 4.17.0.dev0
355
+ use_bfloat16: false
356
+ use_weighted_layer_sum: false
357
+ vocab_size: 32
358
+ xvector_output_dim: 512
359
+ encoder_no_repeat_ngram_size:
360
+ desc: null
361
+ value: 0
362
+ eos_token_id:
363
+ desc: null
364
+ value: 50256
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: None
368
+ eval_batch_size:
369
+ desc: null
370
+ value: 16
371
+ eval_steps:
372
+ desc: null
373
+ value: 500
374
+ evaluation_strategy:
375
+ desc: null
376
+ value: steps
377
+ finetuning_task:
378
+ desc: null
379
+ value: null
380
+ forced_bos_token_id:
381
+ desc: null
382
+ value: null
383
+ forced_eos_token_id:
384
+ desc: null
385
+ value: null
386
+ fp16:
387
+ desc: null
388
+ value: true
389
+ fp16_backend:
390
+ desc: null
391
+ value: auto
392
+ fp16_full_eval:
393
+ desc: null
394
+ value: false
395
+ fp16_opt_level:
396
+ desc: null
397
+ value: O1
398
+ generation_max_length:
399
+ desc: null
400
+ value: 40
401
+ generation_num_beams:
402
+ desc: null
403
+ value: 1
404
+ gradient_accumulation_steps:
405
+ desc: null
406
+ value: 2
407
+ gradient_checkpointing:
408
+ desc: null
409
+ value: true
410
+ greater_is_better:
411
+ desc: null
412
+ value: None
413
+ group_by_length:
414
+ desc: null
415
+ value: true
416
+ half_precision_backend:
417
+ desc: null
418
+ value: amp
419
+ hub_model_id:
420
+ desc: null
421
+ value: None
422
+ hub_strategy:
423
+ desc: null
424
+ value: every_save
425
+ hub_token:
426
+ desc: null
427
+ value: <HUB_TOKEN>
428
+ id2label:
429
+ desc: null
430
+ value:
431
+ '0': LABEL_0
432
+ '1': LABEL_1
433
+ ignore_data_skip:
434
+ desc: null
435
+ value: false
436
+ is_decoder:
437
+ desc: null
438
+ value: false
439
+ is_encoder_decoder:
440
+ desc: null
441
+ value: true
442
+ label2id:
443
+ desc: null
444
+ value:
445
+ LABEL_0: 0
446
+ LABEL_1: 1
447
+ label_names:
448
+ desc: null
449
+ value: None
450
+ label_smoothing_factor:
451
+ desc: null
452
+ value: 0.0
453
+ learning_rate:
454
+ desc: null
455
+ value: 1.0e-05
456
+ length_column_name:
457
+ desc: null
458
+ value: input_length
459
+ length_penalty:
460
+ desc: null
461
+ value: 1.0
462
+ load_best_model_at_end:
463
+ desc: null
464
+ value: false
465
+ local_rank:
466
+ desc: null
467
+ value: -1
468
+ log_level:
469
+ desc: null
470
+ value: -1
471
+ log_level_replica:
472
+ desc: null
473
+ value: -1
474
+ log_on_each_node:
475
+ desc: null
476
+ value: true
477
+ logging_dir:
478
+ desc: null
479
+ value: ./runs/Mar01_01-26-07_sanchit--v100
480
+ logging_first_step:
481
+ desc: null
482
+ value: false
483
+ logging_nan_inf_filter:
484
+ desc: null
485
+ value: true
486
+ logging_steps:
487
+ desc: null
488
+ value: 1
489
+ logging_strategy:
490
+ desc: null
491
+ value: steps
492
+ lr_scheduler_type:
493
+ desc: null
494
+ value: linear
495
+ max_grad_norm:
496
+ desc: null
497
+ value: 1.0
498
+ max_length:
499
+ desc: null
500
+ value: 50
501
+ max_steps:
502
+ desc: null
503
+ value: -1
504
+ metric_for_best_model:
505
+ desc: null
506
+ value: None
507
+ min_length:
508
+ desc: null
509
+ value: 0
510
+ model_type:
511
+ desc: null
512
+ value: speech-encoder-decoder
513
+ mp_parameters:
514
+ desc: null
515
+ value: ''
516
+ no_cuda:
517
+ desc: null
518
+ value: false
519
+ no_repeat_ngram_size:
520
+ desc: null
521
+ value: 0
522
+ num_beam_groups:
523
+ desc: null
524
+ value: 1
525
+ num_beams:
526
+ desc: null
527
+ value: 1
528
+ num_return_sequences:
529
+ desc: null
530
+ value: 1
531
+ num_train_epochs:
532
+ desc: null
533
+ value: 1.0
534
+ optim:
535
+ desc: null
536
+ value: adamw_hf
537
+ output_attentions:
538
+ desc: null
539
+ value: false
540
+ output_dir:
541
+ desc: null
542
+ value: ./
543
+ output_hidden_states:
544
+ desc: null
545
+ value: false
546
+ output_scores:
547
+ desc: null
548
+ value: false
549
+ overwrite_output_dir:
550
+ desc: null
551
+ value: true
552
+ pad_token_id:
553
+ desc: null
554
+ value: 50256
555
+ past_index:
556
+ desc: null
557
+ value: -1
558
+ per_device_eval_batch_size:
559
+ desc: null
560
+ value: 16
561
+ per_device_train_batch_size:
562
+ desc: null
563
+ value: 16
564
+ per_gpu_eval_batch_size:
565
+ desc: null
566
+ value: None
567
+ per_gpu_train_batch_size:
568
+ desc: null
569
+ value: None
570
+ predict_with_generate:
571
+ desc: null
572
+ value: true
573
+ prediction_loss_only:
574
+ desc: null
575
+ value: false
576
+ prefix:
577
+ desc: null
578
+ value: null
579
+ problem_type:
580
+ desc: null
581
+ value: null
582
+ processor_class:
583
+ desc: null
584
+ value: Wav2Vec2Processor
585
+ pruned_heads:
586
+ desc: null
587
+ value: {}
588
+ push_to_hub:
589
+ desc: null
590
+ value: true
591
+ push_to_hub_model_id:
592
+ desc: null
593
+ value: None
594
+ push_to_hub_organization:
595
+ desc: null
596
+ value: None
597
+ push_to_hub_token:
598
+ desc: null
599
+ value: <PUSH_TO_HUB_TOKEN>
600
+ remove_invalid_values:
601
+ desc: null
602
+ value: false
603
+ remove_unused_columns:
604
+ desc: null
605
+ value: true
606
+ repetition_penalty:
607
+ desc: null
608
+ value: 1.0
609
+ report_to:
610
+ desc: null
611
+ value: '[''wandb'']'
612
+ resume_from_checkpoint:
613
+ desc: null
614
+ value: None
615
+ return_dict:
616
+ desc: null
617
+ value: true
618
+ return_dict_in_generate:
619
+ desc: null
620
+ value: false
621
+ run_name:
622
+ desc: null
623
+ value: ./
624
+ save_on_each_node:
625
+ desc: null
626
+ value: false
627
+ save_steps:
628
+ desc: null
629
+ value: 500
630
+ save_strategy:
631
+ desc: null
632
+ value: steps
633
+ save_total_limit:
634
+ desc: null
635
+ value: 1
636
+ seed:
637
+ desc: null
638
+ value: 42
639
+ sep_token_id:
640
+ desc: null
641
+ value: null
642
+ sharded_ddp:
643
+ desc: null
644
+ value: '[]'
645
+ skip_memory_metrics:
646
+ desc: null
647
+ value: true
648
+ sortish_sampler:
649
+ desc: null
650
+ value: false
651
+ task_specific_params:
652
+ desc: null
653
+ value: null
654
+ temperature:
655
+ desc: null
656
+ value: 1.0
657
+ tf32:
658
+ desc: null
659
+ value: None
660
+ tie_encoder_decoder:
661
+ desc: null
662
+ value: false
663
+ tie_word_embeddings:
664
+ desc: null
665
+ value: false
666
+ tokenizer_class:
667
+ desc: null
668
+ value: null
669
+ top_k:
670
+ desc: null
671
+ value: 50
672
+ top_p:
673
+ desc: null
674
+ value: 1.0
675
+ torch_dtype:
676
+ desc: null
677
+ value: torch.float32
678
+ torchscript:
679
+ desc: null
680
+ value: false
681
+ tpu_metrics_debug:
682
+ desc: null
683
+ value: false
684
+ tpu_num_cores:
685
+ desc: null
686
+ value: None
687
+ train_batch_size:
688
+ desc: null
689
+ value: 16
690
+ transformers_version:
691
+ desc: null
692
+ value: null
693
+ use_bfloat16:
694
+ desc: null
695
+ value: false
696
+ use_cache:
697
+ desc: null
698
+ value: false
699
+ use_legacy_prediction_loop:
700
+ desc: null
701
+ value: false
702
+ warmup_ratio:
703
+ desc: null
704
+ value: 0.0
705
+ warmup_steps:
706
+ desc: null
707
+ value: 500
708
+ weight_decay:
709
+ desc: null
710
+ value: 0.0
711
+ xpu_backend:
712
+ desc: null
713
+ value: None
wandb/run-20220301_012648-nzue7jk0/files/output.log ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ 0%| | 0/892 [00:00<?, ?it/s][WARNING|modeling_utils.py:388] 2022-03-01 01:26:54,420 >> Could not estimate the number of tokens of the input, floating-point operations will not be computed
4
+ Traceback (most recent call last):
5
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 539, in <module>
6
+ main()
7
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 491, in main
8
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
9
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1384, in train
10
+ tr_loss_step = self.training_step(model, inputs)
11
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1959, in training_step
12
+ loss = self.compute_loss(model, inputs)
13
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1991, in compute_loss
14
+ outputs = model(**inputs)
15
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
16
+ return forward_call(*input, **kwargs)
17
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py", line 503, in forward
18
+ encoder_outputs = self.encoder(
19
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
20
+ return forward_call(*input, **kwargs)
21
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1346, in forward
22
+ extract_features = self.feature_extractor(input_values)
23
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
24
+ return forward_call(*input, **kwargs)
25
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 514, in forward
26
+ hidden_states = conv_layer(hidden_states)
27
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
28
+ return forward_call(*input, **kwargs)
29
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 389, in forward
30
+ hidden_states = self.layer_norm(hidden_states)
31
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
32
+ return forward_call(*input, **kwargs)
33
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/normalization.py", line 189, in forward
34
+ return F.layer_norm(
35
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/functional.py", line 2347, in layer_norm
36
+ return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
37
+ RuntimeError: CUDA out of memory. Tried to allocate 1.63 GiB (GPU 0; 15.78 GiB total capacity; 10.07 GiB already allocated; 707.31 MiB free; 13.39 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
wandb/run-20220301_012648-nzue7jk0/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220301_012648-nzue7jk0/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-01T01:26:50.071846",
5
+ "startedAt": "2022-03-01T01:26:48.999321",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=16",
24
+ "--per_device_eval_batch_size=16",
25
+ "--gradient_accumulation_steps=2",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "7dec2191123af092d2edc7951267a70d7948caba"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220301_012648-nzue7jk0/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
1
+ {"_wandb": {"runtime": 8}}
wandb/run-20220301_012648-nzue7jk0/logs/debug-internal.log ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 01:26:49,894 INFO MainThread:236498 [internal.py:wandb_internal():89] W&B internal server running at pid: 236498, started at: 2022-03-01 01:26:49.893757
2
+ 2022-03-01 01:26:49,896 INFO WriterThread:236498 [datastore.py:open_for_write():77] open: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/run-nzue7jk0.wandb
3
+ 2022-03-01 01:26:49,896 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: check_version
4
+ 2022-03-01 01:26:49,897 DEBUG SenderThread:236498 [sender.py:send():235] send: header
5
+ 2022-03-01 01:26:49,898 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: check_version
6
+ 2022-03-01 01:26:49,967 DEBUG SenderThread:236498 [sender.py:send():235] send: run
7
+ 2022-03-01 01:26:50,066 INFO SenderThread:236498 [dir_watcher.py:__init__():169] watching files in: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files
8
+ 2022-03-01 01:26:50,066 INFO SenderThread:236498 [sender.py:_start_run_threads():809] run started: nzue7jk0 with start time 1646098009
9
+ 2022-03-01 01:26:50,066 DEBUG SenderThread:236498 [sender.py:send():235] send: summary
10
+ 2022-03-01 01:26:50,066 INFO SenderThread:236498 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
11
+ 2022-03-01 01:26:50,067 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: run_start
12
+ 2022-03-01 01:26:50,071 DEBUG HandlerThread:236498 [meta.py:__init__():36] meta init
13
+ 2022-03-01 01:26:50,071 DEBUG HandlerThread:236498 [meta.py:__init__():50] meta init done
14
+ 2022-03-01 01:26:50,071 DEBUG HandlerThread:236498 [meta.py:probe():210] probe
15
+ 2022-03-01 01:26:50,078 DEBUG HandlerThread:236498 [meta.py:_setup_git():200] setup git
16
+ 2022-03-01 01:26:50,092 DEBUG HandlerThread:236498 [meta.py:_setup_git():207] setup git done
17
+ 2022-03-01 01:26:50,093 DEBUG HandlerThread:236498 [meta.py:_save_pip():54] save pip
18
+ 2022-03-01 01:26:50,093 DEBUG HandlerThread:236498 [meta.py:_save_pip():68] save pip done
19
+ 2022-03-01 01:26:50,093 DEBUG HandlerThread:236498 [meta.py:probe():248] probe done
20
+ 2022-03-01 01:26:50,171 DEBUG SenderThread:236498 [sender.py:send():235] send: files
21
+ 2022-03-01 01:26:50,171 INFO SenderThread:236498 [sender.py:_save_file():944] saving file wandb-metadata.json with policy now
22
+ 2022-03-01 01:26:50,176 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: stop_status
23
+ 2022-03-01 01:26:50,176 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: stop_status
24
+ 2022-03-01 01:26:50,211 DEBUG SenderThread:236498 [sender.py:send():235] send: config
25
+ 2022-03-01 01:26:50,212 DEBUG SenderThread:236498 [sender.py:send():235] send: metric
26
+ 2022-03-01 01:26:50,212 DEBUG SenderThread:236498 [sender.py:send():235] send: metric
27
+ 2022-03-01 01:26:50,212 WARNING SenderThread:236498 [sender.py:send_metric():902] Seen metric with glob (shouldnt happen)
28
+ 2022-03-01 01:26:50,430 INFO Thread-11 :236498 [upload_job.py:push():137] Uploaded file /tmp/tmpscakcml5wandb/3qv9ow3a-wandb-metadata.json
29
+ 2022-03-01 01:26:51,068 INFO Thread-8 :236498 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/wandb-summary.json
30
+ 2022-03-01 01:26:51,068 INFO Thread-8 :236498 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/output.log
31
+ 2022-03-01 01:26:51,068 INFO Thread-8 :236498 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/wandb-metadata.json
32
+ 2022-03-01 01:26:51,068 INFO Thread-8 :236498 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/requirements.txt
33
+ 2022-03-01 01:26:53,067 INFO Thread-8 :236498 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/output.log
34
+ 2022-03-01 01:26:57,068 INFO Thread-8 :236498 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/output.log
35
+ 2022-03-01 01:26:58,513 DEBUG SenderThread:236498 [sender.py:send():235] send: telemetry
36
+ 2022-03-01 01:26:58,513 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: poll_exit
37
+ 2022-03-01 01:26:58,514 DEBUG SenderThread:236498 [sender.py:send():235] send: exit
38
+ 2022-03-01 01:26:58,514 INFO SenderThread:236498 [sender.py:send_exit():371] handling exit code: 1
39
+ 2022-03-01 01:26:58,514 INFO SenderThread:236498 [sender.py:send_exit():373] handling runtime: 8
40
+ 2022-03-01 01:26:58,515 INFO SenderThread:236498 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
41
+ 2022-03-01 01:26:58,515 INFO SenderThread:236498 [sender.py:send_exit():379] send defer
42
+ 2022-03-01 01:26:58,515 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: poll_exit
43
+ 2022-03-01 01:26:58,516 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
44
+ 2022-03-01 01:26:58,516 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 0
45
+ 2022-03-01 01:26:58,516 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
46
+ 2022-03-01 01:26:58,516 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 0
47
+ 2022-03-01 01:26:58,516 INFO SenderThread:236498 [sender.py:transition_state():392] send defer: 1
48
+ 2022-03-01 01:26:58,516 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
49
+ 2022-03-01 01:26:58,516 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 1
50
+ 2022-03-01 01:26:58,601 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
51
+ 2022-03-01 01:26:58,601 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 1
52
+ 2022-03-01 01:26:58,601 INFO SenderThread:236498 [sender.py:transition_state():392] send defer: 2
53
+ 2022-03-01 01:26:58,601 DEBUG SenderThread:236498 [sender.py:send():235] send: stats
54
+ 2022-03-01 01:26:58,602 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
55
+ 2022-03-01 01:26:58,602 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 2
56
+ 2022-03-01 01:26:58,602 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
57
+ 2022-03-01 01:26:58,602 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 2
58
+ 2022-03-01 01:26:58,602 INFO SenderThread:236498 [sender.py:transition_state():392] send defer: 3
59
+ 2022-03-01 01:26:58,603 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
60
+ 2022-03-01 01:26:58,603 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 3
61
+ 2022-03-01 01:26:58,603 DEBUG SenderThread:236498 [sender.py:send():235] send: summary
62
+ 2022-03-01 01:26:58,603 INFO SenderThread:236498 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
63
+ 2022-03-01 01:26:58,603 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
64
+ 2022-03-01 01:26:58,603 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 3
65
+ 2022-03-01 01:26:58,603 INFO SenderThread:236498 [sender.py:transition_state():392] send defer: 4
66
+ 2022-03-01 01:26:58,604 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
67
+ 2022-03-01 01:26:58,604 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 4
68
+ 2022-03-01 01:26:58,604 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
69
+ 2022-03-01 01:26:58,604 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 4
70
+ 2022-03-01 01:26:58,617 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: poll_exit
71
+ 2022-03-01 01:26:58,708 INFO SenderThread:236498 [sender.py:transition_state():392] send defer: 5
72
+ 2022-03-01 01:26:58,708 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: poll_exit
73
+ 2022-03-01 01:26:58,709 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
74
+ 2022-03-01 01:26:58,709 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 5
75
+ 2022-03-01 01:26:58,709 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
76
+ 2022-03-01 01:26:58,709 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 5
77
+ 2022-03-01 01:26:58,709 INFO SenderThread:236498 [dir_watcher.py:finish():283] shutting down directory watcher
78
+ 2022-03-01 01:26:58,809 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: poll_exit
79
+ 2022-03-01 01:26:59,069 INFO Thread-8 :236498 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/config.yaml
80
+ 2022-03-01 01:26:59,070 INFO SenderThread:236498 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/output.log
81
+ 2022-03-01 01:26:59,070 INFO SenderThread:236498 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/wandb-summary.json
82
+ 2022-03-01 01:26:59,070 INFO SenderThread:236498 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files
83
+ 2022-03-01 01:26:59,070 INFO SenderThread:236498 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/wandb-metadata.json wandb-metadata.json
84
+ 2022-03-01 01:26:59,070 INFO SenderThread:236498 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/output.log output.log
85
+ 2022-03-01 01:26:59,070 INFO SenderThread:236498 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/wandb-summary.json wandb-summary.json
86
+ 2022-03-01 01:26:59,073 INFO SenderThread:236498 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/requirements.txt requirements.txt
87
+ 2022-03-01 01:26:59,074 INFO SenderThread:236498 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/config.yaml config.yaml
88
+ 2022-03-01 01:26:59,076 INFO SenderThread:236498 [sender.py:transition_state():392] send defer: 6
89
+ 2022-03-01 01:26:59,076 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: poll_exit
90
+ 2022-03-01 01:26:59,077 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
91
+ 2022-03-01 01:26:59,077 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 6
92
+ 2022-03-01 01:26:59,078 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
93
+ 2022-03-01 01:26:59,084 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 6
94
+ 2022-03-01 01:26:59,084 INFO SenderThread:236498 [file_pusher.py:finish():177] shutting down file pusher
95
+ 2022-03-01 01:26:59,179 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: poll_exit
96
+ 2022-03-01 01:26:59,179 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: poll_exit
97
+ 2022-03-01 01:26:59,281 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: poll_exit
98
+ 2022-03-01 01:26:59,281 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: poll_exit
99
+ 2022-03-01 01:26:59,340 INFO Thread-15 :236498 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/config.yaml
100
+ 2022-03-01 01:26:59,342 INFO Thread-14 :236498 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/requirements.txt
101
+ 2022-03-01 01:26:59,369 INFO Thread-12 :236498 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/output.log
102
+ 2022-03-01 01:26:59,376 INFO Thread-13 :236498 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/files/wandb-summary.json
103
+ 2022-03-01 01:26:59,383 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: poll_exit
104
+ 2022-03-01 01:26:59,383 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: poll_exit
105
+ 2022-03-01 01:26:59,484 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: poll_exit
106
+ 2022-03-01 01:26:59,484 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: poll_exit
107
+ 2022-03-01 01:26:59,577 INFO Thread-7 :236498 [sender.py:transition_state():392] send defer: 7
108
+ 2022-03-01 01:26:59,577 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
109
+ 2022-03-01 01:26:59,577 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 7
110
+ 2022-03-01 01:26:59,578 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
111
+ 2022-03-01 01:26:59,578 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 7
112
+ 2022-03-01 01:26:59,585 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: poll_exit
113
+ 2022-03-01 01:26:59,682 INFO SenderThread:236498 [sender.py:transition_state():392] send defer: 8
114
+ 2022-03-01 01:26:59,683 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: poll_exit
115
+ 2022-03-01 01:26:59,683 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
116
+ 2022-03-01 01:26:59,683 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 8
117
+ 2022-03-01 01:26:59,683 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
118
+ 2022-03-01 01:26:59,684 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 8
119
+ 2022-03-01 01:26:59,684 INFO SenderThread:236498 [sender.py:transition_state():392] send defer: 9
120
+ 2022-03-01 01:26:59,684 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: defer
121
+ 2022-03-01 01:26:59,684 INFO HandlerThread:236498 [handler.py:handle_request_defer():154] handle defer: 9
122
+ 2022-03-01 01:26:59,684 DEBUG SenderThread:236498 [sender.py:send():235] send: final
123
+ 2022-03-01 01:26:59,685 DEBUG SenderThread:236498 [sender.py:send():235] send: footer
124
+ 2022-03-01 01:26:59,685 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: defer
125
+ 2022-03-01 01:26:59,685 INFO SenderThread:236498 [sender.py:send_request_defer():388] handle sender defer: 9
126
+ 2022-03-01 01:26:59,784 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: poll_exit
127
+ 2022-03-01 01:26:59,785 DEBUG SenderThread:236498 [sender.py:send_request():249] send_request: poll_exit
128
+ 2022-03-01 01:26:59,785 INFO SenderThread:236498 [file_pusher.py:join():182] waiting for file pusher
129
+ 2022-03-01 01:26:59,847 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: get_summary
130
+ 2022-03-01 01:26:59,847 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: sampled_history
131
+ 2022-03-01 01:26:59,848 DEBUG HandlerThread:236498 [handler.py:handle_request():131] handle_request: shutdown
132
+ 2022-03-01 01:26:59,848 INFO HandlerThread:236498 [handler.py:finish():739] shutting down handler
133
+ 2022-03-01 01:27:00,685 INFO WriterThread:236498 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/run-nzue7jk0.wandb
134
+ 2022-03-01 01:27:00,846 INFO SenderThread:236498 [sender.py:finish():1075] shutting down sender
135
+ 2022-03-01 01:27:00,846 INFO SenderThread:236498 [file_pusher.py:finish():177] shutting down file pusher
136
+ 2022-03-01 01:27:00,846 INFO SenderThread:236498 [file_pusher.py:join():182] waiting for file pusher
137
+ 2022-03-01 01:27:00,848 INFO MainThread:236498 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220301_012648-nzue7jk0/logs/debug.log ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 01:26:49,008 INFO MainThread:236399 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-01 01:26:49,008 INFO MainThread:236399 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-01 01:26:49,008 INFO MainThread:236399 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-01 01:26:49,008 INFO MainThread:236399 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-01 01:26:49,008 INFO MainThread:236399 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/logs/debug.log
6
+ 2022-03-01 01:26:49,009 INFO MainThread:236399 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012648-nzue7jk0/logs/debug-internal.log
7
+ 2022-03-01 01:26:49,009 INFO MainThread:236399 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-01 01:26:49,009 INFO MainThread:236399 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-01 01:26:49,009 INFO MainThread:236399 [wandb_init.py:init():471] starting backend
11
+ 2022-03-01 01:26:49,009 INFO MainThread:236399 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-01 01:26:49,061 INFO MainThread:236399 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-01 01:26:49,111 INFO MainThread:236399 [backend.py:ensure_launched():224] started backend process with pid: 236498
14
+ 2022-03-01 01:26:49,113 INFO MainThread:236399 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-01 01:26:49,123 INFO MainThread:236399 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-01 01:26:49,246 INFO MainThread:236399 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-01 01:26:49,966 INFO MainThread:236399 [wandb_init.py:init():586] got version response
18
+ 2022-03-01 01:26:49,966 INFO MainThread:236399 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
19
+ 2022-03-01 01:26:50,066 INFO MainThread:236399 [wandb_init.py:init():624] starting run threads in backend
20
+ 2022-03-01 01:26:50,175 INFO MainThread:236399 [wandb_run.py:_console_start():1827] atexit reg
21
+ 2022-03-01 01:26:50,176 INFO MainThread:236399 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
22
+ 2022-03-01 01:26:50,176 INFO MainThread:236399 [wandb_run.py:_redirect():1706] Redirecting console.
23
+ 2022-03-01 01:26:50,178 INFO MainThread:236399 [wandb_run.py:_redirect():1762] Redirects installed.
24
+ 2022-03-01 01:26:50,178 INFO MainThread:236399 [wandb_init.py:init():651] run started, returning control to user process
25
+ 2022-03-01 01:26:50,181 INFO MainThread:236399 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 1e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar01_01-26-07_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 16, 'eval_batch_size': 16}
26
+ 2022-03-01 01:26:50,183 INFO MainThread:236399 [wandb_watch.py:watch():43] Watching
27
+ 2022-03-01 01:26:56,356 INFO MainThread:236399 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
28
+ 2022-03-01 01:26:56,359 INFO MainThread:236399 [wandb_run.py:_restore():1769] restore
29
+ 2022-03-01 01:26:58,516 INFO MainThread:236399 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
30
+ wandb_count: 1
31
+ }
32
+ pusher_stats {
33
+ uploaded_bytes: 2095
34
+ total_bytes: 2095
35
+ }
36
+
37
+ 2022-03-01 01:26:58,709 INFO MainThread:236399 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
38
+ wandb_count: 1
39
+ }
40
+ pusher_stats {
41
+ uploaded_bytes: 2095
42
+ total_bytes: 2095
43
+ }
44
+
45
+ 2022-03-01 01:26:59,078 INFO MainThread:236399 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
46
+ wandb_count: 3
47
+ }
48
+ pusher_stats {
49
+ uploaded_bytes: 2095
50
+ total_bytes: 5481
51
+ }
52
+
53
+ 2022-03-01 01:26:59,180 INFO MainThread:236399 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
54
+ wandb_count: 5
55
+ }
56
+ pusher_stats {
57
+ uploaded_bytes: 21067
58
+ total_bytes: 21067
59
+ }
60
+
61
+ 2022-03-01 01:26:59,282 INFO MainThread:236399 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
62
+ wandb_count: 5
63
+ }
64
+ pusher_stats {
65
+ uploaded_bytes: 21067
66
+ total_bytes: 21067
67
+ }
68
+
69
+ 2022-03-01 01:26:59,383 INFO MainThread:236399 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
70
+ wandb_count: 5
71
+ }
72
+ pusher_stats {
73
+ uploaded_bytes: 21067
74
+ total_bytes: 21067
75
+ }
76
+
77
+ 2022-03-01 01:26:59,485 INFO MainThread:236399 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
78
+ wandb_count: 5
79
+ }
80
+ pusher_stats {
81
+ uploaded_bytes: 21067
82
+ total_bytes: 21067
83
+ }
84
+
85
+ 2022-03-01 01:26:59,683 INFO MainThread:236399 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
86
+ wandb_count: 5
87
+ }
88
+ pusher_stats {
89
+ uploaded_bytes: 21067
90
+ total_bytes: 21067
91
+ }
92
+
93
+ 2022-03-01 01:26:59,846 INFO MainThread:236399 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
94
+ exit_result {
95
+ }
96
+ file_counts {
97
+ wandb_count: 5
98
+ }
99
+ pusher_stats {
100
+ uploaded_bytes: 21067
101
+ total_bytes: 21067
102
+ }
103
+ local_info {
104
+ }
105
+
106
+ 2022-03-01 01:27:00,977 INFO MainThread:236399 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220301_012648-nzue7jk0/run-nzue7jk0.wandb ADDED
Binary file (13.7 kB). View file
wandb/run-20220301_012816-3bh80nzn/files/config.yaml ADDED
@@ -0,0 +1,713 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: ./
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.10
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.9.5
22
+ start_time: 1646098096
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.9.5
37
+ 5: 0.12.10
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ adafactor:
42
+ desc: null
43
+ value: false
44
+ adam_beta1:
45
+ desc: null
46
+ value: 0.9
47
+ adam_beta2:
48
+ desc: null
49
+ value: 0.999
50
+ adam_epsilon:
51
+ desc: null
52
+ value: 1.0e-08
53
+ add_cross_attention:
54
+ desc: null
55
+ value: false
56
+ architectures:
57
+ desc: null
58
+ value:
59
+ - SpeechEncoderDecoderModel
60
+ bad_words_ids:
61
+ desc: null
62
+ value: null
63
+ bf16:
64
+ desc: null
65
+ value: false
66
+ bf16_full_eval:
67
+ desc: null
68
+ value: false
69
+ bos_token_id:
70
+ desc: null
71
+ value: null
72
+ chunk_size_feed_forward:
73
+ desc: null
74
+ value: 0
75
+ cross_attention_hidden_size:
76
+ desc: null
77
+ value: null
78
+ dataloader_drop_last:
79
+ desc: null
80
+ value: false
81
+ dataloader_num_workers:
82
+ desc: null
83
+ value: 0
84
+ dataloader_pin_memory:
85
+ desc: null
86
+ value: true
87
+ ddp_bucket_cap_mb:
88
+ desc: null
89
+ value: None
90
+ ddp_find_unused_parameters:
91
+ desc: null
92
+ value: None
93
+ debug:
94
+ desc: null
95
+ value: '[]'
96
+ decoder:
97
+ desc: null
98
+ value:
99
+ _name_or_path: gpt2-medium
100
+ activation_function: gelu_new
101
+ add_cross_attention: true
102
+ architectures:
103
+ - GPT2LMHeadModel
104
+ attn_pdrop: 0.0
105
+ bad_words_ids: null
106
+ bos_token_id: 50256
107
+ chunk_size_feed_forward: 0
108
+ cross_attention_hidden_size: null
109
+ decoder_start_token_id: null
110
+ diversity_penalty: 0.0
111
+ do_sample: false
112
+ early_stopping: false
113
+ embd_pdrop: 0.0
114
+ encoder_no_repeat_ngram_size: 0
115
+ eos_token_id: 50256
116
+ finetuning_task: null
117
+ forced_bos_token_id: null
118
+ forced_eos_token_id: null
119
+ id2label:
120
+ '0': LABEL_0
121
+ '1': LABEL_1
122
+ initializer_range: 0.02
123
+ is_decoder: true
124
+ is_encoder_decoder: false
125
+ label2id:
126
+ LABEL_0: 0
127
+ LABEL_1: 1
128
+ layer_norm_epsilon: 1.0e-05
129
+ length_penalty: 1.0
130
+ max_length: 20
131
+ min_length: 0
132
+ model_type: gpt2
133
+ n_ctx: 1024
134
+ n_embd: 1024
135
+ n_head: 16
136
+ n_inner: null
137
+ n_layer: 24
138
+ n_positions: 1024
139
+ n_special: 0
140
+ no_repeat_ngram_size: 0
141
+ num_beam_groups: 1
142
+ num_beams: 1
143
+ num_return_sequences: 1
144
+ output_attentions: false
145
+ output_hidden_states: false
146
+ output_scores: false
147
+ pad_token_id: null
148
+ predict_special_tokens: true
149
+ prefix: null
150
+ problem_type: null
151
+ pruned_heads: {}
152
+ remove_invalid_values: false
153
+ reorder_and_upcast_attn: false
154
+ repetition_penalty: 1.0
155
+ resid_pdrop: 0.0
156
+ return_dict: true
157
+ return_dict_in_generate: false
158
+ scale_attn_by_inverse_layer_idx: false
159
+ scale_attn_weights: true
160
+ sep_token_id: null
161
+ summary_activation: null
162
+ summary_first_dropout: 0.0
163
+ summary_proj_to_labels: true
164
+ summary_type: cls_index
165
+ summary_use_proj: true
166
+ task_specific_params:
167
+ text-generation:
168
+ do_sample: true
169
+ max_length: 50
170
+ temperature: 1.0
171
+ tie_encoder_decoder: false
172
+ tie_word_embeddings: true
173
+ tokenizer_class: null
174
+ top_k: 50
175
+ top_p: 1.0
176
+ torch_dtype: null
177
+ torchscript: false
178
+ transformers_version: 4.17.0.dev0
179
+ use_bfloat16: false
180
+ use_cache: false
181
+ vocab_size: 50257
182
+ decoder_start_token_id:
183
+ desc: null
184
+ value: 50256
185
+ deepspeed:
186
+ desc: null
187
+ value: None
188
+ disable_tqdm:
189
+ desc: null
190
+ value: false
191
+ diversity_penalty:
192
+ desc: null
193
+ value: 0.0
194
+ do_eval:
195
+ desc: null
196
+ value: true
197
+ do_predict:
198
+ desc: null
199
+ value: false
200
+ do_sample:
201
+ desc: null
202
+ value: false
203
+ do_train:
204
+ desc: null
205
+ value: true
206
+ early_stopping:
207
+ desc: null
208
+ value: false
209
+ encoder:
210
+ desc: null
211
+ value:
212
+ _name_or_path: facebook/wav2vec2-large-lv60
213
+ activation_dropout: 0.0
214
+ adapter_kernel_size: 3
215
+ adapter_stride: 2
216
+ add_adapter: true
217
+ add_cross_attention: false
218
+ apply_spec_augment: false
219
+ architectures:
220
+ - Wav2Vec2ForPreTraining
221
+ attention_dropout: 0.0
222
+ bad_words_ids: null
223
+ bos_token_id: 1
224
+ chunk_size_feed_forward: 0
225
+ classifier_proj_size: 256
226
+ codevector_dim: 768
227
+ contrastive_logits_temperature: 0.1
228
+ conv_bias: true
229
+ conv_dim:
230
+ - 512
231
+ - 512
232
+ - 512
233
+ - 512
234
+ - 512
235
+ - 512
236
+ - 512
237
+ conv_kernel:
238
+ - 10
239
+ - 3
240
+ - 3
241
+ - 3
242
+ - 3
243
+ - 2
244
+ - 2
245
+ conv_stride:
246
+ - 5
247
+ - 2
248
+ - 2
249
+ - 2
250
+ - 2
251
+ - 2
252
+ - 2
253
+ cross_attention_hidden_size: null
254
+ ctc_loss_reduction: sum
255
+ ctc_zero_infinity: false
256
+ decoder_start_token_id: null
257
+ diversity_loss_weight: 0.1
258
+ diversity_penalty: 0.0
259
+ do_sample: false
260
+ do_stable_layer_norm: true
261
+ early_stopping: false
262
+ encoder_no_repeat_ngram_size: 0
263
+ eos_token_id: 2
264
+ feat_extract_activation: gelu
265
+ feat_extract_dropout: 0.0
266
+ feat_extract_norm: layer
267
+ feat_proj_dropout: 0.0
268
+ feat_quantizer_dropout: 0.0
269
+ final_dropout: 0.0
270
+ finetuning_task: null
271
+ forced_bos_token_id: null
272
+ forced_eos_token_id: null
273
+ gradient_checkpointing: false
274
+ hidden_act: gelu
275
+ hidden_dropout: 0.0
276
+ hidden_dropout_prob: 0.0
277
+ hidden_size: 1024
278
+ id2label:
279
+ '0': LABEL_0
280
+ '1': LABEL_1
281
+ initializer_range: 0.02
282
+ intermediate_size: 4096
283
+ is_decoder: false
284
+ is_encoder_decoder: false
285
+ label2id:
286
+ LABEL_0: 0
287
+ LABEL_1: 1
288
+ layer_norm_eps: 1.0e-05
289
+ layerdrop: 0.0
290
+ length_penalty: 1.0
291
+ mask_feature_length: 10
292
+ mask_feature_min_masks: 0
293
+ mask_feature_prob: 0.0
294
+ mask_time_length: 10
295
+ mask_time_min_masks: 2
296
+ mask_time_prob: 0.0
297
+ max_length: 20
298
+ min_length: 0
299
+ model_type: wav2vec2
300
+ no_repeat_ngram_size: 0
301
+ num_adapter_layers: 3
302
+ num_attention_heads: 16
303
+ num_beam_groups: 1
304
+ num_beams: 1
305
+ num_codevector_groups: 2
306
+ num_codevectors_per_group: 320
307
+ num_conv_pos_embedding_groups: 16
308
+ num_conv_pos_embeddings: 128
309
+ num_feat_extract_layers: 7
310
+ num_hidden_layers: 24
311
+ num_negatives: 100
312
+ num_return_sequences: 1
313
+ output_attentions: false
314
+ output_hidden_size: 1024
315
+ output_hidden_states: false
316
+ output_scores: false
317
+ pad_token_id: 0
318
+ prefix: null
319
+ problem_type: null
320
+ proj_codevector_dim: 768
321
+ pruned_heads: {}
322
+ remove_invalid_values: false
323
+ repetition_penalty: 1.0
324
+ return_dict: true
325
+ return_dict_in_generate: false
326
+ sep_token_id: null
327
+ task_specific_params: null
328
+ tdnn_dilation:
329
+ - 1
330
+ - 2
331
+ - 3
332
+ - 1
333
+ - 1
334
+ tdnn_dim:
335
+ - 512
336
+ - 512
337
+ - 512
338
+ - 512
339
+ - 1500
340
+ tdnn_kernel:
341
+ - 5
342
+ - 3
343
+ - 3
344
+ - 1
345
+ - 1
346
+ temperature: 1.0
347
+ tie_encoder_decoder: false
348
+ tie_word_embeddings: true
349
+ tokenizer_class: null
350
+ top_k: 50
351
+ top_p: 1.0
352
+ torch_dtype: null
353
+ torchscript: false
354
+ transformers_version: 4.17.0.dev0
355
+ use_bfloat16: false
356
+ use_weighted_layer_sum: false
357
+ vocab_size: 32
358
+ xvector_output_dim: 512
359
+ encoder_no_repeat_ngram_size:
360
+ desc: null
361
+ value: 0
362
+ eos_token_id:
363
+ desc: null
364
+ value: 50256
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: None
368
+ eval_batch_size:
369
+ desc: null
370
+ value: 16
371
+ eval_steps:
372
+ desc: null
373
+ value: 500
374
+ evaluation_strategy:
375
+ desc: null
376
+ value: steps
377
+ finetuning_task:
378
+ desc: null
379
+ value: null
380
+ forced_bos_token_id:
381
+ desc: null
382
+ value: null
383
+ forced_eos_token_id:
384
+ desc: null
385
+ value: null
386
+ fp16:
387
+ desc: null
388
+ value: true
389
+ fp16_backend:
390
+ desc: null
391
+ value: auto
392
+ fp16_full_eval:
393
+ desc: null
394
+ value: false
395
+ fp16_opt_level:
396
+ desc: null
397
+ value: O1
398
+ generation_max_length:
399
+ desc: null
400
+ value: 40
401
+ generation_num_beams:
402
+ desc: null
403
+ value: 1
404
+ gradient_accumulation_steps:
405
+ desc: null
406
+ value: 4
407
+ gradient_checkpointing:
408
+ desc: null
409
+ value: true
410
+ greater_is_better:
411
+ desc: null
412
+ value: None
413
+ group_by_length:
414
+ desc: null
415
+ value: true
416
+ half_precision_backend:
417
+ desc: null
418
+ value: amp
419
+ hub_model_id:
420
+ desc: null
421
+ value: None
422
+ hub_strategy:
423
+ desc: null
424
+ value: every_save
425
+ hub_token:
426
+ desc: null
427
+ value: <HUB_TOKEN>
428
+ id2label:
429
+ desc: null
430
+ value:
431
+ '0': LABEL_0
432
+ '1': LABEL_1
433
+ ignore_data_skip:
434
+ desc: null
435
+ value: false
436
+ is_decoder:
437
+ desc: null
438
+ value: false
439
+ is_encoder_decoder:
440
+ desc: null
441
+ value: true
442
+ label2id:
443
+ desc: null
444
+ value:
445
+ LABEL_0: 0
446
+ LABEL_1: 1
447
+ label_names:
448
+ desc: null
449
+ value: None
450
+ label_smoothing_factor:
451
+ desc: null
452
+ value: 0.0
453
+ learning_rate:
454
+ desc: null
455
+ value: 1.0e-05
456
+ length_column_name:
457
+ desc: null
458
+ value: input_length
459
+ length_penalty:
460
+ desc: null
461
+ value: 1.0
462
+ load_best_model_at_end:
463
+ desc: null
464
+ value: false
465
+ local_rank:
466
+ desc: null
467
+ value: -1
468
+ log_level:
469
+ desc: null
470
+ value: -1
471
+ log_level_replica:
472
+ desc: null
473
+ value: -1
474
+ log_on_each_node:
475
+ desc: null
476
+ value: true
477
+ logging_dir:
478
+ desc: null
479
+ value: ./runs/Mar01_01-27-35_sanchit--v100
480
+ logging_first_step:
481
+ desc: null
482
+ value: false
483
+ logging_nan_inf_filter:
484
+ desc: null
485
+ value: true
486
+ logging_steps:
487
+ desc: null
488
+ value: 1
489
+ logging_strategy:
490
+ desc: null
491
+ value: steps
492
+ lr_scheduler_type:
493
+ desc: null
494
+ value: linear
495
+ max_grad_norm:
496
+ desc: null
497
+ value: 1.0
498
+ max_length:
499
+ desc: null
500
+ value: 50
501
+ max_steps:
502
+ desc: null
503
+ value: -1
504
+ metric_for_best_model:
505
+ desc: null
506
+ value: None
507
+ min_length:
508
+ desc: null
509
+ value: 0
510
+ model_type:
511
+ desc: null
512
+ value: speech-encoder-decoder
513
+ mp_parameters:
514
+ desc: null
515
+ value: ''
516
+ no_cuda:
517
+ desc: null
518
+ value: false
519
+ no_repeat_ngram_size:
520
+ desc: null
521
+ value: 0
522
+ num_beam_groups:
523
+ desc: null
524
+ value: 1
525
+ num_beams:
526
+ desc: null
527
+ value: 1
528
+ num_return_sequences:
529
+ desc: null
530
+ value: 1
531
+ num_train_epochs:
532
+ desc: null
533
+ value: 1.0
534
+ optim:
535
+ desc: null
536
+ value: adamw_hf
537
+ output_attentions:
538
+ desc: null
539
+ value: false
540
+ output_dir:
541
+ desc: null
542
+ value: ./
543
+ output_hidden_states:
544
+ desc: null
545
+ value: false
546
+ output_scores:
547
+ desc: null
548
+ value: false
549
+ overwrite_output_dir:
550
+ desc: null
551
+ value: true
552
+ pad_token_id:
553
+ desc: null
554
+ value: 50256
555
+ past_index:
556
+ desc: null
557
+ value: -1
558
+ per_device_eval_batch_size:
559
+ desc: null
560
+ value: 16
561
+ per_device_train_batch_size:
562
+ desc: null
563
+ value: 16
564
+ per_gpu_eval_batch_size:
565
+ desc: null
566
+ value: None
567
+ per_gpu_train_batch_size:
568
+ desc: null
569
+ value: None
570
+ predict_with_generate:
571
+ desc: null
572
+ value: true
573
+ prediction_loss_only:
574
+ desc: null
575
+ value: false
576
+ prefix:
577
+ desc: null
578
+ value: null
579
+ problem_type:
580
+ desc: null
581
+ value: null
582
+ processor_class:
583
+ desc: null
584
+ value: Wav2Vec2Processor
585
+ pruned_heads:
586
+ desc: null
587
+ value: {}
588
+ push_to_hub:
589
+ desc: null
590
+ value: true
591
+ push_to_hub_model_id:
592
+ desc: null
593
+ value: None
594
+ push_to_hub_organization:
595
+ desc: null
596
+ value: None
597
+ push_to_hub_token:
598
+ desc: null
599
+ value: <PUSH_TO_HUB_TOKEN>
600
+ remove_invalid_values:
601
+ desc: null
602
+ value: false
603
+ remove_unused_columns:
604
+ desc: null
605
+ value: true
606
+ repetition_penalty:
607
+ desc: null
608
+ value: 1.0
609
+ report_to:
610
+ desc: null
611
+ value: '[''wandb'']'
612
+ resume_from_checkpoint:
613
+ desc: null
614
+ value: None
615
+ return_dict:
616
+ desc: null
617
+ value: true
618
+ return_dict_in_generate:
619
+ desc: null
620
+ value: false
621
+ run_name:
622
+ desc: null
623
+ value: ./
624
+ save_on_each_node:
625
+ desc: null
626
+ value: false
627
+ save_steps:
628
+ desc: null
629
+ value: 500
630
+ save_strategy:
631
+ desc: null
632
+ value: steps
633
+ save_total_limit:
634
+ desc: null
635
+ value: 1
636
+ seed:
637
+ desc: null
638
+ value: 42
639
+ sep_token_id:
640
+ desc: null
641
+ value: null
642
+ sharded_ddp:
643
+ desc: null
644
+ value: '[]'
645
+ skip_memory_metrics:
646
+ desc: null
647
+ value: true
648
+ sortish_sampler:
649
+ desc: null
650
+ value: false
651
+ task_specific_params:
652
+ desc: null
653
+ value: null
654
+ temperature:
655
+ desc: null
656
+ value: 1.0
657
+ tf32:
658
+ desc: null
659
+ value: None
660
+ tie_encoder_decoder:
661
+ desc: null
662
+ value: false
663
+ tie_word_embeddings:
664
+ desc: null
665
+ value: false
666
+ tokenizer_class:
667
+ desc: null
668
+ value: null
669
+ top_k:
670
+ desc: null
671
+ value: 50
672
+ top_p:
673
+ desc: null
674
+ value: 1.0
675
+ torch_dtype:
676
+ desc: null
677
+ value: torch.float32
678
+ torchscript:
679
+ desc: null
680
+ value: false
681
+ tpu_metrics_debug:
682
+ desc: null
683
+ value: false
684
+ tpu_num_cores:
685
+ desc: null
686
+ value: None
687
+ train_batch_size:
688
+ desc: null
689
+ value: 16
690
+ transformers_version:
691
+ desc: null
692
+ value: null
693
+ use_bfloat16:
694
+ desc: null
695
+ value: false
696
+ use_cache:
697
+ desc: null
698
+ value: false
699
+ use_legacy_prediction_loop:
700
+ desc: null
701
+ value: false
702
+ warmup_ratio:
703
+ desc: null
704
+ value: 0.0
705
+ warmup_steps:
706
+ desc: null
707
+ value: 500
708
+ weight_decay:
709
+ desc: null
710
+ value: 0.0
711
+ xpu_backend:
712
+ desc: null
713
+ value: None
wandb/run-20220301_012816-3bh80nzn/files/output.log ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ 0%| | 0/446 [00:00<?, ?it/s][WARNING|modeling_utils.py:388] 2022-03-01 01:28:21,919 >> Could not estimate the number of tokens of the input, floating-point operations will not be computed
4
+ Traceback (most recent call last):
5
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 539, in <module>
6
+ main()
7
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 491, in main
8
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
9
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1384, in train
10
+ tr_loss_step = self.training_step(model, inputs)
11
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1959, in training_step
12
+ loss = self.compute_loss(model, inputs)
13
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1991, in compute_loss
14
+ outputs = model(**inputs)
15
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
16
+ return forward_call(*input, **kwargs)
17
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py", line 503, in forward
18
+ encoder_outputs = self.encoder(
19
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
20
+ return forward_call(*input, **kwargs)
21
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1346, in forward
22
+ extract_features = self.feature_extractor(input_values)
23
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
24
+ return forward_call(*input, **kwargs)
25
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 514, in forward
26
+ hidden_states = conv_layer(hidden_states)
27
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
28
+ return forward_call(*input, **kwargs)
29
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 389, in forward
30
+ hidden_states = self.layer_norm(hidden_states)
31
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
32
+ return forward_call(*input, **kwargs)
33
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/normalization.py", line 189, in forward
34
+ return F.layer_norm(
35
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/functional.py", line 2347, in layer_norm
36
+ return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
37
+ RuntimeError: CUDA out of memory. Tried to allocate 1.64 GiB (GPU 0; 15.78 GiB total capacity; 10.11 GiB already allocated; 707.31 MiB free; 13.39 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
wandb/run-20220301_012816-3bh80nzn/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220301_012816-3bh80nzn/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-01T01:28:17.642031",
5
+ "startedAt": "2022-03-01T01:28:16.575959",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=16",
24
+ "--per_device_eval_batch_size=16",
25
+ "--gradient_accumulation_steps=4",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=1e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "7dec2191123af092d2edc7951267a70d7948caba"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220301_012816-3bh80nzn/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
1
+ {"_wandb": {"runtime": 8}}
wandb/run-20220301_012816-3bh80nzn/logs/debug-internal.log ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 01:28:17,477 INFO MainThread:236675 [internal.py:wandb_internal():89] W&B internal server running at pid: 236675, started at: 2022-03-01 01:28:17.477147
2
+ 2022-03-01 01:28:17,479 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: check_version
3
+ 2022-03-01 01:28:17,479 INFO WriterThread:236675 [datastore.py:open_for_write():77] open: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/run-3bh80nzn.wandb
4
+ 2022-03-01 01:28:17,481 DEBUG SenderThread:236675 [sender.py:send():235] send: header
5
+ 2022-03-01 01:28:17,481 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: check_version
6
+ 2022-03-01 01:28:17,547 DEBUG SenderThread:236675 [sender.py:send():235] send: run
7
+ 2022-03-01 01:28:17,636 INFO SenderThread:236675 [dir_watcher.py:__init__():169] watching files in: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files
8
+ 2022-03-01 01:28:17,636 INFO SenderThread:236675 [sender.py:_start_run_threads():809] run started: 3bh80nzn with start time 1646098096
9
+ 2022-03-01 01:28:17,636 DEBUG SenderThread:236675 [sender.py:send():235] send: summary
10
+ 2022-03-01 01:28:17,636 INFO SenderThread:236675 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
11
+ 2022-03-01 01:28:17,637 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: run_start
12
+ 2022-03-01 01:28:17,641 DEBUG HandlerThread:236675 [meta.py:__init__():36] meta init
13
+ 2022-03-01 01:28:17,641 DEBUG HandlerThread:236675 [meta.py:__init__():50] meta init done
14
+ 2022-03-01 01:28:17,642 DEBUG HandlerThread:236675 [meta.py:probe():210] probe
15
+ 2022-03-01 01:28:17,648 DEBUG HandlerThread:236675 [meta.py:_setup_git():200] setup git
16
+ 2022-03-01 01:28:17,663 DEBUG HandlerThread:236675 [meta.py:_setup_git():207] setup git done
17
+ 2022-03-01 01:28:17,663 DEBUG HandlerThread:236675 [meta.py:_save_pip():54] save pip
18
+ 2022-03-01 01:28:17,663 DEBUG HandlerThread:236675 [meta.py:_save_pip():68] save pip done
19
+ 2022-03-01 01:28:17,664 DEBUG HandlerThread:236675 [meta.py:probe():248] probe done
20
+ 2022-03-01 01:28:17,744 DEBUG SenderThread:236675 [sender.py:send():235] send: files
21
+ 2022-03-01 01:28:17,744 INFO SenderThread:236675 [sender.py:_save_file():944] saving file wandb-metadata.json with policy now
22
+ 2022-03-01 01:28:17,749 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: stop_status
23
+ 2022-03-01 01:28:17,749 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: stop_status
24
+ 2022-03-01 01:28:17,783 DEBUG SenderThread:236675 [sender.py:send():235] send: config
25
+ 2022-03-01 01:28:17,784 DEBUG SenderThread:236675 [sender.py:send():235] send: metric
26
+ 2022-03-01 01:28:17,784 DEBUG SenderThread:236675 [sender.py:send():235] send: metric
27
+ 2022-03-01 01:28:17,784 WARNING SenderThread:236675 [sender.py:send_metric():902] Seen metric with glob (shouldnt happen)
28
+ 2022-03-01 01:28:18,011 INFO Thread-11 :236675 [upload_job.py:push():137] Uploaded file /tmp/tmpb_wel_s6wandb/1ng2jof2-wandb-metadata.json
29
+ 2022-03-01 01:28:18,638 INFO Thread-8 :236675 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/wandb-metadata.json
30
+ 2022-03-01 01:28:18,638 INFO Thread-8 :236675 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/requirements.txt
31
+ 2022-03-01 01:28:18,638 INFO Thread-8 :236675 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/wandb-summary.json
32
+ 2022-03-01 01:28:18,638 INFO Thread-8 :236675 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/output.log
33
+ 2022-03-01 01:28:20,637 INFO Thread-8 :236675 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/output.log
34
+ 2022-03-01 01:28:24,639 INFO Thread-8 :236675 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/output.log
35
+ 2022-03-01 01:28:25,989 DEBUG SenderThread:236675 [sender.py:send():235] send: telemetry
36
+ 2022-03-01 01:28:25,989 DEBUG SenderThread:236675 [sender.py:send():235] send: exit
37
+ 2022-03-01 01:28:25,989 INFO SenderThread:236675 [sender.py:send_exit():371] handling exit code: 1
38
+ 2022-03-01 01:28:25,989 INFO SenderThread:236675 [sender.py:send_exit():373] handling runtime: 8
39
+ 2022-03-01 01:28:25,990 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
40
+ 2022-03-01 01:28:25,990 INFO SenderThread:236675 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
41
+ 2022-03-01 01:28:25,991 INFO SenderThread:236675 [sender.py:send_exit():379] send defer
42
+ 2022-03-01 01:28:25,991 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
43
+ 2022-03-01 01:28:25,991 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
44
+ 2022-03-01 01:28:25,991 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 0
45
+ 2022-03-01 01:28:25,992 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
46
+ 2022-03-01 01:28:25,992 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 0
47
+ 2022-03-01 01:28:25,992 INFO SenderThread:236675 [sender.py:transition_state():392] send defer: 1
48
+ 2022-03-01 01:28:25,992 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
49
+ 2022-03-01 01:28:25,992 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 1
50
+ 2022-03-01 01:28:26,073 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
51
+ 2022-03-01 01:28:26,073 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 1
52
+ 2022-03-01 01:28:26,073 INFO SenderThread:236675 [sender.py:transition_state():392] send defer: 2
53
+ 2022-03-01 01:28:26,074 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
54
+ 2022-03-01 01:28:26,074 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 2
55
+ 2022-03-01 01:28:26,074 DEBUG SenderThread:236675 [sender.py:send():235] send: stats
56
+ 2022-03-01 01:28:26,075 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
57
+ 2022-03-01 01:28:26,075 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 2
58
+ 2022-03-01 01:28:26,075 INFO SenderThread:236675 [sender.py:transition_state():392] send defer: 3
59
+ 2022-03-01 01:28:26,075 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
60
+ 2022-03-01 01:28:26,075 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 3
61
+ 2022-03-01 01:28:26,075 DEBUG SenderThread:236675 [sender.py:send():235] send: summary
62
+ 2022-03-01 01:28:26,076 INFO SenderThread:236675 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
63
+ 2022-03-01 01:28:26,076 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
64
+ 2022-03-01 01:28:26,076 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 3
65
+ 2022-03-01 01:28:26,076 INFO SenderThread:236675 [sender.py:transition_state():392] send defer: 4
66
+ 2022-03-01 01:28:26,076 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
67
+ 2022-03-01 01:28:26,076 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 4
68
+ 2022-03-01 01:28:26,077 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
69
+ 2022-03-01 01:28:26,077 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 4
70
+ 2022-03-01 01:28:26,093 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
71
+ 2022-03-01 01:28:26,179 INFO SenderThread:236675 [sender.py:transition_state():392] send defer: 5
72
+ 2022-03-01 01:28:26,179 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
73
+ 2022-03-01 01:28:26,180 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
74
+ 2022-03-01 01:28:26,180 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 5
75
+ 2022-03-01 01:28:26,180 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
76
+ 2022-03-01 01:28:26,180 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 5
77
+ 2022-03-01 01:28:26,180 INFO SenderThread:236675 [dir_watcher.py:finish():283] shutting down directory watcher
78
+ 2022-03-01 01:28:26,281 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
79
+ 2022-03-01 01:28:26,639 INFO Thread-8 :236675 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/wandb-summary.json
80
+ 2022-03-01 01:28:26,640 INFO SenderThread:236675 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/output.log
81
+ 2022-03-01 01:28:26,640 INFO SenderThread:236675 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/config.yaml
82
+ 2022-03-01 01:28:26,640 INFO SenderThread:236675 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files
83
+ 2022-03-01 01:28:26,641 INFO SenderThread:236675 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/wandb-metadata.json wandb-metadata.json
84
+ 2022-03-01 01:28:26,641 INFO SenderThread:236675 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/output.log output.log
85
+ 2022-03-01 01:28:26,641 INFO SenderThread:236675 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/wandb-summary.json wandb-summary.json
86
+ 2022-03-01 01:28:26,644 INFO SenderThread:236675 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/requirements.txt requirements.txt
87
+ 2022-03-01 01:28:26,644 INFO SenderThread:236675 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/config.yaml config.yaml
88
+ 2022-03-01 01:28:26,644 INFO SenderThread:236675 [sender.py:transition_state():392] send defer: 6
89
+ 2022-03-01 01:28:26,645 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
90
+ 2022-03-01 01:28:26,648 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
91
+ 2022-03-01 01:28:26,651 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 6
92
+ 2022-03-01 01:28:26,653 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
93
+ 2022-03-01 01:28:26,653 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 6
94
+ 2022-03-01 01:28:26,654 INFO SenderThread:236675 [file_pusher.py:finish():177] shutting down file pusher
95
+ 2022-03-01 01:28:26,752 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
96
+ 2022-03-01 01:28:26,752 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
97
+ 2022-03-01 01:28:26,854 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
98
+ 2022-03-01 01:28:26,854 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
99
+ 2022-03-01 01:28:26,905 INFO Thread-14 :236675 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/requirements.txt
100
+ 2022-03-01 01:28:26,912 INFO Thread-13 :236675 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/wandb-summary.json
101
+ 2022-03-01 01:28:26,932 INFO Thread-12 :236675 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/output.log
102
+ 2022-03-01 01:28:26,955 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
103
+ 2022-03-01 01:28:26,955 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
104
+ 2022-03-01 01:28:26,961 INFO Thread-15 :236675 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/files/config.yaml
105
+ 2022-03-01 01:28:27,057 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
106
+ 2022-03-01 01:28:27,057 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
107
+ 2022-03-01 01:28:27,158 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
108
+ 2022-03-01 01:28:27,158 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
109
+ 2022-03-01 01:28:27,162 INFO Thread-7 :236675 [sender.py:transition_state():392] send defer: 7
110
+ 2022-03-01 01:28:27,162 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
111
+ 2022-03-01 01:28:27,162 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 7
112
+ 2022-03-01 01:28:27,162 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
113
+ 2022-03-01 01:28:27,162 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 7
114
+ 2022-03-01 01:28:27,259 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
115
+ 2022-03-01 01:28:28,140 INFO SenderThread:236675 [sender.py:transition_state():392] send defer: 8
116
+ 2022-03-01 01:28:28,140 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
117
+ 2022-03-01 01:28:28,141 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
118
+ 2022-03-01 01:28:28,141 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 8
119
+ 2022-03-01 01:28:28,141 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
120
+ 2022-03-01 01:28:28,141 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 8
121
+ 2022-03-01 01:28:28,141 INFO SenderThread:236675 [sender.py:transition_state():392] send defer: 9
122
+ 2022-03-01 01:28:28,142 DEBUG SenderThread:236675 [sender.py:send():235] send: final
123
+ 2022-03-01 01:28:28,142 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: defer
124
+ 2022-03-01 01:28:28,142 INFO HandlerThread:236675 [handler.py:handle_request_defer():154] handle defer: 9
125
+ 2022-03-01 01:28:28,142 DEBUG SenderThread:236675 [sender.py:send():235] send: footer
126
+ 2022-03-01 01:28:28,142 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: defer
127
+ 2022-03-01 01:28:28,142 INFO SenderThread:236675 [sender.py:send_request_defer():388] handle sender defer: 9
128
+ 2022-03-01 01:28:28,241 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: poll_exit
129
+ 2022-03-01 01:28:28,242 DEBUG SenderThread:236675 [sender.py:send_request():249] send_request: poll_exit
130
+ 2022-03-01 01:28:28,242 INFO SenderThread:236675 [file_pusher.py:join():182] waiting for file pusher
131
+ 2022-03-01 01:28:28,300 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: get_summary
132
+ 2022-03-01 01:28:28,301 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: sampled_history
133
+ 2022-03-01 01:28:28,302 DEBUG HandlerThread:236675 [handler.py:handle_request():131] handle_request: shutdown
134
+ 2022-03-01 01:28:28,302 INFO HandlerThread:236675 [handler.py:finish():739] shutting down handler
135
+ 2022-03-01 01:28:29,142 INFO WriterThread:236675 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_012816-3bh80nzn/run-3bh80nzn.wandb
136
+ 2022-03-01 01:28:29,299 INFO SenderThread:236675 [sender.py:finish():1075] shutting down sender
137
+ 2022-03-01 01:28:29,300 INFO SenderThread:236675 [file_pusher.py:finish():177] shutting down file pusher
138
+ 2022-03-01 01:28:29,300 INFO SenderThread:236675 [file_pusher.py:join():182] waiting for file pusher
139
+ 2022-03-01 01:28:29,302 INFO MainThread:236675 [internal.py:handle_exit():79] Internal process exited