sanchit-gandhi HF staff commited on
Commit
bcb2332
1 Parent(s): 8c71811

Training in progress, step 500

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. pytorch_model.bin +1 -1
  3. training_args.bin +1 -1
  4. wandb/debug-internal.log +1 -1
  5. wandb/debug.log +1 -1
  6. wandb/latest-run +1 -1
  7. wandb/run-20220302_200036-31e4k99c/files/config.yaml +7 -0
  8. wandb/run-20220302_200036-31e4k99c/files/output.log +7 -0
  9. wandb/run-20220302_200036-31e4k99c/files/wandb-summary.json +0 -0
  10. wandb/run-20220302_200036-31e4k99c/logs/debug-internal.log +112 -0
  11. wandb/run-20220302_200036-31e4k99c/logs/debug.log +98 -0
  12. wandb/run-20220302_200036-31e4k99c/run-31e4k99c.wandb +2 -2
  13. wandb/run-20220302_214437-2u4nhnsf/files/config.yaml +0 -0
  14. wandb/run-20220302_214437-2u4nhnsf/files/output.log +0 -0
  15. wandb/run-20220302_214437-2u4nhnsf/files/requirements.txt +184 -0
  16. wandb/run-20220302_214437-2u4nhnsf/files/wandb-metadata.json +60 -0
  17. wandb/run-20220302_214437-2u4nhnsf/files/wandb-summary.json +0 -0
  18. wandb/run-20220302_214437-2u4nhnsf/logs/debug-internal.log +0 -0
  19. wandb/run-20220302_214437-2u4nhnsf/logs/debug.log +149 -0
  20. wandb/run-20220302_214437-2u4nhnsf/run-2u4nhnsf.wandb +3 -0
  21. wandb/run-20220302_222605-10glutwr/files/config.yaml +0 -0
  22. wandb/run-20220302_222605-10glutwr/files/output.log +0 -0
  23. wandb/run-20220302_222605-10glutwr/files/requirements.txt +184 -0
  24. wandb/run-20220302_222605-10glutwr/files/wandb-metadata.json +60 -0
  25. wandb/run-20220302_222605-10glutwr/files/wandb-summary.json +0 -0
  26. wandb/run-20220302_222605-10glutwr/logs/debug-internal.log +0 -0
  27. wandb/run-20220302_222605-10glutwr/logs/debug.log +125 -0
  28. wandb/run-20220302_222605-10glutwr/run-10glutwr.wandb +3 -0
  29. wandb/run-20220302_233655-33dtvgaa/files/config.yaml +0 -0
  30. wandb/run-20220302_233655-33dtvgaa/files/output.log +0 -0
  31. wandb/run-20220302_233655-33dtvgaa/files/requirements.txt +184 -0
  32. wandb/run-20220302_233655-33dtvgaa/files/wandb-metadata.json +60 -0
  33. wandb/run-20220302_233655-33dtvgaa/files/wandb-summary.json +0 -0
  34. wandb/run-20220302_233655-33dtvgaa/logs/debug-internal.log +0 -0
  35. wandb/run-20220302_233655-33dtvgaa/logs/debug.log +133 -0
  36. wandb/run-20220302_233655-33dtvgaa/run-33dtvgaa.wandb +3 -0
  37. wandb/run-20220303_004039-1a18zsso/files/config.yaml +713 -0
  38. wandb/run-20220303_004039-1a18zsso/files/output.log +37 -0
  39. wandb/run-20220303_004039-1a18zsso/files/requirements.txt +184 -0
  40. wandb/run-20220303_004039-1a18zsso/files/wandb-metadata.json +60 -0
  41. wandb/run-20220303_004039-1a18zsso/files/wandb-summary.json +1 -0
  42. wandb/run-20220303_004039-1a18zsso/logs/debug-internal.log +142 -0
  43. wandb/run-20220303_004039-1a18zsso/logs/debug.log +123 -0
  44. wandb/run-20220303_004039-1a18zsso/run-1a18zsso.wandb +0 -0
  45. wandb/run-20220303_004215-2ttcbe3s/files/config.yaml +713 -0
  46. wandb/run-20220303_004215-2ttcbe3s/files/output.log +37 -0
  47. wandb/run-20220303_004215-2ttcbe3s/files/requirements.txt +184 -0
  48. wandb/run-20220303_004215-2ttcbe3s/files/wandb-metadata.json +60 -0
  49. wandb/run-20220303_004215-2ttcbe3s/files/wandb-summary.json +1 -0
  50. wandb/run-20220303_004215-2ttcbe3s/logs/debug-internal.log +138 -0
.gitattributes CHANGED
@@ -58,3 +58,7 @@ wandb/run-20220302_135216-v36bieky/run-v36bieky.wandb filter=lfs diff=lfs merge=
58
  wandb/run-20220302_154455-17zs7rwf/run-17zs7rwf.wandb filter=lfs diff=lfs merge=lfs -text
59
  wandb/run-20220302_180214-gd4yxtv7/run-gd4yxtv7.wandb filter=lfs diff=lfs merge=lfs -text
60
  wandb/run-20220302_200036-31e4k99c/run-31e4k99c.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
58
  wandb/run-20220302_154455-17zs7rwf/run-17zs7rwf.wandb filter=lfs diff=lfs merge=lfs -text
59
  wandb/run-20220302_180214-gd4yxtv7/run-gd4yxtv7.wandb filter=lfs diff=lfs merge=lfs -text
60
  wandb/run-20220302_200036-31e4k99c/run-31e4k99c.wandb filter=lfs diff=lfs merge=lfs -text
61
+ wandb/run-20220302_214437-2u4nhnsf/run-2u4nhnsf.wandb filter=lfs diff=lfs merge=lfs -text
62
+ wandb/run-20220302_222605-10glutwr/run-10glutwr.wandb filter=lfs diff=lfs merge=lfs -text
63
+ wandb/run-20220302_233655-33dtvgaa/run-33dtvgaa.wandb filter=lfs diff=lfs merge=lfs -text
64
+ wandb/run-20220303_004520-25bnjrx1/run-25bnjrx1.wandb filter=lfs diff=lfs merge=lfs -text
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:156ed52f7b75b3b9e2a76618995ada1f013231ce2df078df8390d1a85e4b86d7
3
  size 3210531882
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b81709a6ce9426bdaff484adc1608f8d99dbf7740bc8e82cddfc67464982a7e9
3
  size 3210531882
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1761634195e3a1808d695078d68cd84c32a851cb17c65f31d192af379bc3784b
3
  size 3119
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d43aac09004b2df048bb8c8c596b53ee14f8c791e701e3e501cb95b28c53f18
3
  size 3119
wandb/debug-internal.log CHANGED
@@ -1 +1 @@
1
- run-20220302_200036-31e4k99c/logs/debug-internal.log
1
+ run-20220303_004520-25bnjrx1/logs/debug-internal.log
wandb/debug.log CHANGED
@@ -1 +1 @@
1
- run-20220302_200036-31e4k99c/logs/debug.log
1
+ run-20220303_004520-25bnjrx1/logs/debug.log
wandb/latest-run CHANGED
@@ -1 +1 @@
1
- run-20220302_200036-31e4k99c
1
+ run-20220303_004520-25bnjrx1
wandb/run-20220302_200036-31e4k99c/files/config.yaml CHANGED
@@ -10673,7 +10673,14 @@ _wandb:
10673
  - 1
10674
  - 5
10675
  - 11
 
 
 
 
 
10676
  3:
 
 
10677
  - 13
10678
  4: 3.9.5
10679
  5: 0.12.10
10673
  - 1
10674
  - 5
10675
  - 11
10676
+ 2:
10677
+ - 1
10678
+ - 5
10679
+ - 11
10680
+ - 12
10681
  3:
10682
+ - 1
10683
+ - 7
10684
  - 13
10685
  4: 3.9.5
10686
  5: 0.12.10
wandb/run-20220302_200036-31e4k99c/files/output.log CHANGED
@@ -2792,3 +2792,10 @@ Upload file wandb/run-20220302_200036-31e4k99c/run-31e4k99c.wandb: 43%|██
2792
  eval_samples_per_second = 2.793
2793
  eval_steps_per_second = 0.234
2794
  [INFO|modeling_utils.py:1081] 2022-03-02 21:42:29,550 >> Model weights saved in ./pytorch_model.bin:11<11:47, 3.26s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
 
 
 
 
 
 
 
2792
  eval_samples_per_second = 2.793
2793
  eval_steps_per_second = 0.234
2794
  [INFO|modeling_utils.py:1081] 2022-03-02 21:42:29,550 >> Model weights saved in ./pytorch_model.bin:11<11:47, 3.26s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2795
+ Upload file wandb/run-20220302_200036-31e4k99c/run-31e4k99c.wandb: 0%| | 32.0k/34.7M [00:00<?, ?B/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2796
+ Upload file wandb/run-20220302_200036-31e4k99c/run-31e4k99c.wandb: 0%| | 32.0k/34.7M [00:00<?, ?B/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2797
+ 03/02/2022 21:42:58 - WARNING - huggingface_hub.repository - To https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search
2798
+ Upload file wandb/run-20220302_200036-31e4k99c/run-31e4k99c.wandb: 100%|███████████| 34.7M/34.7M [00:02<00:00, 18.1MB/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2799
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2800
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2801
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
wandb/run-20220302_200036-31e4k99c/files/wandb-summary.json CHANGED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_200036-31e4k99c/logs/debug-internal.log CHANGED
@@ -7522,3 +7522,115 @@
7522
  2022-03-02 21:42:43,423 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: stop_status
7523
  2022-03-02 21:42:43,569 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: stop_status
7524
  2022-03-02 21:42:44,190 INFO Thread-8 :264383 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/config.yaml
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7522
  2022-03-02 21:42:43,423 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: stop_status
7523
  2022-03-02 21:42:43,569 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: stop_status
7524
  2022-03-02 21:42:44,190 INFO Thread-8 :264383 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/config.yaml
7525
+ 2022-03-02 21:42:58,627 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: stop_status
7526
+ 2022-03-02 21:42:58,627 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: stop_status
7527
+ 2022-03-02 21:42:59,195 INFO Thread-8 :264383 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/output.log
7528
+ 2022-03-02 21:43:01,196 INFO Thread-8 :264383 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/output.log
7529
+ 2022-03-02 21:43:02,805 DEBUG SenderThread:264383 [sender.py:send():235] send: stats
7530
+ 2022-03-02 21:43:04,931 DEBUG SenderThread:264383 [sender.py:send():235] send: telemetry
7531
+ 2022-03-02 21:43:04,931 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7532
+ 2022-03-02 21:43:04,931 DEBUG SenderThread:264383 [sender.py:send():235] send: exit
7533
+ 2022-03-02 21:43:04,932 INFO SenderThread:264383 [sender.py:send_exit():371] handling exit code: 1
7534
+ 2022-03-02 21:43:04,932 INFO SenderThread:264383 [sender.py:send_exit():373] handling runtime: 6147
7535
+ 2022-03-02 21:43:04,983 INFO SenderThread:264383 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
7536
+ 2022-03-02 21:43:04,983 INFO SenderThread:264383 [sender.py:send_exit():379] send defer
7537
+ 2022-03-02 21:43:04,984 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7538
+ 2022-03-02 21:43:04,984 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7539
+ 2022-03-02 21:43:04,984 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 0
7540
+ 2022-03-02 21:43:04,984 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7541
+ 2022-03-02 21:43:04,985 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 0
7542
+ 2022-03-02 21:43:04,985 INFO SenderThread:264383 [sender.py:transition_state():392] send defer: 1
7543
+ 2022-03-02 21:43:04,985 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7544
+ 2022-03-02 21:43:04,985 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 1
7545
+ 2022-03-02 21:43:05,048 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7546
+ 2022-03-02 21:43:05,048 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 1
7547
+ 2022-03-02 21:43:05,048 INFO SenderThread:264383 [sender.py:transition_state():392] send defer: 2
7548
+ 2022-03-02 21:43:05,048 DEBUG SenderThread:264383 [sender.py:send():235] send: stats
7549
+ 2022-03-02 21:43:05,049 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7550
+ 2022-03-02 21:43:05,049 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 2
7551
+ 2022-03-02 21:43:05,050 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7552
+ 2022-03-02 21:43:05,050 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 2
7553
+ 2022-03-02 21:43:05,050 INFO SenderThread:264383 [sender.py:transition_state():392] send defer: 3
7554
+ 2022-03-02 21:43:05,050 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7555
+ 2022-03-02 21:43:05,050 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 3
7556
+ 2022-03-02 21:43:05,108 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7557
+ 2022-03-02 21:43:05,111 DEBUG SenderThread:264383 [sender.py:send():235] send: summary
7558
+ 2022-03-02 21:43:05,195 INFO SenderThread:264383 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
7559
+ 2022-03-02 21:43:05,195 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7560
+ 2022-03-02 21:43:05,195 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 3
7561
+ 2022-03-02 21:43:05,195 INFO SenderThread:264383 [sender.py:transition_state():392] send defer: 4
7562
+ 2022-03-02 21:43:05,195 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7563
+ 2022-03-02 21:43:05,196 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7564
+ 2022-03-02 21:43:05,196 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 4
7565
+ 2022-03-02 21:43:05,196 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7566
+ 2022-03-02 21:43:05,196 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 4
7567
+ 2022-03-02 21:43:05,209 INFO Thread-8 :264383 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/output.log
7568
+ 2022-03-02 21:43:05,209 INFO Thread-8 :264383 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/wandb-summary.json
7569
+ 2022-03-02 21:43:05,297 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7570
+ 2022-03-02 21:43:06,125 INFO SenderThread:264383 [sender.py:transition_state():392] send defer: 5
7571
+ 2022-03-02 21:43:06,125 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7572
+ 2022-03-02 21:43:06,126 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7573
+ 2022-03-02 21:43:06,126 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 5
7574
+ 2022-03-02 21:43:06,126 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7575
+ 2022-03-02 21:43:06,126 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 5
7576
+ 2022-03-02 21:43:06,126 INFO SenderThread:264383 [dir_watcher.py:finish():283] shutting down directory watcher
7577
+ 2022-03-02 21:43:06,209 INFO Thread-8 :264383 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/config.yaml
7578
+ 2022-03-02 21:43:06,209 INFO SenderThread:264383 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files
7579
+ 2022-03-02 21:43:06,210 INFO SenderThread:264383 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/wandb-metadata.json wandb-metadata.json
7580
+ 2022-03-02 21:43:06,210 INFO SenderThread:264383 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/output.log output.log
7581
+ 2022-03-02 21:43:06,210 INFO SenderThread:264383 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/wandb-summary.json wandb-summary.json
7582
+ 2022-03-02 21:43:06,210 INFO SenderThread:264383 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/requirements.txt requirements.txt
7583
+ 2022-03-02 21:43:06,211 INFO SenderThread:264383 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/config.yaml config.yaml
7584
+ 2022-03-02 21:43:06,214 INFO SenderThread:264383 [sender.py:transition_state():392] send defer: 6
7585
+ 2022-03-02 21:43:06,222 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7586
+ 2022-03-02 21:43:06,223 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 6
7587
+ 2022-03-02 21:43:06,223 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7588
+ 2022-03-02 21:43:06,223 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 6
7589
+ 2022-03-02 21:43:06,223 INFO SenderThread:264383 [file_pusher.py:finish():177] shutting down file pusher
7590
+ 2022-03-02 21:43:06,227 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7591
+ 2022-03-02 21:43:06,227 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7592
+ 2022-03-02 21:43:06,328 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7593
+ 2022-03-02 21:43:06,329 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7594
+ 2022-03-02 21:43:06,430 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7595
+ 2022-03-02 21:43:06,430 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7596
+ 2022-03-02 21:43:06,527 INFO Thread-14 :264383 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/requirements.txt
7597
+ 2022-03-02 21:43:06,532 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7598
+ 2022-03-02 21:43:06,532 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7599
+ 2022-03-02 21:43:06,540 INFO Thread-12 :264383 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/output.log
7600
+ 2022-03-02 21:43:06,575 INFO Thread-13 :264383 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/wandb-summary.json
7601
+ 2022-03-02 21:43:06,584 INFO Thread-15 :264383 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/files/config.yaml
7602
+ 2022-03-02 21:43:06,634 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7603
+ 2022-03-02 21:43:06,634 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7604
+ 2022-03-02 21:43:06,735 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7605
+ 2022-03-02 21:43:06,736 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7606
+ 2022-03-02 21:43:06,784 INFO Thread-7 :264383 [sender.py:transition_state():392] send defer: 7
7607
+ 2022-03-02 21:43:06,785 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7608
+ 2022-03-02 21:43:06,785 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 7
7609
+ 2022-03-02 21:43:06,785 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7610
+ 2022-03-02 21:43:06,785 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 7
7611
+ 2022-03-02 21:43:06,837 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7612
+ 2022-03-02 21:43:08,117 INFO SenderThread:264383 [sender.py:transition_state():392] send defer: 8
7613
+ 2022-03-02 21:43:08,117 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7614
+ 2022-03-02 21:43:08,118 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7615
+ 2022-03-02 21:43:08,118 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 8
7616
+ 2022-03-02 21:43:08,118 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7617
+ 2022-03-02 21:43:08,118 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 8
7618
+ 2022-03-02 21:43:08,118 INFO SenderThread:264383 [sender.py:transition_state():392] send defer: 9
7619
+ 2022-03-02 21:43:08,119 DEBUG SenderThread:264383 [sender.py:send():235] send: final
7620
+ 2022-03-02 21:43:08,120 DEBUG SenderThread:264383 [sender.py:send():235] send: footer
7621
+ 2022-03-02 21:43:08,120 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: defer
7622
+ 2022-03-02 21:43:08,121 INFO HandlerThread:264383 [handler.py:handle_request_defer():154] handle defer: 9
7623
+ 2022-03-02 21:43:08,121 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: defer
7624
+ 2022-03-02 21:43:08,121 INFO SenderThread:264383 [sender.py:send_request_defer():388] handle sender defer: 9
7625
+ 2022-03-02 21:43:08,219 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: poll_exit
7626
+ 2022-03-02 21:43:08,219 DEBUG SenderThread:264383 [sender.py:send_request():249] send_request: poll_exit
7627
+ 2022-03-02 21:43:08,219 INFO SenderThread:264383 [file_pusher.py:join():182] waiting for file pusher
7628
+ 2022-03-02 21:43:08,285 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: get_summary
7629
+ 2022-03-02 21:43:08,384 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: sampled_history
7630
+ 2022-03-02 21:43:08,387 DEBUG HandlerThread:264383 [handler.py:handle_request():131] handle_request: shutdown
7631
+ 2022-03-02 21:43:08,387 INFO HandlerThread:264383 [handler.py:finish():739] shutting down handler
7632
+ 2022-03-02 21:43:09,120 INFO WriterThread:264383 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_200036-31e4k99c/run-31e4k99c.wandb
7633
+ 2022-03-02 21:43:09,284 INFO SenderThread:264383 [sender.py:finish():1075] shutting down sender
7634
+ 2022-03-02 21:43:09,284 INFO SenderThread:264383 [file_pusher.py:finish():177] shutting down file pusher
7635
+ 2022-03-02 21:43:09,284 INFO SenderThread:264383 [file_pusher.py:join():182] waiting for file pusher
7636
+ 2022-03-02 21:43:09,291 INFO MainThread:264383 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220302_200036-31e4k99c/logs/debug.log CHANGED
@@ -25,3 +25,101 @@ config: {}
25
  2022-03-02 20:00:37,941 INFO MainThread:264283 [wandb_init.py:init():651] run started, returning control to user process
26
  2022-03-02 20:00:37,944 INFO MainThread:264283 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0003, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_19-59-53_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 12, 'eval_batch_size': 12}
27
  2022-03-02 20:00:37,946 INFO MainThread:264283 [wandb_watch.py:watch():43] Watching
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  2022-03-02 20:00:37,941 INFO MainThread:264283 [wandb_init.py:init():651] run started, returning control to user process
26
  2022-03-02 20:00:37,944 INFO MainThread:264283 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0003, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_19-59-53_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 12, 'eval_batch_size': 12}
27
  2022-03-02 20:00:37,946 INFO MainThread:264283 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-02 21:43:02,636 INFO MainThread:264283 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-02 21:43:02,640 INFO MainThread:264283 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-02 21:43:04,984 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-02 21:43:05,196 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-02 21:43:06,126 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 2095
52
+ }
53
+
54
+ 2022-03-02 21:43:06,227 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 5
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2095
59
+ total_bytes: 2674763
60
+ }
61
+
62
+ 2022-03-02 21:43:06,329 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 2041516
67
+ total_bytes: 2674763
68
+ }
69
+
70
+ 2022-03-02 21:43:06,431 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 2674763
75
+ total_bytes: 2674763
76
+ }
77
+
78
+ 2022-03-02 21:43:06,533 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 2674763
83
+ total_bytes: 2674763
84
+ }
85
+
86
+ 2022-03-02 21:43:06,634 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 2674763
91
+ total_bytes: 2674763
92
+ }
93
+
94
+ 2022-03-02 21:43:06,736 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 2674763
99
+ total_bytes: 2674763
100
+ }
101
+
102
+ 2022-03-02 21:43:08,118 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 2674763
107
+ total_bytes: 2674763
108
+ }
109
+
110
+ 2022-03-02 21:43:08,284 INFO MainThread:264283 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
111
+ exit_result {
112
+ }
113
+ file_counts {
114
+ wandb_count: 5
115
+ }
116
+ pusher_stats {
117
+ uploaded_bytes: 2674763
118
+ total_bytes: 2674763
119
+ }
120
+ local_info {
121
+ }
122
+
123
+ 2022-03-02 21:43:09,438 INFO MainThread:264283 [wandb_run.py:_append_history():2144] rendering history
124
+ 2022-03-02 21:43:09,439 INFO MainThread:264283 [wandb_run.py:_append_summary():2102] rendering summary
125
+ 2022-03-02 21:43:09,440 INFO MainThread:264283 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220302_200036-31e4k99c/run-31e4k99c.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa51a0a26ce5e172e476442812ae6212e91addd95b58f5626c318ed39b664595
3
- size 36405850
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7bba159d4dadd3685bda07a160b03aef6a022bcd1ffe05c918057d8d7e27558
3
+ size 37907046
wandb/run-20220302_214437-2u4nhnsf/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_214437-2u4nhnsf/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_214437-2u4nhnsf/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220302_214437-2u4nhnsf/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-02T21:44:38.711819",
5
+ "startedAt": "2022-03-02T21:44:37.567182",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=2",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "8c7181143c175387040dc1a6ac2ddbc9179b550c"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220302_214437-2u4nhnsf/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_214437-2u4nhnsf/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_214437-2u4nhnsf/logs/debug.log ADDED
@@ -0,0 +1,149 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-02 21:44:37,569 INFO MainThread:265590 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-02 21:44:37,569 INFO MainThread:265590 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-02 21:44:37,570 INFO MainThread:265590 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-02 21:44:37,570 INFO MainThread:265590 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-02 21:44:37,570 INFO MainThread:265590 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_214437-2u4nhnsf/logs/debug.log
6
+ 2022-03-02 21:44:37,570 INFO MainThread:265590 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_214437-2u4nhnsf/logs/debug-internal.log
7
+ 2022-03-02 21:44:37,570 INFO MainThread:265590 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-02 21:44:37,570 INFO MainThread:265590 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-02 21:44:37,570 INFO MainThread:265590 [wandb_init.py:init():471] starting backend
11
+ 2022-03-02 21:44:37,570 INFO MainThread:265590 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-02 21:44:37,641 INFO MainThread:265590 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-02 21:44:37,709 INFO MainThread:265590 [backend.py:ensure_launched():224] started backend process with pid: 265689
14
+ 2022-03-02 21:44:37,711 INFO MainThread:265590 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-02 21:44:37,721 INFO MainThread:265590 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-02 21:44:37,884 INFO MainThread:265590 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-02 21:44:38,595 INFO MainThread:265590 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.11 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
18
+
19
+ 2022-03-02 21:44:38,595 INFO MainThread:265590 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
20
+ 2022-03-02 21:44:38,705 INFO MainThread:265590 [wandb_init.py:init():624] starting run threads in backend
21
+ 2022-03-02 21:44:38,844 INFO MainThread:265590 [wandb_run.py:_console_start():1827] atexit reg
22
+ 2022-03-02 21:44:38,844 INFO MainThread:265590 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
23
+ 2022-03-02 21:44:38,845 INFO MainThread:265590 [wandb_run.py:_redirect():1706] Redirecting console.
24
+ 2022-03-02 21:44:38,846 INFO MainThread:265590 [wandb_run.py:_redirect():1762] Redirects installed.
25
+ 2022-03-02 21:44:38,847 INFO MainThread:265590 [wandb_init.py:init():651] run started, returning control to user process
26
+ 2022-03-02 21:44:38,849 INFO MainThread:265590 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0003, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_21-43-55_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
27
+ 2022-03-02 21:44:38,852 INFO MainThread:265590 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-02 22:24:38,308 INFO MainThread:265590 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-02 22:24:38,309 INFO MainThread:265590 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-02 22:24:40,737 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-02 22:24:40,855 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-02 22:24:41,005 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 2095
52
+ }
53
+
54
+ 2022-03-02 22:24:41,904 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 1
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2095
59
+ total_bytes: 2095
60
+ }
61
+
62
+ 2022-03-02 22:24:42,648 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 2095
67
+ total_bytes: 2248104
68
+ }
69
+
70
+ 2022-03-02 22:24:42,750 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 2095
75
+ total_bytes: 2248104
76
+ }
77
+
78
+ 2022-03-02 22:24:42,851 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 2244867
83
+ total_bytes: 2248104
84
+ }
85
+
86
+ 2022-03-02 22:24:42,953 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 2248104
91
+ total_bytes: 2248104
92
+ }
93
+
94
+ 2022-03-02 22:24:43,055 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 2248104
99
+ total_bytes: 2248104
100
+ }
101
+
102
+ 2022-03-02 22:24:43,157 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 2248104
107
+ total_bytes: 2248104
108
+ }
109
+
110
+ 2022-03-02 22:24:43,258 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
111
+ wandb_count: 5
112
+ }
113
+ pusher_stats {
114
+ uploaded_bytes: 2248104
115
+ total_bytes: 2248104
116
+ }
117
+
118
+ 2022-03-02 22:24:43,360 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
119
+ wandb_count: 5
120
+ }
121
+ pusher_stats {
122
+ uploaded_bytes: 2248104
123
+ total_bytes: 2248104
124
+ }
125
+
126
+ 2022-03-02 22:24:44,699 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
127
+ wandb_count: 5
128
+ }
129
+ pusher_stats {
130
+ uploaded_bytes: 2248104
131
+ total_bytes: 2248104
132
+ }
133
+
134
+ 2022-03-02 22:24:44,867 INFO MainThread:265590 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
135
+ exit_result {
136
+ }
137
+ file_counts {
138
+ wandb_count: 5
139
+ }
140
+ pusher_stats {
141
+ uploaded_bytes: 2248104
142
+ total_bytes: 2248104
143
+ }
144
+ local_info {
145
+ }
146
+
147
+ 2022-03-02 22:24:46,025 INFO MainThread:265590 [wandb_run.py:_append_history():2144] rendering history
148
+ 2022-03-02 22:24:46,026 INFO MainThread:265590 [wandb_run.py:_append_summary():2102] rendering summary
149
+ 2022-03-02 22:24:46,027 INFO MainThread:265590 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220302_214437-2u4nhnsf/run-2u4nhnsf.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2097f4493d43dc5464038250e585dd89e0b00de7e3c675d9afcbfe58a143e87
3
+ size 17687292
wandb/run-20220302_222605-10glutwr/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_222605-10glutwr/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_222605-10glutwr/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220302_222605-10glutwr/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-02T22:26:06.164288",
5
+ "startedAt": "2022-03-02T22:26:05.008436",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=4",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "8c7181143c175387040dc1a6ac2ddbc9179b550c"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220302_222605-10glutwr/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_222605-10glutwr/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_222605-10glutwr/logs/debug.log ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_222605-10glutwr/logs/debug.log
6
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_222605-10glutwr/logs/debug-internal.log
7
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [wandb_init.py:init():471] starting backend
11
+ 2022-03-02 22:26:05,010 INFO MainThread:266400 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-02 22:26:05,078 INFO MainThread:266400 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-02 22:26:05,143 INFO MainThread:266400 [backend.py:ensure_launched():224] started backend process with pid: 266499
14
+ 2022-03-02 22:26:05,145 INFO MainThread:266400 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-02 22:26:05,156 INFO MainThread:266400 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-02 22:26:05,315 INFO MainThread:266400 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-02 22:26:06,059 INFO MainThread:266400 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.11 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
18
+
19
+ 2022-03-02 22:26:06,060 INFO MainThread:266400 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
20
+ 2022-03-02 22:26:06,158 INFO MainThread:266400 [wandb_init.py:init():624] starting run threads in backend
21
+ 2022-03-02 22:26:06,273 INFO MainThread:266400 [wandb_run.py:_console_start():1827] atexit reg
22
+ 2022-03-02 22:26:06,274 INFO MainThread:266400 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
23
+ 2022-03-02 22:26:06,274 INFO MainThread:266400 [wandb_run.py:_redirect():1706] Redirecting console.
24
+ 2022-03-02 22:26:06,276 INFO MainThread:266400 [wandb_run.py:_redirect():1762] Redirects installed.
25
+ 2022-03-02 22:26:06,276 INFO MainThread:266400 [wandb_init.py:init():651] run started, returning control to user process
26
+ 2022-03-02 22:26:06,280 INFO MainThread:266400 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0003, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_22-25-22_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
27
+ 2022-03-02 22:26:06,285 INFO MainThread:266400 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-02 23:35:20,661 INFO MainThread:266400 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-02 23:35:20,662 INFO MainThread:266400 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-02 23:35:23,045 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-02 23:35:23,244 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-02 23:35:24,169 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 2095
52
+ }
53
+
54
+ 2022-03-02 23:35:24,471 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 3
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2095
59
+ total_bytes: 2051200
60
+ }
61
+
62
+ 2022-03-02 23:35:24,574 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 174127
67
+ total_bytes: 2352011
68
+ }
69
+
70
+ 2022-03-02 23:35:24,676 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 2352011
75
+ total_bytes: 2352011
76
+ }
77
+
78
+ 2022-03-02 23:35:24,777 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 2352011
83
+ total_bytes: 2352011
84
+ }
85
+
86
+ 2022-03-02 23:35:24,879 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 2352011
91
+ total_bytes: 2352011
92
+ }
93
+
94
+ 2022-03-02 23:35:24,981 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 2352011
99
+ total_bytes: 2352011
100
+ }
101
+
102
+ 2022-03-02 23:35:26,630 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 2352011
107
+ total_bytes: 2352011
108
+ }
109
+
110
+ 2022-03-02 23:35:26,787 INFO MainThread:266400 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
111
+ exit_result {
112
+ }
113
+ file_counts {
114
+ wandb_count: 5
115
+ }
116
+ pusher_stats {
117
+ uploaded_bytes: 2352011
118
+ total_bytes: 2352011
119
+ }
120
+ local_info {
121
+ }
122
+
123
+ 2022-03-02 23:35:27,944 INFO MainThread:266400 [wandb_run.py:_append_history():2144] rendering history
124
+ 2022-03-02 23:35:27,945 INFO MainThread:266400 [wandb_run.py:_append_summary():2102] rendering summary
125
+ 2022-03-02 23:35:27,946 INFO MainThread:266400 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220302_222605-10glutwr/run-10glutwr.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac6d6670d416568eed6d71f6125deb820b6762806b784e9aa0f2962ef5170dc2
3
+ size 29171596
wandb/run-20220302_233655-33dtvgaa/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_233655-33dtvgaa/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_233655-33dtvgaa/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220302_233655-33dtvgaa/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-02T23:36:56.434217",
5
+ "startedAt": "2022-03-02T23:36:55.283203",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=8",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "8c7181143c175387040dc1a6ac2ddbc9179b550c"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220302_233655-33dtvgaa/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_233655-33dtvgaa/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220302_233655-33dtvgaa/logs/debug.log ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-02 23:36:55,285 INFO MainThread:266634 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-02 23:36:55,285 INFO MainThread:266634 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-02 23:36:55,285 INFO MainThread:266634 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-02 23:36:55,285 INFO MainThread:266634 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-02 23:36:55,285 INFO MainThread:266634 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_233655-33dtvgaa/logs/debug.log
6
+ 2022-03-02 23:36:55,285 INFO MainThread:266634 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220302_233655-33dtvgaa/logs/debug-internal.log
7
+ 2022-03-02 23:36:55,285 INFO MainThread:266634 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-02 23:36:55,285 INFO MainThread:266634 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-02 23:36:55,285 INFO MainThread:266634 [wandb_init.py:init():471] starting backend
11
+ 2022-03-02 23:36:55,286 INFO MainThread:266634 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-02 23:36:55,362 INFO MainThread:266634 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-02 23:36:55,432 INFO MainThread:266634 [backend.py:ensure_launched():224] started backend process with pid: 266733
14
+ 2022-03-02 23:36:55,434 INFO MainThread:266634 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-02 23:36:55,444 INFO MainThread:266634 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-02 23:36:55,608 INFO MainThread:266634 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-02 23:36:56,328 INFO MainThread:266634 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.11 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
18
+
19
+ 2022-03-02 23:36:56,328 INFO MainThread:266634 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
20
+ 2022-03-02 23:36:56,427 INFO MainThread:266634 [wandb_init.py:init():624] starting run threads in backend
21
+ 2022-03-02 23:36:56,564 INFO MainThread:266634 [wandb_run.py:_console_start():1827] atexit reg
22
+ 2022-03-02 23:36:56,564 INFO MainThread:266634 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
23
+ 2022-03-02 23:36:56,565 INFO MainThread:266634 [wandb_run.py:_redirect():1706] Redirecting console.
24
+ 2022-03-02 23:36:56,566 INFO MainThread:266634 [wandb_run.py:_redirect():1762] Redirects installed.
25
+ 2022-03-02 23:36:56,567 INFO MainThread:266634 [wandb_init.py:init():651] run started, returning control to user process
26
+ 2022-03-02 23:36:56,569 INFO MainThread:266634 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0003, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar02_23-36-13_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
27
+ 2022-03-02 23:36:56,572 INFO MainThread:266634 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-03 00:38:55,561 INFO MainThread:266634 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-03 00:38:55,563 INFO MainThread:266634 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-03 00:38:58,456 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-03 00:38:58,586 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-03 00:38:58,772 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 2095
52
+ }
53
+
54
+ 2022-03-03 00:38:59,764 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 1
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2095
59
+ total_bytes: 2095
60
+ }
61
+
62
+ 2022-03-03 00:39:00,513 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 4
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 2095
67
+ total_bytes: 1999756
68
+ }
69
+
70
+ 2022-03-03 00:39:00,615 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 299669
75
+ total_bytes: 2297330
76
+ }
77
+
78
+ 2022-03-03 00:39:00,718 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 2297330
83
+ total_bytes: 2297330
84
+ }
85
+
86
+ 2022-03-03 00:39:00,820 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 2297330
91
+ total_bytes: 2297330
92
+ }
93
+
94
+ 2022-03-03 00:39:00,923 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 2297330
99
+ total_bytes: 2297330
100
+ }
101
+
102
+ 2022-03-03 00:39:01,026 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 2297330
107
+ total_bytes: 2297330
108
+ }
109
+
110
+ 2022-03-03 00:39:02,762 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
111
+ wandb_count: 5
112
+ }
113
+ pusher_stats {
114
+ uploaded_bytes: 2297330
115
+ total_bytes: 2297330
116
+ }
117
+
118
+ 2022-03-03 00:39:02,929 INFO MainThread:266634 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
119
+ exit_result {
120
+ }
121
+ file_counts {
122
+ wandb_count: 5
123
+ }
124
+ pusher_stats {
125
+ uploaded_bytes: 2297330
126
+ total_bytes: 2297330
127
+ }
128
+ local_info {
129
+ }
130
+
131
+ 2022-03-03 00:39:04,125 INFO MainThread:266634 [wandb_run.py:_append_history():2144] rendering history
132
+ 2022-03-03 00:39:04,126 INFO MainThread:266634 [wandb_run.py:_append_summary():2102] rendering summary
133
+ 2022-03-03 00:39:04,127 INFO MainThread:266634 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220302_233655-33dtvgaa/run-33dtvgaa.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12a5d36de07abad2de3fa12743b03c3105d2a4cedeb9e2f608b915a435c08667
3
+ size 26660446
wandb/run-20220303_004039-1a18zsso/files/config.yaml ADDED
@@ -0,0 +1,713 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: ./
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.10
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.9.5
22
+ start_time: 1646268039
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.9.5
37
+ 5: 0.12.10
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ adafactor:
42
+ desc: null
43
+ value: false
44
+ adam_beta1:
45
+ desc: null
46
+ value: 0.9
47
+ adam_beta2:
48
+ desc: null
49
+ value: 0.999
50
+ adam_epsilon:
51
+ desc: null
52
+ value: 1.0e-08
53
+ add_cross_attention:
54
+ desc: null
55
+ value: false
56
+ architectures:
57
+ desc: null
58
+ value:
59
+ - SpeechEncoderDecoderModel
60
+ bad_words_ids:
61
+ desc: null
62
+ value: null
63
+ bf16:
64
+ desc: null
65
+ value: false
66
+ bf16_full_eval:
67
+ desc: null
68
+ value: false
69
+ bos_token_id:
70
+ desc: null
71
+ value: null
72
+ chunk_size_feed_forward:
73
+ desc: null
74
+ value: 0
75
+ cross_attention_hidden_size:
76
+ desc: null
77
+ value: null
78
+ dataloader_drop_last:
79
+ desc: null
80
+ value: false
81
+ dataloader_num_workers:
82
+ desc: null
83
+ value: 0
84
+ dataloader_pin_memory:
85
+ desc: null
86
+ value: true
87
+ ddp_bucket_cap_mb:
88
+ desc: null
89
+ value: None
90
+ ddp_find_unused_parameters:
91
+ desc: null
92
+ value: None
93
+ debug:
94
+ desc: null
95
+ value: '[]'
96
+ decoder:
97
+ desc: null
98
+ value:
99
+ _name_or_path: gpt2-medium
100
+ activation_function: gelu_new
101
+ add_cross_attention: true
102
+ architectures:
103
+ - GPT2LMHeadModel
104
+ attn_pdrop: 0.0
105
+ bad_words_ids: null
106
+ bos_token_id: 50256
107
+ chunk_size_feed_forward: 0
108
+ cross_attention_hidden_size: null
109
+ decoder_start_token_id: null
110
+ diversity_penalty: 0.0
111
+ do_sample: false
112
+ early_stopping: false
113
+ embd_pdrop: 0.0
114
+ encoder_no_repeat_ngram_size: 0
115
+ eos_token_id: 50256
116
+ finetuning_task: null
117
+ forced_bos_token_id: null
118
+ forced_eos_token_id: null
119
+ id2label:
120
+ '0': LABEL_0
121
+ '1': LABEL_1
122
+ initializer_range: 0.02
123
+ is_decoder: true
124
+ is_encoder_decoder: false
125
+ label2id:
126
+ LABEL_0: 0
127
+ LABEL_1: 1
128
+ layer_norm_epsilon: 1.0e-05
129
+ length_penalty: 1.0
130
+ max_length: 20
131
+ min_length: 0
132
+ model_type: gpt2
133
+ n_ctx: 1024
134
+ n_embd: 1024
135
+ n_head: 16
136
+ n_inner: null
137
+ n_layer: 24
138
+ n_positions: 1024
139
+ n_special: 0
140
+ no_repeat_ngram_size: 0
141
+ num_beam_groups: 1
142
+ num_beams: 1
143
+ num_return_sequences: 1
144
+ output_attentions: false
145
+ output_hidden_states: false
146
+ output_scores: false
147
+ pad_token_id: null
148
+ predict_special_tokens: true
149
+ prefix: null
150
+ problem_type: null
151
+ pruned_heads: {}
152
+ remove_invalid_values: false
153
+ reorder_and_upcast_attn: false
154
+ repetition_penalty: 1.0
155
+ resid_pdrop: 0.0
156
+ return_dict: true
157
+ return_dict_in_generate: false
158
+ scale_attn_by_inverse_layer_idx: false
159
+ scale_attn_weights: true
160
+ sep_token_id: null
161
+ summary_activation: null
162
+ summary_first_dropout: 0.0
163
+ summary_proj_to_labels: true
164
+ summary_type: cls_index
165
+ summary_use_proj: true
166
+ task_specific_params:
167
+ text-generation:
168
+ do_sample: true
169
+ max_length: 50
170
+ temperature: 1.0
171
+ tie_encoder_decoder: false
172
+ tie_word_embeddings: true
173
+ tokenizer_class: null
174
+ top_k: 50
175
+ top_p: 1.0
176
+ torch_dtype: null
177
+ torchscript: false
178
+ transformers_version: 4.17.0.dev0
179
+ use_bfloat16: false
180
+ use_cache: false
181
+ vocab_size: 50257
182
+ decoder_start_token_id:
183
+ desc: null
184
+ value: 50256
185
+ deepspeed:
186
+ desc: null
187
+ value: None
188
+ disable_tqdm:
189
+ desc: null
190
+ value: false
191
+ diversity_penalty:
192
+ desc: null
193
+ value: 0.0
194
+ do_eval:
195
+ desc: null
196
+ value: true
197
+ do_predict:
198
+ desc: null
199
+ value: false
200
+ do_sample:
201
+ desc: null
202
+ value: false
203
+ do_train:
204
+ desc: null
205
+ value: true
206
+ early_stopping:
207
+ desc: null
208
+ value: false
209
+ encoder:
210
+ desc: null
211
+ value:
212
+ _name_or_path: facebook/wav2vec2-large-lv60
213
+ activation_dropout: 0.0
214
+ adapter_kernel_size: 3
215
+ adapter_stride: 2
216
+ add_adapter: true
217
+ add_cross_attention: false
218
+ apply_spec_augment: false
219
+ architectures:
220
+ - Wav2Vec2ForPreTraining
221
+ attention_dropout: 0.0
222
+ bad_words_ids: null
223
+ bos_token_id: 1
224
+ chunk_size_feed_forward: 0
225
+ classifier_proj_size: 256
226
+ codevector_dim: 768
227
+ contrastive_logits_temperature: 0.1
228
+ conv_bias: true
229
+ conv_dim:
230
+ - 512
231
+ - 512
232
+ - 512
233
+ - 512
234
+ - 512
235
+ - 512
236
+ - 512
237
+ conv_kernel:
238
+ - 10
239
+ - 3
240
+ - 3
241
+ - 3
242
+ - 3
243
+ - 2
244
+ - 2
245
+ conv_stride:
246
+ - 5
247
+ - 2
248
+ - 2
249
+ - 2
250
+ - 2
251
+ - 2
252
+ - 2
253
+ cross_attention_hidden_size: null
254
+ ctc_loss_reduction: sum
255
+ ctc_zero_infinity: false
256
+ decoder_start_token_id: null
257
+ diversity_loss_weight: 0.1
258
+ diversity_penalty: 0.0
259
+ do_sample: false
260
+ do_stable_layer_norm: true
261
+ early_stopping: false
262
+ encoder_no_repeat_ngram_size: 0
263
+ eos_token_id: 2
264
+ feat_extract_activation: gelu
265
+ feat_extract_dropout: 0.0
266
+ feat_extract_norm: layer
267
+ feat_proj_dropout: 0.0
268
+ feat_quantizer_dropout: 0.0
269
+ final_dropout: 0.0
270
+ finetuning_task: null
271
+ forced_bos_token_id: null
272
+ forced_eos_token_id: null
273
+ gradient_checkpointing: false
274
+ hidden_act: gelu
275
+ hidden_dropout: 0.0
276
+ hidden_dropout_prob: 0.0
277
+ hidden_size: 1024
278
+ id2label:
279
+ '0': LABEL_0
280
+ '1': LABEL_1
281
+ initializer_range: 0.02
282
+ intermediate_size: 4096
283
+ is_decoder: false
284
+ is_encoder_decoder: false
285
+ label2id:
286
+ LABEL_0: 0
287
+ LABEL_1: 1
288
+ layer_norm_eps: 1.0e-05
289
+ layerdrop: 0.0
290
+ length_penalty: 1.0
291
+ mask_feature_length: 10
292
+ mask_feature_min_masks: 0
293
+ mask_feature_prob: 0.0
294
+ mask_time_length: 10
295
+ mask_time_min_masks: 2
296
+ mask_time_prob: 0.0
297
+ max_length: 20
298
+ min_length: 0
299
+ model_type: wav2vec2
300
+ no_repeat_ngram_size: 0
301
+ num_adapter_layers: 3
302
+ num_attention_heads: 16
303
+ num_beam_groups: 1
304
+ num_beams: 1
305
+ num_codevector_groups: 2
306
+ num_codevectors_per_group: 320
307
+ num_conv_pos_embedding_groups: 16
308
+ num_conv_pos_embeddings: 128
309
+ num_feat_extract_layers: 7
310
+ num_hidden_layers: 24
311
+ num_negatives: 100
312
+ num_return_sequences: 1
313
+ output_attentions: false
314
+ output_hidden_size: 1024
315
+ output_hidden_states: false
316
+ output_scores: false
317
+ pad_token_id: 0
318
+ prefix: null
319
+ problem_type: null
320
+ proj_codevector_dim: 768
321
+ pruned_heads: {}
322
+ remove_invalid_values: false
323
+ repetition_penalty: 1.0
324
+ return_dict: true
325
+ return_dict_in_generate: false
326
+ sep_token_id: null
327
+ task_specific_params: null
328
+ tdnn_dilation:
329
+ - 1
330
+ - 2
331
+ - 3
332
+ - 1
333
+ - 1
334
+ tdnn_dim:
335
+ - 512
336
+ - 512
337
+ - 512
338
+ - 512
339
+ - 1500
340
+ tdnn_kernel:
341
+ - 5
342
+ - 3
343
+ - 3
344
+ - 1
345
+ - 1
346
+ temperature: 1.0
347
+ tie_encoder_decoder: false
348
+ tie_word_embeddings: true
349
+ tokenizer_class: null
350
+ top_k: 50
351
+ top_p: 1.0
352
+ torch_dtype: null
353
+ torchscript: false
354
+ transformers_version: 4.17.0.dev0
355
+ use_bfloat16: false
356
+ use_weighted_layer_sum: false
357
+ vocab_size: 32
358
+ xvector_output_dim: 512
359
+ encoder_no_repeat_ngram_size:
360
+ desc: null
361
+ value: 0
362
+ eos_token_id:
363
+ desc: null
364
+ value: 50256
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: None
368
+ eval_batch_size:
369
+ desc: null
370
+ value: 16
371
+ eval_steps:
372
+ desc: null
373
+ value: 500
374
+ evaluation_strategy:
375
+ desc: null
376
+ value: steps
377
+ finetuning_task:
378
+ desc: null
379
+ value: null
380
+ forced_bos_token_id:
381
+ desc: null
382
+ value: null
383
+ forced_eos_token_id:
384
+ desc: null
385
+ value: null
386
+ fp16:
387
+ desc: null
388
+ value: true
389
+ fp16_backend:
390
+ desc: null
391
+ value: auto
392
+ fp16_full_eval:
393
+ desc: null
394
+ value: false
395
+ fp16_opt_level:
396
+ desc: null
397
+ value: O1
398
+ generation_max_length:
399
+ desc: null
400
+ value: 40
401
+ generation_num_beams:
402
+ desc: null
403
+ value: 1
404
+ gradient_accumulation_steps:
405
+ desc: null
406
+ value: 2
407
+ gradient_checkpointing:
408
+ desc: null
409
+ value: true
410
+ greater_is_better:
411
+ desc: null
412
+ value: None
413
+ group_by_length:
414
+ desc: null
415
+ value: true
416
+ half_precision_backend:
417
+ desc: null
418
+ value: amp
419
+ hub_model_id:
420
+ desc: null
421
+ value: None
422
+ hub_strategy:
423
+ desc: null
424
+ value: every_save
425
+ hub_token:
426
+ desc: null
427
+ value: <HUB_TOKEN>
428
+ id2label:
429
+ desc: null
430
+ value:
431
+ '0': LABEL_0
432
+ '1': LABEL_1
433
+ ignore_data_skip:
434
+ desc: null
435
+ value: false
436
+ is_decoder:
437
+ desc: null
438
+ value: false
439
+ is_encoder_decoder:
440
+ desc: null
441
+ value: true
442
+ label2id:
443
+ desc: null
444
+ value:
445
+ LABEL_0: 0
446
+ LABEL_1: 1
447
+ label_names:
448
+ desc: null
449
+ value: None
450
+ label_smoothing_factor:
451
+ desc: null
452
+ value: 0.0
453
+ learning_rate:
454
+ desc: null
455
+ value: 0.0003
456
+ length_column_name:
457
+ desc: null
458
+ value: input_length
459
+ length_penalty:
460
+ desc: null
461
+ value: 1.0
462
+ load_best_model_at_end:
463
+ desc: null
464
+ value: false
465
+ local_rank:
466
+ desc: null
467
+ value: -1
468
+ log_level:
469
+ desc: null
470
+ value: -1
471
+ log_level_replica:
472
+ desc: null
473
+ value: -1
474
+ log_on_each_node:
475
+ desc: null
476
+ value: true
477
+ logging_dir:
478
+ desc: null
479
+ value: ./runs/Mar03_00-39-53_sanchit--v100
480
+ logging_first_step:
481
+ desc: null
482
+ value: false
483
+ logging_nan_inf_filter:
484
+ desc: null
485
+ value: true
486
+ logging_steps:
487
+ desc: null
488
+ value: 1
489
+ logging_strategy:
490
+ desc: null
491
+ value: steps
492
+ lr_scheduler_type:
493
+ desc: null
494
+ value: linear
495
+ max_grad_norm:
496
+ desc: null
497
+ value: 1.0
498
+ max_length:
499
+ desc: null
500
+ value: 50
501
+ max_steps:
502
+ desc: null
503
+ value: -1
504
+ metric_for_best_model:
505
+ desc: null
506
+ value: None
507
+ min_length:
508
+ desc: null
509
+ value: 0
510
+ model_type:
511
+ desc: null
512
+ value: speech-encoder-decoder
513
+ mp_parameters:
514
+ desc: null
515
+ value: ''
516
+ no_cuda:
517
+ desc: null
518
+ value: false
519
+ no_repeat_ngram_size:
520
+ desc: null
521
+ value: 0
522
+ num_beam_groups:
523
+ desc: null
524
+ value: 1
525
+ num_beams:
526
+ desc: null
527
+ value: 1
528
+ num_return_sequences:
529
+ desc: null
530
+ value: 1
531
+ num_train_epochs:
532
+ desc: null
533
+ value: 1.0
534
+ optim:
535
+ desc: null
536
+ value: adamw_hf
537
+ output_attentions:
538
+ desc: null
539
+ value: false
540
+ output_dir:
541
+ desc: null
542
+ value: ./
543
+ output_hidden_states:
544
+ desc: null
545
+ value: false
546
+ output_scores:
547
+ desc: null
548
+ value: false
549
+ overwrite_output_dir:
550
+ desc: null
551
+ value: true
552
+ pad_token_id:
553
+ desc: null
554
+ value: 50256
555
+ past_index:
556
+ desc: null
557
+ value: -1
558
+ per_device_eval_batch_size:
559
+ desc: null
560
+ value: 16
561
+ per_device_train_batch_size:
562
+ desc: null
563
+ value: 16
564
+ per_gpu_eval_batch_size:
565
+ desc: null
566
+ value: None
567
+ per_gpu_train_batch_size:
568
+ desc: null
569
+ value: None
570
+ predict_with_generate:
571
+ desc: null
572
+ value: true
573
+ prediction_loss_only:
574
+ desc: null
575
+ value: false
576
+ prefix:
577
+ desc: null
578
+ value: null
579
+ problem_type:
580
+ desc: null
581
+ value: null
582
+ processor_class:
583
+ desc: null
584
+ value: Wav2Vec2Processor
585
+ pruned_heads:
586
+ desc: null
587
+ value: {}
588
+ push_to_hub:
589
+ desc: null
590
+ value: true
591
+ push_to_hub_model_id:
592
+ desc: null
593
+ value: None
594
+ push_to_hub_organization:
595
+ desc: null
596
+ value: None
597
+ push_to_hub_token:
598
+ desc: null
599
+ value: <PUSH_TO_HUB_TOKEN>
600
+ remove_invalid_values:
601
+ desc: null
602
+ value: false
603
+ remove_unused_columns:
604
+ desc: null
605
+ value: true
606
+ repetition_penalty:
607
+ desc: null
608
+ value: 1.0
609
+ report_to:
610
+ desc: null
611
+ value: '[''wandb'']'
612
+ resume_from_checkpoint:
613
+ desc: null
614
+ value: None
615
+ return_dict:
616
+ desc: null
617
+ value: true
618
+ return_dict_in_generate:
619
+ desc: null
620
+ value: false
621
+ run_name:
622
+ desc: null
623
+ value: ./
624
+ save_on_each_node:
625
+ desc: null
626
+ value: false
627
+ save_steps:
628
+ desc: null
629
+ value: 500
630
+ save_strategy:
631
+ desc: null
632
+ value: steps
633
+ save_total_limit:
634
+ desc: null
635
+ value: 1
636
+ seed:
637
+ desc: null
638
+ value: 42
639
+ sep_token_id:
640
+ desc: null
641
+ value: null
642
+ sharded_ddp:
643
+ desc: null
644
+ value: '[]'
645
+ skip_memory_metrics:
646
+ desc: null
647
+ value: true
648
+ sortish_sampler:
649
+ desc: null
650
+ value: false
651
+ task_specific_params:
652
+ desc: null
653
+ value: null
654
+ temperature:
655
+ desc: null
656
+ value: 1.0
657
+ tf32:
658
+ desc: null
659
+ value: None
660
+ tie_encoder_decoder:
661
+ desc: null
662
+ value: false
663
+ tie_word_embeddings:
664
+ desc: null
665
+ value: false
666
+ tokenizer_class:
667
+ desc: null
668
+ value: null
669
+ top_k:
670
+ desc: null
671
+ value: 50
672
+ top_p:
673
+ desc: null
674
+ value: 1.0
675
+ torch_dtype:
676
+ desc: null
677
+ value: torch.float32
678
+ torchscript:
679
+ desc: null
680
+ value: false
681
+ tpu_metrics_debug:
682
+ desc: null
683
+ value: false
684
+ tpu_num_cores:
685
+ desc: null
686
+ value: None
687
+ train_batch_size:
688
+ desc: null
689
+ value: 16
690
+ transformers_version:
691
+ desc: null
692
+ value: null
693
+ use_bfloat16:
694
+ desc: null
695
+ value: false
696
+ use_cache:
697
+ desc: null
698
+ value: false
699
+ use_legacy_prediction_loop:
700
+ desc: null
701
+ value: false
702
+ warmup_ratio:
703
+ desc: null
704
+ value: 0.0
705
+ warmup_steps:
706
+ desc: null
707
+ value: 500
708
+ weight_decay:
709
+ desc: null
710
+ value: 0.0
711
+ xpu_backend:
712
+ desc: null
713
+ value: None
wandb/run-20220303_004039-1a18zsso/files/output.log ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ 0%| | 0/892 [00:00<?, ?it/s][WARNING|modeling_utils.py:388] 2022-03-03 00:40:45,319 >> Could not estimate the number of tokens of the input, floating-point operations will not be computed
4
+ Traceback (most recent call last):
5
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 539, in <module>
6
+ main()
7
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 491, in main
8
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
9
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1384, in train
10
+ tr_loss_step = self.training_step(model, inputs)
11
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1959, in training_step
12
+ loss = self.compute_loss(model, inputs)
13
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1991, in compute_loss
14
+ outputs = model(**inputs)
15
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
16
+ return forward_call(*input, **kwargs)
17
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py", line 503, in forward
18
+ encoder_outputs = self.encoder(
19
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
20
+ return forward_call(*input, **kwargs)
21
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1346, in forward
22
+ extract_features = self.feature_extractor(input_values)
23
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
24
+ return forward_call(*input, **kwargs)
25
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 514, in forward
26
+ hidden_states = conv_layer(hidden_states)
27
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
28
+ return forward_call(*input, **kwargs)
29
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 389, in forward
30
+ hidden_states = self.layer_norm(hidden_states)
31
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
32
+ return forward_call(*input, **kwargs)
33
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/normalization.py", line 189, in forward
34
+ return F.layer_norm(
35
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/functional.py", line 2347, in layer_norm
36
+ return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
37
+ RuntimeError: CUDA out of memory. Tried to allocate 1.63 GiB (GPU 0; 15.78 GiB total capacity; 10.07 GiB already allocated; 707.31 MiB free; 13.39 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
wandb/run-20220303_004039-1a18zsso/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220303_004039-1a18zsso/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-03T00:40:40.798823",
5
+ "startedAt": "2022-03-03T00:40:39.514688",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=16",
24
+ "--per_device_eval_batch_size=16",
25
+ "--gradient_accumulation_steps=2",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "8c7181143c175387040dc1a6ac2ddbc9179b550c"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220303_004039-1a18zsso/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
1
+ {"_wandb": {"runtime": 9}}
wandb/run-20220303_004039-1a18zsso/logs/debug-internal.log ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-03 00:40:40,587 INFO MainThread:267195 [internal.py:wandb_internal():89] W&B internal server running at pid: 267195, started at: 2022-03-03 00:40:40.586996
2
+ 2022-03-03 00:40:40,590 INFO WriterThread:267195 [datastore.py:open_for_write():77] open: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/run-1a18zsso.wandb
3
+ 2022-03-03 00:40:40,590 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: check_version
4
+ 2022-03-03 00:40:40,591 DEBUG SenderThread:267195 [sender.py:send():235] send: header
5
+ 2022-03-03 00:40:40,591 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: check_version
6
+ 2022-03-03 00:40:40,668 DEBUG SenderThread:267195 [sender.py:send():235] send: run
7
+ 2022-03-03 00:40:40,791 INFO SenderThread:267195 [dir_watcher.py:__init__():169] watching files in: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files
8
+ 2022-03-03 00:40:40,791 INFO SenderThread:267195 [sender.py:_start_run_threads():809] run started: 1a18zsso with start time 1646268039
9
+ 2022-03-03 00:40:40,791 DEBUG SenderThread:267195 [sender.py:send():235] send: summary
10
+ 2022-03-03 00:40:40,791 INFO SenderThread:267195 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
11
+ 2022-03-03 00:40:40,793 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: run_start
12
+ 2022-03-03 00:40:40,798 DEBUG HandlerThread:267195 [meta.py:__init__():36] meta init
13
+ 2022-03-03 00:40:40,798 DEBUG HandlerThread:267195 [meta.py:__init__():50] meta init done
14
+ 2022-03-03 00:40:40,798 DEBUG HandlerThread:267195 [meta.py:probe():210] probe
15
+ 2022-03-03 00:40:40,805 DEBUG HandlerThread:267195 [meta.py:_setup_git():200] setup git
16
+ 2022-03-03 00:40:40,821 DEBUG HandlerThread:267195 [meta.py:_setup_git():207] setup git done
17
+ 2022-03-03 00:40:40,822 DEBUG HandlerThread:267195 [meta.py:_save_pip():54] save pip
18
+ 2022-03-03 00:40:40,823 DEBUG HandlerThread:267195 [meta.py:_save_pip():68] save pip done
19
+ 2022-03-03 00:40:40,823 DEBUG HandlerThread:267195 [meta.py:probe():248] probe done
20
+ 2022-03-03 00:40:40,946 DEBUG SenderThread:267195 [sender.py:send():235] send: files
21
+ 2022-03-03 00:40:40,946 INFO SenderThread:267195 [sender.py:_save_file():944] saving file wandb-metadata.json with policy now
22
+ 2022-03-03 00:40:40,951 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: stop_status
23
+ 2022-03-03 00:40:40,952 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: stop_status
24
+ 2022-03-03 00:40:40,999 DEBUG SenderThread:267195 [sender.py:send():235] send: config
25
+ 2022-03-03 00:40:41,000 DEBUG SenderThread:267195 [sender.py:send():235] send: metric
26
+ 2022-03-03 00:40:41,000 DEBUG SenderThread:267195 [sender.py:send():235] send: metric
27
+ 2022-03-03 00:40:41,000 WARNING SenderThread:267195 [sender.py:send_metric():902] Seen metric with glob (shouldnt happen)
28
+ 2022-03-03 00:40:41,240 INFO Thread-11 :267195 [upload_job.py:push():137] Uploaded file /tmp/tmp9mhkg11ywandb/2vus7xrr-wandb-metadata.json
29
+ 2022-03-03 00:40:41,793 INFO Thread-8 :267195 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/output.log
30
+ 2022-03-03 00:40:41,793 INFO Thread-8 :267195 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/wandb-metadata.json
31
+ 2022-03-03 00:40:41,794 INFO Thread-8 :267195 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/requirements.txt
32
+ 2022-03-03 00:40:41,794 INFO Thread-8 :267195 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/wandb-summary.json
33
+ 2022-03-03 00:40:43,792 INFO Thread-8 :267195 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/output.log
34
+ 2022-03-03 00:40:47,793 INFO Thread-8 :267195 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/output.log
35
+ 2022-03-03 00:40:49,794 INFO Thread-8 :267195 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/output.log
36
+ 2022-03-03 00:40:49,872 DEBUG SenderThread:267195 [sender.py:send():235] send: telemetry
37
+ 2022-03-03 00:40:49,872 DEBUG SenderThread:267195 [sender.py:send():235] send: exit
38
+ 2022-03-03 00:40:49,872 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
39
+ 2022-03-03 00:40:49,873 INFO SenderThread:267195 [sender.py:send_exit():371] handling exit code: 1
40
+ 2022-03-03 00:40:49,873 INFO SenderThread:267195 [sender.py:send_exit():373] handling runtime: 9
41
+ 2022-03-03 00:40:49,873 INFO SenderThread:267195 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
42
+ 2022-03-03 00:40:49,873 INFO SenderThread:267195 [sender.py:send_exit():379] send defer
43
+ 2022-03-03 00:40:49,874 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
44
+ 2022-03-03 00:40:49,874 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
45
+ 2022-03-03 00:40:49,875 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 0
46
+ 2022-03-03 00:40:49,875 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
47
+ 2022-03-03 00:40:49,875 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 0
48
+ 2022-03-03 00:40:49,875 INFO SenderThread:267195 [sender.py:transition_state():392] send defer: 1
49
+ 2022-03-03 00:40:49,875 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
50
+ 2022-03-03 00:40:49,876 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 1
51
+ 2022-03-03 00:40:49,994 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
52
+ 2022-03-03 00:40:49,994 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
53
+ 2022-03-03 00:40:49,994 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 1
54
+ 2022-03-03 00:40:49,994 INFO SenderThread:267195 [sender.py:transition_state():392] send defer: 2
55
+ 2022-03-03 00:40:49,995 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
56
+ 2022-03-03 00:40:49,995 DEBUG SenderThread:267195 [sender.py:send():235] send: stats
57
+ 2022-03-03 00:40:49,996 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
58
+ 2022-03-03 00:40:49,996 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 2
59
+ 2022-03-03 00:40:49,996 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
60
+ 2022-03-03 00:40:49,996 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 2
61
+ 2022-03-03 00:40:49,996 INFO SenderThread:267195 [sender.py:transition_state():392] send defer: 3
62
+ 2022-03-03 00:40:49,996 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
63
+ 2022-03-03 00:40:49,997 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 3
64
+ 2022-03-03 00:40:49,997 DEBUG SenderThread:267195 [sender.py:send():235] send: summary
65
+ 2022-03-03 00:40:49,998 INFO SenderThread:267195 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
66
+ 2022-03-03 00:40:49,998 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
67
+ 2022-03-03 00:40:49,998 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 3
68
+ 2022-03-03 00:40:49,998 INFO SenderThread:267195 [sender.py:transition_state():392] send defer: 4
69
+ 2022-03-03 00:40:49,998 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
70
+ 2022-03-03 00:40:49,998 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 4
71
+ 2022-03-03 00:40:49,998 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
72
+ 2022-03-03 00:40:49,998 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 4
73
+ 2022-03-03 00:40:50,102 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
74
+ 2022-03-03 00:40:50,128 INFO SenderThread:267195 [sender.py:transition_state():392] send defer: 5
75
+ 2022-03-03 00:40:50,128 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
76
+ 2022-03-03 00:40:50,129 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
77
+ 2022-03-03 00:40:50,129 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 5
78
+ 2022-03-03 00:40:50,129 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
79
+ 2022-03-03 00:40:50,130 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 5
80
+ 2022-03-03 00:40:50,130 INFO SenderThread:267195 [dir_watcher.py:finish():283] shutting down directory watcher
81
+ 2022-03-03 00:40:50,231 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
82
+ 2022-03-03 00:40:50,795 INFO Thread-8 :267195 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/output.log
83
+ 2022-03-03 00:40:50,795 INFO SenderThread:267195 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/wandb-summary.json
84
+ 2022-03-03 00:40:50,796 INFO SenderThread:267195 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/config.yaml
85
+ 2022-03-03 00:40:50,796 INFO SenderThread:267195 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files
86
+ 2022-03-03 00:40:50,796 INFO SenderThread:267195 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/wandb-metadata.json wandb-metadata.json
87
+ 2022-03-03 00:40:50,796 INFO SenderThread:267195 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/output.log output.log
88
+ 2022-03-03 00:40:50,797 INFO SenderThread:267195 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/wandb-summary.json wandb-summary.json
89
+ 2022-03-03 00:40:50,800 INFO SenderThread:267195 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/requirements.txt requirements.txt
90
+ 2022-03-03 00:40:50,803 INFO SenderThread:267195 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/config.yaml config.yaml
91
+ 2022-03-03 00:40:50,807 INFO SenderThread:267195 [sender.py:transition_state():392] send defer: 6
92
+ 2022-03-03 00:40:50,807 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
93
+ 2022-03-03 00:40:50,808 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
94
+ 2022-03-03 00:40:50,809 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 6
95
+ 2022-03-03 00:40:50,812 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
96
+ 2022-03-03 00:40:50,813 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 6
97
+ 2022-03-03 00:40:50,813 INFO SenderThread:267195 [file_pusher.py:finish():177] shutting down file pusher
98
+ 2022-03-03 00:40:50,910 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
99
+ 2022-03-03 00:40:50,911 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
100
+ 2022-03-03 00:40:51,013 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
101
+ 2022-03-03 00:40:51,013 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
102
+ 2022-03-03 00:40:51,068 INFO Thread-15 :267195 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/config.yaml
103
+ 2022-03-03 00:40:51,071 INFO Thread-13 :267195 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/wandb-summary.json
104
+ 2022-03-03 00:40:51,116 INFO Thread-12 :267195 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/output.log
105
+ 2022-03-03 00:40:51,117 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
106
+ 2022-03-03 00:40:51,117 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
107
+ 2022-03-03 00:40:51,172 INFO Thread-14 :267195 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/files/requirements.txt
108
+ 2022-03-03 00:40:51,219 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
109
+ 2022-03-03 00:40:51,219 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
110
+ 2022-03-03 00:40:51,321 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
111
+ 2022-03-03 00:40:51,322 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
112
+ 2022-03-03 00:40:51,372 INFO Thread-7 :267195 [sender.py:transition_state():392] send defer: 7
113
+ 2022-03-03 00:40:51,373 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
114
+ 2022-03-03 00:40:51,373 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 7
115
+ 2022-03-03 00:40:51,373 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
116
+ 2022-03-03 00:40:51,373 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 7
117
+ 2022-03-03 00:40:51,424 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
118
+ 2022-03-03 00:40:51,429 INFO SenderThread:267195 [sender.py:transition_state():392] send defer: 8
119
+ 2022-03-03 00:40:51,429 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
120
+ 2022-03-03 00:40:51,430 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
121
+ 2022-03-03 00:40:51,430 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 8
122
+ 2022-03-03 00:40:51,431 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
123
+ 2022-03-03 00:40:51,431 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 8
124
+ 2022-03-03 00:40:51,431 INFO SenderThread:267195 [sender.py:transition_state():392] send defer: 9
125
+ 2022-03-03 00:40:51,431 DEBUG SenderThread:267195 [sender.py:send():235] send: final
126
+ 2022-03-03 00:40:51,432 DEBUG SenderThread:267195 [sender.py:send():235] send: footer
127
+ 2022-03-03 00:40:51,432 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: defer
128
+ 2022-03-03 00:40:51,432 INFO HandlerThread:267195 [handler.py:handle_request_defer():154] handle defer: 9
129
+ 2022-03-03 00:40:51,432 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: defer
130
+ 2022-03-03 00:40:51,433 INFO SenderThread:267195 [sender.py:send_request_defer():388] handle sender defer: 9
131
+ 2022-03-03 00:40:51,531 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: poll_exit
132
+ 2022-03-03 00:40:51,532 DEBUG SenderThread:267195 [sender.py:send_request():249] send_request: poll_exit
133
+ 2022-03-03 00:40:51,532 INFO SenderThread:267195 [file_pusher.py:join():182] waiting for file pusher
134
+ 2022-03-03 00:40:51,631 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: get_summary
135
+ 2022-03-03 00:40:51,632 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: sampled_history
136
+ 2022-03-03 00:40:51,633 DEBUG HandlerThread:267195 [handler.py:handle_request():131] handle_request: shutdown
137
+ 2022-03-03 00:40:51,633 INFO HandlerThread:267195 [handler.py:finish():739] shutting down handler
138
+ 2022-03-03 00:40:52,432 INFO WriterThread:267195 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/run-1a18zsso.wandb
139
+ 2022-03-03 00:40:52,630 INFO SenderThread:267195 [sender.py:finish():1075] shutting down sender
140
+ 2022-03-03 00:40:52,630 INFO SenderThread:267195 [file_pusher.py:finish():177] shutting down file pusher
141
+ 2022-03-03 00:40:52,630 INFO SenderThread:267195 [file_pusher.py:join():182] waiting for file pusher
142
+ 2022-03-03 00:40:52,632 INFO MainThread:267195 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220303_004039-1a18zsso/logs/debug.log ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-03 00:40:39,518 INFO MainThread:267096 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-03 00:40:39,518 INFO MainThread:267096 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-03 00:40:39,518 INFO MainThread:267096 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-03 00:40:39,518 INFO MainThread:267096 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-03 00:40:39,518 INFO MainThread:267096 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/logs/debug.log
6
+ 2022-03-03 00:40:39,518 INFO MainThread:267096 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004039-1a18zsso/logs/debug-internal.log
7
+ 2022-03-03 00:40:39,518 INFO MainThread:267096 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-03 00:40:39,518 INFO MainThread:267096 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-03 00:40:39,519 INFO MainThread:267096 [wandb_init.py:init():471] starting backend
11
+ 2022-03-03 00:40:39,519 INFO MainThread:267096 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-03 00:40:39,592 INFO MainThread:267096 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-03 00:40:39,663 INFO MainThread:267096 [backend.py:ensure_launched():224] started backend process with pid: 267195
14
+ 2022-03-03 00:40:39,665 INFO MainThread:267096 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-03 00:40:39,677 INFO MainThread:267096 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-03 00:40:39,846 INFO MainThread:267096 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-03 00:40:40,667 INFO MainThread:267096 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.11 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
18
+
19
+ 2022-03-03 00:40:40,667 INFO MainThread:267096 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
20
+ 2022-03-03 00:40:40,792 INFO MainThread:267096 [wandb_init.py:init():624] starting run threads in backend
21
+ 2022-03-03 00:40:40,950 INFO MainThread:267096 [wandb_run.py:_console_start():1827] atexit reg
22
+ 2022-03-03 00:40:40,951 INFO MainThread:267096 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
23
+ 2022-03-03 00:40:40,952 INFO MainThread:267096 [wandb_run.py:_redirect():1706] Redirecting console.
24
+ 2022-03-03 00:40:40,954 INFO MainThread:267096 [wandb_run.py:_redirect():1762] Redirects installed.
25
+ 2022-03-03 00:40:40,954 INFO MainThread:267096 [wandb_init.py:init():651] run started, returning control to user process
26
+ 2022-03-03 00:40:40,956 INFO MainThread:267096 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0003, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar03_00-39-53_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 16, 'eval_batch_size': 16}
27
+ 2022-03-03 00:40:40,960 INFO MainThread:267096 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-03 00:40:47,275 INFO MainThread:267096 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-03 00:40:47,278 INFO MainThread:267096 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-03 00:40:49,875 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-03 00:40:49,996 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-03 00:40:50,129 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 1
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 2095
52
+ }
53
+
54
+ 2022-03-03 00:40:50,809 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 5
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 2095
59
+ total_bytes: 21066
60
+ }
61
+
62
+ 2022-03-03 00:40:50,912 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 17829
67
+ total_bytes: 21066
68
+ }
69
+
70
+ 2022-03-03 00:40:51,015 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 21066
75
+ total_bytes: 21066
76
+ }
77
+
78
+ 2022-03-03 00:40:51,118 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 21066
83
+ total_bytes: 21066
84
+ }
85
+
86
+ 2022-03-03 00:40:51,220 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 21066
91
+ total_bytes: 21066
92
+ }
93
+
94
+ 2022-03-03 00:40:51,323 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
95
+ wandb_count: 5
96
+ }
97
+ pusher_stats {
98
+ uploaded_bytes: 21066
99
+ total_bytes: 21066
100
+ }
101
+
102
+ 2022-03-03 00:40:51,430 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
103
+ wandb_count: 5
104
+ }
105
+ pusher_stats {
106
+ uploaded_bytes: 21066
107
+ total_bytes: 21066
108
+ }
109
+
110
+ 2022-03-03 00:40:51,630 INFO MainThread:267096 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
111
+ exit_result {
112
+ }
113
+ file_counts {
114
+ wandb_count: 5
115
+ }
116
+ pusher_stats {
117
+ uploaded_bytes: 21066
118
+ total_bytes: 21066
119
+ }
120
+ local_info {
121
+ }
122
+
123
+ 2022-03-03 00:40:52,787 INFO MainThread:267096 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220303_004039-1a18zsso/run-1a18zsso.wandb ADDED
Binary file (13.7 kB). View file
wandb/run-20220303_004215-2ttcbe3s/files/config.yaml ADDED
@@ -0,0 +1,713 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: ./
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.10
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.9.5
22
+ start_time: 1646268136
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.9.5
37
+ 5: 0.12.10
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ adafactor:
42
+ desc: null
43
+ value: false
44
+ adam_beta1:
45
+ desc: null
46
+ value: 0.9
47
+ adam_beta2:
48
+ desc: null
49
+ value: 0.999
50
+ adam_epsilon:
51
+ desc: null
52
+ value: 1.0e-08
53
+ add_cross_attention:
54
+ desc: null
55
+ value: false
56
+ architectures:
57
+ desc: null
58
+ value:
59
+ - SpeechEncoderDecoderModel
60
+ bad_words_ids:
61
+ desc: null
62
+ value: null
63
+ bf16:
64
+ desc: null
65
+ value: false
66
+ bf16_full_eval:
67
+ desc: null
68
+ value: false
69
+ bos_token_id:
70
+ desc: null
71
+ value: null
72
+ chunk_size_feed_forward:
73
+ desc: null
74
+ value: 0
75
+ cross_attention_hidden_size:
76
+ desc: null
77
+ value: null
78
+ dataloader_drop_last:
79
+ desc: null
80
+ value: false
81
+ dataloader_num_workers:
82
+ desc: null
83
+ value: 0
84
+ dataloader_pin_memory:
85
+ desc: null
86
+ value: true
87
+ ddp_bucket_cap_mb:
88
+ desc: null
89
+ value: None
90
+ ddp_find_unused_parameters:
91
+ desc: null
92
+ value: None
93
+ debug:
94
+ desc: null
95
+ value: '[]'
96
+ decoder:
97
+ desc: null
98
+ value:
99
+ _name_or_path: gpt2-medium
100
+ activation_function: gelu_new
101
+ add_cross_attention: true
102
+ architectures:
103
+ - GPT2LMHeadModel
104
+ attn_pdrop: 0.0
105
+ bad_words_ids: null
106
+ bos_token_id: 50256
107
+ chunk_size_feed_forward: 0
108
+ cross_attention_hidden_size: null
109
+ decoder_start_token_id: null
110
+ diversity_penalty: 0.0
111
+ do_sample: false
112
+ early_stopping: false
113
+ embd_pdrop: 0.0
114
+ encoder_no_repeat_ngram_size: 0
115
+ eos_token_id: 50256
116
+ finetuning_task: null
117
+ forced_bos_token_id: null
118
+ forced_eos_token_id: null
119
+ id2label:
120
+ '0': LABEL_0
121
+ '1': LABEL_1
122
+ initializer_range: 0.02
123
+ is_decoder: true
124
+ is_encoder_decoder: false
125
+ label2id:
126
+ LABEL_0: 0
127
+ LABEL_1: 1
128
+ layer_norm_epsilon: 1.0e-05
129
+ length_penalty: 1.0
130
+ max_length: 20
131
+ min_length: 0
132
+ model_type: gpt2
133
+ n_ctx: 1024
134
+ n_embd: 1024
135
+ n_head: 16
136
+ n_inner: null
137
+ n_layer: 24
138
+ n_positions: 1024
139
+ n_special: 0
140
+ no_repeat_ngram_size: 0
141
+ num_beam_groups: 1
142
+ num_beams: 1
143
+ num_return_sequences: 1
144
+ output_attentions: false
145
+ output_hidden_states: false
146
+ output_scores: false
147
+ pad_token_id: null
148
+ predict_special_tokens: true
149
+ prefix: null
150
+ problem_type: null
151
+ pruned_heads: {}
152
+ remove_invalid_values: false
153
+ reorder_and_upcast_attn: false
154
+ repetition_penalty: 1.0
155
+ resid_pdrop: 0.0
156
+ return_dict: true
157
+ return_dict_in_generate: false
158
+ scale_attn_by_inverse_layer_idx: false
159
+ scale_attn_weights: true
160
+ sep_token_id: null
161
+ summary_activation: null
162
+ summary_first_dropout: 0.0
163
+ summary_proj_to_labels: true
164
+ summary_type: cls_index
165
+ summary_use_proj: true
166
+ task_specific_params:
167
+ text-generation:
168
+ do_sample: true
169
+ max_length: 50
170
+ temperature: 1.0
171
+ tie_encoder_decoder: false
172
+ tie_word_embeddings: true
173
+ tokenizer_class: null
174
+ top_k: 50
175
+ top_p: 1.0
176
+ torch_dtype: null
177
+ torchscript: false
178
+ transformers_version: 4.17.0.dev0
179
+ use_bfloat16: false
180
+ use_cache: false
181
+ vocab_size: 50257
182
+ decoder_start_token_id:
183
+ desc: null
184
+ value: 50256
185
+ deepspeed:
186
+ desc: null
187
+ value: None
188
+ disable_tqdm:
189
+ desc: null
190
+ value: false
191
+ diversity_penalty:
192
+ desc: null
193
+ value: 0.0
194
+ do_eval:
195
+ desc: null
196
+ value: true
197
+ do_predict:
198
+ desc: null
199
+ value: false
200
+ do_sample:
201
+ desc: null
202
+ value: false
203
+ do_train:
204
+ desc: null
205
+ value: true
206
+ early_stopping:
207
+ desc: null
208
+ value: false
209
+ encoder:
210
+ desc: null
211
+ value:
212
+ _name_or_path: facebook/wav2vec2-large-lv60
213
+ activation_dropout: 0.0
214
+ adapter_kernel_size: 3
215
+ adapter_stride: 2
216
+ add_adapter: true
217
+ add_cross_attention: false
218
+ apply_spec_augment: false
219
+ architectures:
220
+ - Wav2Vec2ForPreTraining
221
+ attention_dropout: 0.0
222
+ bad_words_ids: null
223
+ bos_token_id: 1
224
+ chunk_size_feed_forward: 0
225
+ classifier_proj_size: 256
226
+ codevector_dim: 768
227
+ contrastive_logits_temperature: 0.1
228
+ conv_bias: true
229
+ conv_dim:
230
+ - 512
231
+ - 512
232
+ - 512
233
+ - 512
234
+ - 512
235
+ - 512
236
+ - 512
237
+ conv_kernel:
238
+ - 10
239
+ - 3
240
+ - 3
241
+ - 3
242
+ - 3
243
+ - 2
244
+ - 2
245
+ conv_stride:
246
+ - 5
247
+ - 2
248
+ - 2
249
+ - 2
250
+ - 2
251
+ - 2
252
+ - 2
253
+ cross_attention_hidden_size: null
254
+ ctc_loss_reduction: sum
255
+ ctc_zero_infinity: false
256
+ decoder_start_token_id: null
257
+ diversity_loss_weight: 0.1
258
+ diversity_penalty: 0.0
259
+ do_sample: false
260
+ do_stable_layer_norm: true
261
+ early_stopping: false
262
+ encoder_no_repeat_ngram_size: 0
263
+ eos_token_id: 2
264
+ feat_extract_activation: gelu
265
+ feat_extract_dropout: 0.0
266
+ feat_extract_norm: layer
267
+ feat_proj_dropout: 0.0
268
+ feat_quantizer_dropout: 0.0
269
+ final_dropout: 0.0
270
+ finetuning_task: null
271
+ forced_bos_token_id: null
272
+ forced_eos_token_id: null
273
+ gradient_checkpointing: false
274
+ hidden_act: gelu
275
+ hidden_dropout: 0.0
276
+ hidden_dropout_prob: 0.0
277
+ hidden_size: 1024
278
+ id2label:
279
+ '0': LABEL_0
280
+ '1': LABEL_1
281
+ initializer_range: 0.02
282
+ intermediate_size: 4096
283
+ is_decoder: false
284
+ is_encoder_decoder: false
285
+ label2id:
286
+ LABEL_0: 0
287
+ LABEL_1: 1
288
+ layer_norm_eps: 1.0e-05
289
+ layerdrop: 0.0
290
+ length_penalty: 1.0
291
+ mask_feature_length: 10
292
+ mask_feature_min_masks: 0
293
+ mask_feature_prob: 0.0
294
+ mask_time_length: 10
295
+ mask_time_min_masks: 2
296
+ mask_time_prob: 0.0
297
+ max_length: 20
298
+ min_length: 0
299
+ model_type: wav2vec2
300
+ no_repeat_ngram_size: 0
301
+ num_adapter_layers: 3
302
+ num_attention_heads: 16
303
+ num_beam_groups: 1
304
+ num_beams: 1
305
+ num_codevector_groups: 2
306
+ num_codevectors_per_group: 320
307
+ num_conv_pos_embedding_groups: 16
308
+ num_conv_pos_embeddings: 128
309
+ num_feat_extract_layers: 7
310
+ num_hidden_layers: 24
311
+ num_negatives: 100
312
+ num_return_sequences: 1
313
+ output_attentions: false
314
+ output_hidden_size: 1024
315
+ output_hidden_states: false
316
+ output_scores: false
317
+ pad_token_id: 0
318
+ prefix: null
319
+ problem_type: null
320
+ proj_codevector_dim: 768
321
+ pruned_heads: {}
322
+ remove_invalid_values: false
323
+ repetition_penalty: 1.0
324
+ return_dict: true
325
+ return_dict_in_generate: false
326
+ sep_token_id: null
327
+ task_specific_params: null
328
+ tdnn_dilation:
329
+ - 1
330
+ - 2
331
+ - 3
332
+ - 1
333
+ - 1
334
+ tdnn_dim:
335
+ - 512
336
+ - 512
337
+ - 512
338
+ - 512
339
+ - 1500
340
+ tdnn_kernel:
341
+ - 5
342
+ - 3
343
+ - 3
344
+ - 1
345
+ - 1
346
+ temperature: 1.0
347
+ tie_encoder_decoder: false
348
+ tie_word_embeddings: true
349
+ tokenizer_class: null
350
+ top_k: 50
351
+ top_p: 1.0
352
+ torch_dtype: null
353
+ torchscript: false
354
+ transformers_version: 4.17.0.dev0
355
+ use_bfloat16: false
356
+ use_weighted_layer_sum: false
357
+ vocab_size: 32
358
+ xvector_output_dim: 512
359
+ encoder_no_repeat_ngram_size:
360
+ desc: null
361
+ value: 0
362
+ eos_token_id:
363
+ desc: null
364
+ value: 50256
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: None
368
+ eval_batch_size:
369
+ desc: null
370
+ value: 16
371
+ eval_steps:
372
+ desc: null
373
+ value: 500
374
+ evaluation_strategy:
375
+ desc: null
376
+ value: steps
377
+ finetuning_task:
378
+ desc: null
379
+ value: null
380
+ forced_bos_token_id:
381
+ desc: null
382
+ value: null
383
+ forced_eos_token_id:
384
+ desc: null
385
+ value: null
386
+ fp16:
387
+ desc: null
388
+ value: true
389
+ fp16_backend:
390
+ desc: null
391
+ value: auto
392
+ fp16_full_eval:
393
+ desc: null
394
+ value: false
395
+ fp16_opt_level:
396
+ desc: null
397
+ value: O1
398
+ generation_max_length:
399
+ desc: null
400
+ value: 40
401
+ generation_num_beams:
402
+ desc: null
403
+ value: 1
404
+ gradient_accumulation_steps:
405
+ desc: null
406
+ value: 4
407
+ gradient_checkpointing:
408
+ desc: null
409
+ value: true
410
+ greater_is_better:
411
+ desc: null
412
+ value: None
413
+ group_by_length:
414
+ desc: null
415
+ value: true
416
+ half_precision_backend:
417
+ desc: null
418
+ value: amp
419
+ hub_model_id:
420
+ desc: null
421
+ value: None
422
+ hub_strategy:
423
+ desc: null
424
+ value: every_save
425
+ hub_token:
426
+ desc: null
427
+ value: <HUB_TOKEN>
428
+ id2label:
429
+ desc: null
430
+ value:
431
+ '0': LABEL_0
432
+ '1': LABEL_1
433
+ ignore_data_skip:
434
+ desc: null
435
+ value: false
436
+ is_decoder:
437
+ desc: null
438
+ value: false
439
+ is_encoder_decoder:
440
+ desc: null
441
+ value: true
442
+ label2id:
443
+ desc: null
444
+ value:
445
+ LABEL_0: 0
446
+ LABEL_1: 1
447
+ label_names:
448
+ desc: null
449
+ value: None
450
+ label_smoothing_factor:
451
+ desc: null
452
+ value: 0.0
453
+ learning_rate:
454
+ desc: null
455
+ value: 0.0003
456
+ length_column_name:
457
+ desc: null
458
+ value: input_length
459
+ length_penalty:
460
+ desc: null
461
+ value: 1.0
462
+ load_best_model_at_end:
463
+ desc: null
464
+ value: false
465
+ local_rank:
466
+ desc: null
467
+ value: -1
468
+ log_level:
469
+ desc: null
470
+ value: -1
471
+ log_level_replica:
472
+ desc: null
473
+ value: -1
474
+ log_on_each_node:
475
+ desc: null
476
+ value: true
477
+ logging_dir:
478
+ desc: null
479
+ value: ./runs/Mar03_00-41-32_sanchit--v100
480
+ logging_first_step:
481
+ desc: null
482
+ value: false
483
+ logging_nan_inf_filter:
484
+ desc: null
485
+ value: true
486
+ logging_steps:
487
+ desc: null
488
+ value: 1
489
+ logging_strategy:
490
+ desc: null
491
+ value: steps
492
+ lr_scheduler_type:
493
+ desc: null
494
+ value: linear
495
+ max_grad_norm:
496
+ desc: null
497
+ value: 1.0
498
+ max_length:
499
+ desc: null
500
+ value: 50
501
+ max_steps:
502
+ desc: null
503
+ value: -1
504
+ metric_for_best_model:
505
+ desc: null
506
+ value: None
507
+ min_length:
508
+ desc: null
509
+ value: 0
510
+ model_type:
511
+ desc: null
512
+ value: speech-encoder-decoder
513
+ mp_parameters:
514
+ desc: null
515
+ value: ''
516
+ no_cuda:
517
+ desc: null
518
+ value: false
519
+ no_repeat_ngram_size:
520
+ desc: null
521
+ value: 0
522
+ num_beam_groups:
523
+ desc: null
524
+ value: 1
525
+ num_beams:
526
+ desc: null
527
+ value: 1
528
+ num_return_sequences:
529
+ desc: null
530
+ value: 1
531
+ num_train_epochs:
532
+ desc: null
533
+ value: 1.0
534
+ optim:
535
+ desc: null
536
+ value: adamw_hf
537
+ output_attentions:
538
+ desc: null
539
+ value: false
540
+ output_dir:
541
+ desc: null
542
+ value: ./
543
+ output_hidden_states:
544
+ desc: null
545
+ value: false
546
+ output_scores:
547
+ desc: null
548
+ value: false
549
+ overwrite_output_dir:
550
+ desc: null
551
+ value: true
552
+ pad_token_id:
553
+ desc: null
554
+ value: 50256
555
+ past_index:
556
+ desc: null
557
+ value: -1
558
+ per_device_eval_batch_size:
559
+ desc: null
560
+ value: 16
561
+ per_device_train_batch_size:
562
+ desc: null
563
+ value: 16
564
+ per_gpu_eval_batch_size:
565
+ desc: null
566
+ value: None
567
+ per_gpu_train_batch_size:
568
+ desc: null
569
+ value: None
570
+ predict_with_generate:
571
+ desc: null
572
+ value: true
573
+ prediction_loss_only:
574
+ desc: null
575
+ value: false
576
+ prefix:
577
+ desc: null
578
+ value: null
579
+ problem_type:
580
+ desc: null
581
+ value: null
582
+ processor_class:
583
+ desc: null
584
+ value: Wav2Vec2Processor
585
+ pruned_heads:
586
+ desc: null
587
+ value: {}
588
+ push_to_hub:
589
+ desc: null
590
+ value: true
591
+ push_to_hub_model_id:
592
+ desc: null
593
+ value: None
594
+ push_to_hub_organization:
595
+ desc: null
596
+ value: None
597
+ push_to_hub_token:
598
+ desc: null
599
+ value: <PUSH_TO_HUB_TOKEN>
600
+ remove_invalid_values:
601
+ desc: null
602
+ value: false
603
+ remove_unused_columns:
604
+ desc: null
605
+ value: true
606
+ repetition_penalty:
607
+ desc: null
608
+ value: 1.0
609
+ report_to:
610
+ desc: null
611
+ value: '[''wandb'']'
612
+ resume_from_checkpoint:
613
+ desc: null
614
+ value: None
615
+ return_dict:
616
+ desc: null
617
+ value: true
618
+ return_dict_in_generate:
619
+ desc: null
620
+ value: false
621
+ run_name:
622
+ desc: null
623
+ value: ./
624
+ save_on_each_node:
625
+ desc: null
626
+ value: false
627
+ save_steps:
628
+ desc: null
629
+ value: 500
630
+ save_strategy:
631
+ desc: null
632
+ value: steps
633
+ save_total_limit:
634
+ desc: null
635
+ value: 1
636
+ seed:
637
+ desc: null
638
+ value: 42
639
+ sep_token_id:
640
+ desc: null
641
+ value: null
642
+ sharded_ddp:
643
+ desc: null
644
+ value: '[]'
645
+ skip_memory_metrics:
646
+ desc: null
647
+ value: true
648
+ sortish_sampler:
649
+ desc: null
650
+ value: false
651
+ task_specific_params:
652
+ desc: null
653
+ value: null
654
+ temperature:
655
+ desc: null
656
+ value: 1.0
657
+ tf32:
658
+ desc: null
659
+ value: None
660
+ tie_encoder_decoder:
661
+ desc: null
662
+ value: false
663
+ tie_word_embeddings:
664
+ desc: null
665
+ value: false
666
+ tokenizer_class:
667
+ desc: null
668
+ value: null
669
+ top_k:
670
+ desc: null
671
+ value: 50
672
+ top_p:
673
+ desc: null
674
+ value: 1.0
675
+ torch_dtype:
676
+ desc: null
677
+ value: torch.float32
678
+ torchscript:
679
+ desc: null
680
+ value: false
681
+ tpu_metrics_debug:
682
+ desc: null
683
+ value: false
684
+ tpu_num_cores:
685
+ desc: null
686
+ value: None
687
+ train_batch_size:
688
+ desc: null
689
+ value: 16
690
+ transformers_version:
691
+ desc: null
692
+ value: null
693
+ use_bfloat16:
694
+ desc: null
695
+ value: false
696
+ use_cache:
697
+ desc: null
698
+ value: false
699
+ use_legacy_prediction_loop:
700
+ desc: null
701
+ value: false
702
+ warmup_ratio:
703
+ desc: null
704
+ value: 0.0
705
+ warmup_steps:
706
+ desc: null
707
+ value: 500
708
+ weight_decay:
709
+ desc: null
710
+ value: 0.0
711
+ xpu_backend:
712
+ desc: null
713
+ value: None
wandb/run-20220303_004215-2ttcbe3s/files/output.log ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ 0%| | 0/446 [00:00<?, ?it/s][WARNING|modeling_utils.py:388] 2022-03-03 00:42:21,600 >> Could not estimate the number of tokens of the input, floating-point operations will not be computed
4
+ Traceback (most recent call last):
5
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 539, in <module>
6
+ main()
7
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 491, in main
8
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
9
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1384, in train
10
+ tr_loss_step = self.training_step(model, inputs)
11
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1959, in training_step
12
+ loss = self.compute_loss(model, inputs)
13
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1991, in compute_loss
14
+ outputs = model(**inputs)
15
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
16
+ return forward_call(*input, **kwargs)
17
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py", line 503, in forward
18
+ encoder_outputs = self.encoder(
19
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
20
+ return forward_call(*input, **kwargs)
21
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1346, in forward
22
+ extract_features = self.feature_extractor(input_values)
23
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
24
+ return forward_call(*input, **kwargs)
25
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 514, in forward
26
+ hidden_states = conv_layer(hidden_states)
27
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
28
+ return forward_call(*input, **kwargs)
29
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 389, in forward
30
+ hidden_states = self.layer_norm(hidden_states)
31
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
32
+ return forward_call(*input, **kwargs)
33
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/normalization.py", line 189, in forward
34
+ return F.layer_norm(
35
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/functional.py", line 2347, in layer_norm
36
+ return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
37
+ RuntimeError: CUDA out of memory. Tried to allocate 1.64 GiB (GPU 0; 15.78 GiB total capacity; 10.11 GiB already allocated; 707.31 MiB free; 13.39 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
wandb/run-20220303_004215-2ttcbe3s/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220303_004215-2ttcbe3s/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-03T00:42:17.150413",
5
+ "startedAt": "2022-03-03T00:42:15.987965",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=16",
24
+ "--per_device_eval_batch_size=16",
25
+ "--gradient_accumulation_steps=4",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-4",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "8c7181143c175387040dc1a6ac2ddbc9179b550c"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220303_004215-2ttcbe3s/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
1
+ {"_wandb": {"runtime": 8}}
wandb/run-20220303_004215-2ttcbe3s/logs/debug-internal.log ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-03 00:42:16,953 INFO MainThread:267379 [internal.py:wandb_internal():89] W&B internal server running at pid: 267379, started at: 2022-03-03 00:42:16.952980
2
+ 2022-03-03 00:42:16,956 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: check_version
3
+ 2022-03-03 00:42:16,956 INFO WriterThread:267379 [datastore.py:open_for_write():77] open: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/run-2ttcbe3s.wandb
4
+ 2022-03-03 00:42:16,957 DEBUG SenderThread:267379 [sender.py:send():235] send: header
5
+ 2022-03-03 00:42:16,957 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: check_version
6
+ 2022-03-03 00:42:17,024 DEBUG SenderThread:267379 [sender.py:send():235] send: run
7
+ 2022-03-03 00:42:17,144 INFO SenderThread:267379 [dir_watcher.py:__init__():169] watching files in: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files
8
+ 2022-03-03 00:42:17,144 INFO SenderThread:267379 [sender.py:_start_run_threads():809] run started: 2ttcbe3s with start time 1646268136
9
+ 2022-03-03 00:42:17,145 DEBUG SenderThread:267379 [sender.py:send():235] send: summary
10
+ 2022-03-03 00:42:17,145 INFO SenderThread:267379 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
11
+ 2022-03-03 00:42:17,145 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: run_start
12
+ 2022-03-03 00:42:17,150 DEBUG HandlerThread:267379 [meta.py:__init__():36] meta init
13
+ 2022-03-03 00:42:17,150 DEBUG HandlerThread:267379 [meta.py:__init__():50] meta init done
14
+ 2022-03-03 00:42:17,150 DEBUG HandlerThread:267379 [meta.py:probe():210] probe
15
+ 2022-03-03 00:42:17,156 DEBUG HandlerThread:267379 [meta.py:_setup_git():200] setup git
16
+ 2022-03-03 00:42:17,172 DEBUG HandlerThread:267379 [meta.py:_setup_git():207] setup git done
17
+ 2022-03-03 00:42:17,172 DEBUG HandlerThread:267379 [meta.py:_save_pip():54] save pip
18
+ 2022-03-03 00:42:17,173 DEBUG HandlerThread:267379 [meta.py:_save_pip():68] save pip done
19
+ 2022-03-03 00:42:17,173 DEBUG HandlerThread:267379 [meta.py:probe():248] probe done
20
+ 2022-03-03 00:42:17,254 DEBUG SenderThread:267379 [sender.py:send():235] send: files
21
+ 2022-03-03 00:42:17,254 INFO SenderThread:267379 [sender.py:_save_file():944] saving file wandb-metadata.json with policy now
22
+ 2022-03-03 00:42:17,259 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: stop_status
23
+ 2022-03-03 00:42:17,259 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: stop_status
24
+ 2022-03-03 00:42:17,295 DEBUG SenderThread:267379 [sender.py:send():235] send: config
25
+ 2022-03-03 00:42:17,296 DEBUG SenderThread:267379 [sender.py:send():235] send: metric
26
+ 2022-03-03 00:42:17,296 DEBUG SenderThread:267379 [sender.py:send():235] send: metric
27
+ 2022-03-03 00:42:17,296 WARNING SenderThread:267379 [sender.py:send_metric():902] Seen metric with glob (shouldnt happen)
28
+ 2022-03-03 00:42:17,551 INFO Thread-11 :267379 [upload_job.py:push():137] Uploaded file /tmp/tmpamllbtqswandb/2k16no24-wandb-metadata.json
29
+ 2022-03-03 00:42:18,148 INFO Thread-8 :267379 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/wandb-metadata.json
30
+ 2022-03-03 00:42:18,148 INFO Thread-8 :267379 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/wandb-summary.json
31
+ 2022-03-03 00:42:18,148 INFO Thread-8 :267379 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/requirements.txt
32
+ 2022-03-03 00:42:18,148 INFO Thread-8 :267379 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/output.log
33
+ 2022-03-03 00:42:20,146 INFO Thread-8 :267379 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/output.log
34
+ 2022-03-03 00:42:24,147 INFO Thread-8 :267379 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/output.log
35
+ 2022-03-03 00:42:26,115 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: poll_exit
36
+ 2022-03-03 00:42:26,115 DEBUG SenderThread:267379 [sender.py:send():235] send: telemetry
37
+ 2022-03-03 00:42:26,115 DEBUG SenderThread:267379 [sender.py:send():235] send: exit
38
+ 2022-03-03 00:42:26,116 INFO SenderThread:267379 [sender.py:send_exit():371] handling exit code: 1
39
+ 2022-03-03 00:42:26,116 INFO SenderThread:267379 [sender.py:send_exit():373] handling runtime: 8
40
+ 2022-03-03 00:42:26,116 INFO SenderThread:267379 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
41
+ 2022-03-03 00:42:26,116 INFO SenderThread:267379 [sender.py:send_exit():379] send defer
42
+ 2022-03-03 00:42:26,116 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: poll_exit
43
+ 2022-03-03 00:42:26,117 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
44
+ 2022-03-03 00:42:26,117 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 0
45
+ 2022-03-03 00:42:26,117 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
46
+ 2022-03-03 00:42:26,117 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 0
47
+ 2022-03-03 00:42:26,117 INFO SenderThread:267379 [sender.py:transition_state():392] send defer: 1
48
+ 2022-03-03 00:42:26,118 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
49
+ 2022-03-03 00:42:26,118 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 1
50
+ 2022-03-03 00:42:26,148 INFO Thread-8 :267379 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/wandb-summary.json
51
+ 2022-03-03 00:42:26,148 INFO Thread-8 :267379 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/output.log
52
+ 2022-03-03 00:42:26,201 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
53
+ 2022-03-03 00:42:26,201 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 1
54
+ 2022-03-03 00:42:26,201 INFO SenderThread:267379 [sender.py:transition_state():392] send defer: 2
55
+ 2022-03-03 00:42:26,202 DEBUG SenderThread:267379 [sender.py:send():235] send: stats
56
+ 2022-03-03 00:42:26,202 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
57
+ 2022-03-03 00:42:26,203 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 2
58
+ 2022-03-03 00:42:26,203 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
59
+ 2022-03-03 00:42:26,203 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 2
60
+ 2022-03-03 00:42:26,203 INFO SenderThread:267379 [sender.py:transition_state():392] send defer: 3
61
+ 2022-03-03 00:42:26,203 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
62
+ 2022-03-03 00:42:26,203 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 3
63
+ 2022-03-03 00:42:26,203 DEBUG SenderThread:267379 [sender.py:send():235] send: summary
64
+ 2022-03-03 00:42:26,204 INFO SenderThread:267379 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
65
+ 2022-03-03 00:42:26,204 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
66
+ 2022-03-03 00:42:26,204 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 3
67
+ 2022-03-03 00:42:26,204 INFO SenderThread:267379 [sender.py:transition_state():392] send defer: 4
68
+ 2022-03-03 00:42:26,204 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
69
+ 2022-03-03 00:42:26,205 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 4
70
+ 2022-03-03 00:42:26,205 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
71
+ 2022-03-03 00:42:26,205 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 4
72
+ 2022-03-03 00:42:26,218 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: poll_exit
73
+ 2022-03-03 00:42:26,313 INFO SenderThread:267379 [sender.py:transition_state():392] send defer: 5
74
+ 2022-03-03 00:42:26,313 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: poll_exit
75
+ 2022-03-03 00:42:26,314 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
76
+ 2022-03-03 00:42:26,314 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 5
77
+ 2022-03-03 00:42:26,314 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
78
+ 2022-03-03 00:42:26,315 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 5
79
+ 2022-03-03 00:42:26,315 INFO SenderThread:267379 [dir_watcher.py:finish():283] shutting down directory watcher
80
+ 2022-03-03 00:42:26,415 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: poll_exit
81
+ 2022-03-03 00:42:27,148 INFO Thread-8 :267379 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/wandb-summary.json
82
+ 2022-03-03 00:42:27,149 INFO SenderThread:267379 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/config.yaml
83
+ 2022-03-03 00:42:27,149 INFO SenderThread:267379 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files
84
+ 2022-03-03 00:42:27,150 INFO SenderThread:267379 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/wandb-metadata.json wandb-metadata.json
85
+ 2022-03-03 00:42:27,150 INFO SenderThread:267379 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/output.log output.log
86
+ 2022-03-03 00:42:27,150 INFO SenderThread:267379 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/wandb-summary.json wandb-summary.json
87
+ 2022-03-03 00:42:27,152 INFO SenderThread:267379 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/requirements.txt requirements.txt
88
+ 2022-03-03 00:42:27,156 INFO SenderThread:267379 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/config.yaml config.yaml
89
+ 2022-03-03 00:42:27,156 INFO SenderThread:267379 [sender.py:transition_state():392] send defer: 6
90
+ 2022-03-03 00:42:27,156 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: poll_exit
91
+ 2022-03-03 00:42:27,157 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
92
+ 2022-03-03 00:42:27,157 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 6
93
+ 2022-03-03 00:42:27,160 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
94
+ 2022-03-03 00:42:27,160 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 6
95
+ 2022-03-03 00:42:27,160 INFO SenderThread:267379 [file_pusher.py:finish():177] shutting down file pusher
96
+ 2022-03-03 00:42:27,258 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: poll_exit
97
+ 2022-03-03 00:42:27,258 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: poll_exit
98
+ 2022-03-03 00:42:27,360 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: poll_exit
99
+ 2022-03-03 00:42:27,360 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: poll_exit
100
+ 2022-03-03 00:42:27,434 INFO Thread-12 :267379 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/output.log
101
+ 2022-03-03 00:42:27,457 INFO Thread-13 :267379 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/wandb-summary.json
102
+ 2022-03-03 00:42:27,460 INFO Thread-15 :267379 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/config.yaml
103
+ 2022-03-03 00:42:27,462 INFO Thread-14 :267379 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/files/requirements.txt
104
+ 2022-03-03 00:42:27,462 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: poll_exit
105
+ 2022-03-03 00:42:27,463 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: poll_exit
106
+ 2022-03-03 00:42:27,565 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: poll_exit
107
+ 2022-03-03 00:42:27,565 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: poll_exit
108
+ 2022-03-03 00:42:27,663 INFO Thread-7 :267379 [sender.py:transition_state():392] send defer: 7
109
+ 2022-03-03 00:42:27,664 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
110
+ 2022-03-03 00:42:27,664 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 7
111
+ 2022-03-03 00:42:27,664 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
112
+ 2022-03-03 00:42:27,664 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 7
113
+ 2022-03-03 00:42:27,667 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: poll_exit
114
+ 2022-03-03 00:42:27,834 INFO SenderThread:267379 [sender.py:transition_state():392] send defer: 8
115
+ 2022-03-03 00:42:27,834 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: poll_exit
116
+ 2022-03-03 00:42:27,835 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
117
+ 2022-03-03 00:42:27,835 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 8
118
+ 2022-03-03 00:42:27,835 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
119
+ 2022-03-03 00:42:27,835 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 8
120
+ 2022-03-03 00:42:27,835 INFO SenderThread:267379 [sender.py:transition_state():392] send defer: 9
121
+ 2022-03-03 00:42:27,836 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: defer
122
+ 2022-03-03 00:42:27,836 INFO HandlerThread:267379 [handler.py:handle_request_defer():154] handle defer: 9
123
+ 2022-03-03 00:42:27,836 DEBUG SenderThread:267379 [sender.py:send():235] send: final
124
+ 2022-03-03 00:42:27,836 DEBUG SenderThread:267379 [sender.py:send():235] send: footer
125
+ 2022-03-03 00:42:27,836 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: defer
126
+ 2022-03-03 00:42:27,836 INFO SenderThread:267379 [sender.py:send_request_defer():388] handle sender defer: 9
127
+ 2022-03-03 00:42:27,936 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: poll_exit
128
+ 2022-03-03 00:42:27,937 DEBUG SenderThread:267379 [sender.py:send_request():249] send_request: poll_exit
129
+ 2022-03-03 00:42:27,937 INFO SenderThread:267379 [file_pusher.py:join():182] waiting for file pusher
130
+ 2022-03-03 00:42:28,000 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: get_summary
131
+ 2022-03-03 00:42:28,001 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: sampled_history
132
+ 2022-03-03 00:42:28,002 DEBUG HandlerThread:267379 [handler.py:handle_request():131] handle_request: shutdown
133
+ 2022-03-03 00:42:28,002 INFO HandlerThread:267379 [handler.py:finish():739] shutting down handler
134
+ 2022-03-03 00:42:28,836 INFO WriterThread:267379 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220303_004215-2ttcbe3s/run-2ttcbe3s.wandb
135
+ 2022-03-03 00:42:28,999 INFO SenderThread:267379 [sender.py:finish():1075] shutting down sender
136
+ 2022-03-03 00:42:28,999 INFO SenderThread:267379 [file_pusher.py:finish():177] shutting down file pusher
137
+ 2022-03-03 00:42:28,999 INFO SenderThread:267379 [file_pusher.py:join():182] waiting for file pusher
138
+ 2022-03-03 00:42:29,001 INFO MainThread:267379 [internal.py:handle_exit():79] Internal process exited