sanchit-gandhi HF staff commited on
Commit
53a9ef4
1 Parent(s): 3350b71

Training in progress, step 500

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. pytorch_model.bin +1 -1
  3. training_args.bin +1 -1
  4. wandb/debug-internal.log +1 -1
  5. wandb/debug.log +1 -1
  6. wandb/latest-run +1 -1
  7. wandb/run-20220301_123331-3cwoccr3/files/config.yaml +7 -0
  8. wandb/run-20220301_123331-3cwoccr3/files/output.log +7 -0
  9. wandb/run-20220301_123331-3cwoccr3/files/wandb-summary.json +0 -0
  10. wandb/run-20220301_123331-3cwoccr3/logs/debug-internal.log +116 -0
  11. wandb/run-20220301_123331-3cwoccr3/logs/debug.log +106 -0
  12. wandb/run-20220301_123331-3cwoccr3/run-3cwoccr3.wandb +2 -2
  13. wandb/run-20220301_141622-dv1pefph/files/config.yaml +0 -0
  14. wandb/run-20220301_141622-dv1pefph/files/output.log +0 -0
  15. wandb/run-20220301_141622-dv1pefph/files/requirements.txt +184 -0
  16. wandb/run-20220301_141622-dv1pefph/files/wandb-metadata.json +60 -0
  17. wandb/run-20220301_141622-dv1pefph/files/wandb-summary.json +0 -0
  18. wandb/run-20220301_141622-dv1pefph/logs/debug-internal.log +0 -0
  19. wandb/run-20220301_141622-dv1pefph/logs/debug.log +124 -0
  20. wandb/run-20220301_141622-dv1pefph/run-dv1pefph.wandb +3 -0
  21. wandb/run-20220301_145657-3jv2086j/files/config.yaml +0 -0
  22. wandb/run-20220301_145657-3jv2086j/files/output.log +0 -0
  23. wandb/run-20220301_145657-3jv2086j/files/requirements.txt +184 -0
  24. wandb/run-20220301_145657-3jv2086j/files/wandb-metadata.json +60 -0
  25. wandb/run-20220301_145657-3jv2086j/files/wandb-summary.json +0 -0
  26. wandb/run-20220301_145657-3jv2086j/logs/debug-internal.log +0 -0
  27. wandb/run-20220301_145657-3jv2086j/logs/debug.log +124 -0
  28. wandb/run-20220301_145657-3jv2086j/run-3jv2086j.wandb +3 -0
  29. wandb/run-20220301_160718-1tlgvk9e/files/config.yaml +0 -0
  30. wandb/run-20220301_160718-1tlgvk9e/files/output.log +0 -0
  31. wandb/run-20220301_160718-1tlgvk9e/files/requirements.txt +184 -0
  32. wandb/run-20220301_160718-1tlgvk9e/files/wandb-metadata.json +60 -0
  33. wandb/run-20220301_160718-1tlgvk9e/files/wandb-summary.json +0 -0
  34. wandb/run-20220301_160718-1tlgvk9e/logs/debug-internal.log +0 -0
  35. wandb/run-20220301_160718-1tlgvk9e/logs/debug.log +132 -0
  36. wandb/run-20220301_160718-1tlgvk9e/run-1tlgvk9e.wandb +3 -0
  37. wandb/run-20220301_171032-6le9s370/files/config.yaml +713 -0
  38. wandb/run-20220301_171032-6le9s370/files/output.log +37 -0
  39. wandb/run-20220301_171032-6le9s370/files/requirements.txt +184 -0
  40. wandb/run-20220301_171032-6le9s370/files/wandb-metadata.json +60 -0
  41. wandb/run-20220301_171032-6le9s370/files/wandb-summary.json +1 -0
  42. wandb/run-20220301_171032-6le9s370/logs/debug-internal.log +137 -0
  43. wandb/run-20220301_171032-6le9s370/logs/debug.log +107 -0
  44. wandb/run-20220301_171032-6le9s370/run-6le9s370.wandb +0 -0
  45. wandb/run-20220301_171200-3h52ha0p/files/config.yaml +713 -0
  46. wandb/run-20220301_171200-3h52ha0p/files/output.log +37 -0
  47. wandb/run-20220301_171200-3h52ha0p/files/requirements.txt +184 -0
  48. wandb/run-20220301_171200-3h52ha0p/files/wandb-metadata.json +60 -0
  49. wandb/run-20220301_171200-3h52ha0p/files/wandb-summary.json +1 -0
  50. wandb/run-20220301_171200-3h52ha0p/logs/debug-internal.log +140 -0
.gitattributes CHANGED
@@ -40,3 +40,7 @@ wandb/run-20220301_062854-elbn80d0/run-elbn80d0.wandb filter=lfs diff=lfs merge=
40
  wandb/run-20220301_082052-1y1by2yf/run-1y1by2yf.wandb filter=lfs diff=lfs merge=lfs -text
41
  wandb/run-20220301_103527-1wkgn37c/run-1wkgn37c.wandb filter=lfs diff=lfs merge=lfs -text
42
  wandb/run-20220301_123331-3cwoccr3/run-3cwoccr3.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
 
40
  wandb/run-20220301_082052-1y1by2yf/run-1y1by2yf.wandb filter=lfs diff=lfs merge=lfs -text
41
  wandb/run-20220301_103527-1wkgn37c/run-1wkgn37c.wandb filter=lfs diff=lfs merge=lfs -text
42
  wandb/run-20220301_123331-3cwoccr3/run-3cwoccr3.wandb filter=lfs diff=lfs merge=lfs -text
43
+ wandb/run-20220301_141622-dv1pefph/run-dv1pefph.wandb filter=lfs diff=lfs merge=lfs -text
44
+ wandb/run-20220301_145657-3jv2086j/run-3jv2086j.wandb filter=lfs diff=lfs merge=lfs -text
45
+ wandb/run-20220301_160718-1tlgvk9e/run-1tlgvk9e.wandb filter=lfs diff=lfs merge=lfs -text
46
+ wandb/run-20220301_171456-1jxdkzs7/run-1jxdkzs7.wandb filter=lfs diff=lfs merge=lfs -text
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f4de1aaa889324c95783d73aed24cd73a1ece76550e594efb3d6dc79db63ae2
3
  size 3210531882
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:163e2a8f6a9a3cf5d3c7899f308e2350e38d82f3a1c799d3363f86333955dee9
3
  size 3210531882
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c56ed8ff2200e382bcb83dc4fb515c294487cb31fd96a2420cdda4b9dee61420
3
  size 3119
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e841c874653bfe8fcfe53fd1724eb324c10616992a45664ed489f76836a9e834
3
  size 3119
wandb/debug-internal.log CHANGED
@@ -1 +1 @@
1
- run-20220301_123331-3cwoccr3/logs/debug-internal.log
1
+ run-20220301_171456-1jxdkzs7/logs/debug-internal.log
wandb/debug.log CHANGED
@@ -1 +1 @@
1
- run-20220301_123331-3cwoccr3/logs/debug.log
1
+ run-20220301_171456-1jxdkzs7/logs/debug.log
wandb/latest-run CHANGED
@@ -1 +1 @@
1
- run-20220301_123331-3cwoccr3
1
+ run-20220301_171456-1jxdkzs7
wandb/run-20220301_123331-3cwoccr3/files/config.yaml CHANGED
@@ -10673,7 +10673,14 @@ _wandb:
10673
  - 1
10674
  - 5
10675
  - 11
 
 
 
 
 
10676
  3:
 
 
10677
  - 13
10678
  4: 3.9.5
10679
  5: 0.12.10
10673
  - 1
10674
  - 5
10675
  - 11
10676
+ 2:
10677
+ - 1
10678
+ - 5
10679
+ - 11
10680
+ - 12
10681
  3:
10682
+ - 1
10683
+ - 7
10684
  - 13
10685
  4: 3.9.5
10686
  5: 0.12.10
wandb/run-20220301_123331-3cwoccr3/files/output.log CHANGED
@@ -2765,3 +2765,10 @@ Upload file wandb/run-20220301_123331-3cwoccr3/run-3cwoccr3.wandb: 39%|██
2765
  eval_steps_per_second = 0.236
2766
  [INFO|trainer.py:2114] 2022-03-01 14:14:09,310 >> Saving model checkpoint to ./ | 4/221 [00:10<10:41, 2.96s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2767
  [INFO|modeling_utils.py:1081] 2022-03-01 14:14:25,817 >> Model weights saved in ./pytorch_model.bin:10<10:41, 2.96s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
 
 
 
 
 
 
 
2765
  eval_steps_per_second = 0.236
2766
  [INFO|trainer.py:2114] 2022-03-01 14:14:09,310 >> Saving model checkpoint to ./ | 4/221 [00:10<10:41, 2.96s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2767
  [INFO|modeling_utils.py:1081] 2022-03-01 14:14:25,817 >> Model weights saved in ./pytorch_model.bin:10<10:41, 2.96s/it] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2768
+ Upload file wandb/run-20220301_123331-3cwoccr3/run-3cwoccr3.wandb: 0%| | 32.0k/34.8M [00:00<?, ?B/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2769
+ Upload file wandb/run-20220301_123331-3cwoccr3/run-3cwoccr3.wandb: 0%| | 32.0k/34.8M [00:00<?, ?B/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2770
+ 03/01/2022 14:14:54 - WARNING - huggingface_hub.repository - To https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search
2771
+ Upload file wandb/run-20220301_123331-3cwoccr3/run-3cwoccr3.wandb: 100%|███████████| 34.8M/34.8M [00:02<00:00, 18.2MB/s] argument in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2772
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2773
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
2774
+ return ModelInfo(**d)f.finetuned_from)formers/src/transformers/modelcard.py", line 611, in from_trainercard31, in mainule>ent in `SpeechEncoderDecoderModel.forward` and have been ignored: input_length. If input_length are not expected by `SpeechEncoderDecoderModel.forward`, you can safely ignore this message.
wandb/run-20220301_123331-3cwoccr3/files/wandb-summary.json CHANGED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_123331-3cwoccr3/logs/debug-internal.log CHANGED
@@ -7491,3 +7491,119 @@
7491
  2022-03-01 14:14:27,950 INFO Thread-8 :243622 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/output.log
7492
  2022-03-01 14:14:38,156 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: stop_status
7493
  2022-03-01 14:14:38,157 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: stop_status
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7491
  2022-03-01 14:14:27,950 INFO Thread-8 :243622 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/output.log
7492
  2022-03-01 14:14:38,156 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: stop_status
7493
  2022-03-01 14:14:38,157 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: stop_status
7494
+ 2022-03-01 14:14:52,101 DEBUG SenderThread:243622 [sender.py:send():235] send: stats
7495
+ 2022-03-01 14:14:53,206 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: stop_status
7496
+ 2022-03-01 14:14:53,207 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: stop_status
7497
+ 2022-03-01 14:14:53,959 INFO Thread-8 :243622 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/output.log
7498
+ 2022-03-01 14:14:54,959 INFO Thread-8 :243622 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/output.log
7499
+ 2022-03-01 14:14:55,959 INFO Thread-8 :243622 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/output.log
7500
+ 2022-03-01 14:14:59,961 INFO Thread-8 :243622 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/output.log
7501
+ 2022-03-01 14:15:00,428 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7502
+ 2022-03-01 14:15:00,429 DEBUG SenderThread:243622 [sender.py:send():235] send: telemetry
7503
+ 2022-03-01 14:15:00,429 DEBUG SenderThread:243622 [sender.py:send():235] send: exit
7504
+ 2022-03-01 14:15:00,429 INFO SenderThread:243622 [sender.py:send_exit():371] handling exit code: 1
7505
+ 2022-03-01 14:15:00,429 INFO SenderThread:243622 [sender.py:send_exit():373] handling runtime: 6088
7506
+ 2022-03-01 14:15:00,482 INFO SenderThread:243622 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
7507
+ 2022-03-01 14:15:00,483 INFO SenderThread:243622 [sender.py:send_exit():379] send defer
7508
+ 2022-03-01 14:15:00,483 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7509
+ 2022-03-01 14:15:00,483 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7510
+ 2022-03-01 14:15:00,483 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 0
7511
+ 2022-03-01 14:15:00,484 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7512
+ 2022-03-01 14:15:00,484 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 0
7513
+ 2022-03-01 14:15:00,484 INFO SenderThread:243622 [sender.py:transition_state():392] send defer: 1
7514
+ 2022-03-01 14:15:00,484 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7515
+ 2022-03-01 14:15:00,484 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 1
7516
+ 2022-03-01 14:15:00,506 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7517
+ 2022-03-01 14:15:00,506 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 1
7518
+ 2022-03-01 14:15:00,506 INFO SenderThread:243622 [sender.py:transition_state():392] send defer: 2
7519
+ 2022-03-01 14:15:00,506 DEBUG SenderThread:243622 [sender.py:send():235] send: stats
7520
+ 2022-03-01 14:15:00,507 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7521
+ 2022-03-01 14:15:00,507 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 2
7522
+ 2022-03-01 14:15:00,507 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7523
+ 2022-03-01 14:15:00,507 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 2
7524
+ 2022-03-01 14:15:00,507 INFO SenderThread:243622 [sender.py:transition_state():392] send defer: 3
7525
+ 2022-03-01 14:15:00,508 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7526
+ 2022-03-01 14:15:00,508 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 3
7527
+ 2022-03-01 14:15:00,565 DEBUG SenderThread:243622 [sender.py:send():235] send: summary
7528
+ 2022-03-01 14:15:00,588 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7529
+ 2022-03-01 14:15:00,648 INFO SenderThread:243622 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
7530
+ 2022-03-01 14:15:00,648 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7531
+ 2022-03-01 14:15:00,648 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 3
7532
+ 2022-03-01 14:15:00,648 INFO SenderThread:243622 [sender.py:transition_state():392] send defer: 4
7533
+ 2022-03-01 14:15:00,648 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7534
+ 2022-03-01 14:15:00,649 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7535
+ 2022-03-01 14:15:00,649 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 4
7536
+ 2022-03-01 14:15:00,649 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7537
+ 2022-03-01 14:15:00,649 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 4
7538
+ 2022-03-01 14:15:00,750 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7539
+ 2022-03-01 14:15:01,012 INFO Thread-8 :243622 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/wandb-summary.json
7540
+ 2022-03-01 14:15:01,023 INFO Thread-8 :243622 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/output.log
7541
+ 2022-03-01 14:15:01,559 INFO SenderThread:243622 [sender.py:transition_state():392] send defer: 5
7542
+ 2022-03-01 14:15:01,559 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7543
+ 2022-03-01 14:15:01,560 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7544
+ 2022-03-01 14:15:01,560 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 5
7545
+ 2022-03-01 14:15:01,560 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7546
+ 2022-03-01 14:15:01,560 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 5
7547
+ 2022-03-01 14:15:01,560 INFO SenderThread:243622 [dir_watcher.py:finish():283] shutting down directory watcher
7548
+ 2022-03-01 14:15:01,661 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7549
+ 2022-03-01 14:15:02,003 INFO Thread-8 :243622 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/config.yaml
7550
+ 2022-03-01 14:15:02,003 INFO SenderThread:243622 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files
7551
+ 2022-03-01 14:15:02,004 INFO SenderThread:243622 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/wandb-metadata.json wandb-metadata.json
7552
+ 2022-03-01 14:15:02,004 INFO SenderThread:243622 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/output.log output.log
7553
+ 2022-03-01 14:15:02,004 INFO SenderThread:243622 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/wandb-summary.json wandb-summary.json
7554
+ 2022-03-01 14:15:02,004 INFO SenderThread:243622 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/requirements.txt requirements.txt
7555
+ 2022-03-01 14:15:02,005 INFO SenderThread:243622 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/config.yaml config.yaml
7556
+ 2022-03-01 14:15:02,011 INFO SenderThread:243622 [sender.py:transition_state():392] send defer: 6
7557
+ 2022-03-01 14:15:02,013 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7558
+ 2022-03-01 14:15:02,014 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7559
+ 2022-03-01 14:15:02,014 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 6
7560
+ 2022-03-01 14:15:02,018 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7561
+ 2022-03-01 14:15:02,018 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 6
7562
+ 2022-03-01 14:15:02,018 INFO SenderThread:243622 [file_pusher.py:finish():177] shutting down file pusher
7563
+ 2022-03-01 14:15:02,115 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7564
+ 2022-03-01 14:15:02,115 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7565
+ 2022-03-01 14:15:02,217 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7566
+ 2022-03-01 14:15:02,217 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7567
+ 2022-03-01 14:15:02,318 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7568
+ 2022-03-01 14:15:02,319 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7569
+ 2022-03-01 14:15:02,370 INFO Thread-14 :243622 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/requirements.txt
7570
+ 2022-03-01 14:15:02,403 INFO Thread-12 :243622 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/output.log
7571
+ 2022-03-01 14:15:02,420 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7572
+ 2022-03-01 14:15:02,420 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7573
+ 2022-03-01 14:15:02,423 INFO Thread-15 :243622 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/config.yaml
7574
+ 2022-03-01 14:15:02,496 INFO Thread-13 :243622 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/files/wandb-summary.json
7575
+ 2022-03-01 14:15:02,521 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7576
+ 2022-03-01 14:15:02,522 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7577
+ 2022-03-01 14:15:02,623 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7578
+ 2022-03-01 14:15:02,623 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7579
+ 2022-03-01 14:15:02,696 INFO Thread-7 :243622 [sender.py:transition_state():392] send defer: 7
7580
+ 2022-03-01 14:15:02,697 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7581
+ 2022-03-01 14:15:02,697 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 7
7582
+ 2022-03-01 14:15:02,697 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7583
+ 2022-03-01 14:15:02,697 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 7
7584
+ 2022-03-01 14:15:02,724 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7585
+ 2022-03-01 14:15:04,010 INFO SenderThread:243622 [sender.py:transition_state():392] send defer: 8
7586
+ 2022-03-01 14:15:04,011 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7587
+ 2022-03-01 14:15:04,011 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7588
+ 2022-03-01 14:15:04,011 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 8
7589
+ 2022-03-01 14:15:04,012 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7590
+ 2022-03-01 14:15:04,012 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 8
7591
+ 2022-03-01 14:15:04,012 INFO SenderThread:243622 [sender.py:transition_state():392] send defer: 9
7592
+ 2022-03-01 14:15:04,012 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: defer
7593
+ 2022-03-01 14:15:04,012 INFO HandlerThread:243622 [handler.py:handle_request_defer():154] handle defer: 9
7594
+ 2022-03-01 14:15:04,013 DEBUG SenderThread:243622 [sender.py:send():235] send: final
7595
+ 2022-03-01 14:15:04,014 DEBUG SenderThread:243622 [sender.py:send():235] send: footer
7596
+ 2022-03-01 14:15:04,014 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: defer
7597
+ 2022-03-01 14:15:04,014 INFO SenderThread:243622 [sender.py:send_request_defer():388] handle sender defer: 9
7598
+ 2022-03-01 14:15:04,112 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: poll_exit
7599
+ 2022-03-01 14:15:04,113 DEBUG SenderThread:243622 [sender.py:send_request():249] send_request: poll_exit
7600
+ 2022-03-01 14:15:04,113 INFO SenderThread:243622 [file_pusher.py:join():182] waiting for file pusher
7601
+ 2022-03-01 14:15:04,207 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: get_summary
7602
+ 2022-03-01 14:15:04,307 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: sampled_history
7603
+ 2022-03-01 14:15:04,309 DEBUG HandlerThread:243622 [handler.py:handle_request():131] handle_request: shutdown
7604
+ 2022-03-01 14:15:04,309 INFO HandlerThread:243622 [handler.py:finish():739] shutting down handler
7605
+ 2022-03-01 14:15:05,014 INFO WriterThread:243622 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_123331-3cwoccr3/run-3cwoccr3.wandb
7606
+ 2022-03-01 14:15:05,206 INFO SenderThread:243622 [sender.py:finish():1075] shutting down sender
7607
+ 2022-03-01 14:15:05,206 INFO SenderThread:243622 [file_pusher.py:finish():177] shutting down file pusher
7608
+ 2022-03-01 14:15:05,206 INFO SenderThread:243622 [file_pusher.py:join():182] waiting for file pusher
7609
+ 2022-03-01 14:15:05,213 INFO MainThread:243622 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220301_123331-3cwoccr3/logs/debug.log CHANGED
@@ -24,3 +24,109 @@ config: {}
24
  2022-03-01 12:33:32,484 INFO MainThread:243523 [wandb_init.py:init():651] run started, returning control to user process
25
  2022-03-01 12:33:32,487 INFO MainThread:243523 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 3e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar01_12-32-47_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 12, 'eval_batch_size': 12}
26
  2022-03-01 12:33:32,493 INFO MainThread:243523 [wandb_watch.py:watch():43] Watching
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  2022-03-01 12:33:32,484 INFO MainThread:243523 [wandb_init.py:init():651] run started, returning control to user process
25
  2022-03-01 12:33:32,487 INFO MainThread:243523 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 12, 'per_device_eval_batch_size': 12, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 3e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar01_12-32-47_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 12, 'eval_batch_size': 12}
26
  2022-03-01 12:33:32,493 INFO MainThread:243523 [wandb_watch.py:watch():43] Watching
27
+ 2022-03-01 14:14:57,863 INFO MainThread:243523 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
28
+ 2022-03-01 14:14:57,866 INFO MainThread:243523 [wandb_run.py:_restore():1769] restore
29
+ 2022-03-01 14:15:00,483 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
30
+ wandb_count: 1
31
+ }
32
+ pusher_stats {
33
+ uploaded_bytes: 2095
34
+ total_bytes: 2095
35
+ }
36
+
37
+ 2022-03-01 14:15:00,649 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
38
+ wandb_count: 1
39
+ }
40
+ pusher_stats {
41
+ uploaded_bytes: 2095
42
+ total_bytes: 2095
43
+ }
44
+
45
+ 2022-03-01 14:15:01,560 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
46
+ wandb_count: 1
47
+ }
48
+ pusher_stats {
49
+ uploaded_bytes: 2095
50
+ total_bytes: 2095
51
+ }
52
+
53
+ 2022-03-01 14:15:02,014 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
54
+ wandb_count: 4
55
+ }
56
+ pusher_stats {
57
+ uploaded_bytes: 2095
58
+ total_bytes: 2354993
59
+ }
60
+
61
+ 2022-03-01 14:15:02,116 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
62
+ wandb_count: 5
63
+ }
64
+ pusher_stats {
65
+ uploaded_bytes: 2095
66
+ total_bytes: 2653161
67
+ }
68
+
69
+ 2022-03-01 14:15:02,218 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
70
+ wandb_count: 5
71
+ }
72
+ pusher_stats {
73
+ uploaded_bytes: 2653161
74
+ total_bytes: 2653161
75
+ }
76
+
77
+ 2022-03-01 14:15:02,319 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
78
+ wandb_count: 5
79
+ }
80
+ pusher_stats {
81
+ uploaded_bytes: 2653161
82
+ total_bytes: 2653161
83
+ }
84
+
85
+ 2022-03-01 14:15:02,421 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
86
+ wandb_count: 5
87
+ }
88
+ pusher_stats {
89
+ uploaded_bytes: 2653161
90
+ total_bytes: 2653161
91
+ }
92
+
93
+ 2022-03-01 14:15:02,522 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
94
+ wandb_count: 5
95
+ }
96
+ pusher_stats {
97
+ uploaded_bytes: 2653161
98
+ total_bytes: 2653161
99
+ }
100
+
101
+ 2022-03-01 14:15:02,623 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
102
+ wandb_count: 5
103
+ }
104
+ pusher_stats {
105
+ uploaded_bytes: 2653161
106
+ total_bytes: 2653161
107
+ }
108
+
109
+ 2022-03-01 14:15:04,011 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
110
+ wandb_count: 5
111
+ }
112
+ pusher_stats {
113
+ uploaded_bytes: 2653161
114
+ total_bytes: 2653161
115
+ }
116
+
117
+ 2022-03-01 14:15:04,206 INFO MainThread:243523 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
118
+ exit_result {
119
+ }
120
+ file_counts {
121
+ wandb_count: 5
122
+ }
123
+ pusher_stats {
124
+ uploaded_bytes: 2653161
125
+ total_bytes: 2653161
126
+ }
127
+ local_info {
128
+ }
129
+
130
+ 2022-03-01 14:15:05,350 INFO MainThread:243523 [wandb_run.py:_append_history():2144] rendering history
131
+ 2022-03-01 14:15:05,351 INFO MainThread:243523 [wandb_run.py:_append_summary():2102] rendering summary
132
+ 2022-03-01 14:15:05,352 INFO MainThread:243523 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220301_123331-3cwoccr3/run-3cwoccr3.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e99c29342cc96693fa49f17adaa4386cf2e7df077c82d3b8880bc3711001683
3
- size 36442048
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1b79b9c9ea1d106f02cdd3fcdc736bec22e3974959d390c48fbe032330f6835
3
+ size 37922404
wandb/run-20220301_141622-dv1pefph/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_141622-dv1pefph/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_141622-dv1pefph/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220301_141622-dv1pefph/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-01T14:16:23.586659",
5
+ "startedAt": "2022-03-01T14:16:22.498686",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=2",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "3350b717ff05e187076e5bb0722b6b02f9837858"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220301_141622-dv1pefph/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_141622-dv1pefph/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_141622-dv1pefph/logs/debug.log ADDED
@@ -0,0 +1,124 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_141622-dv1pefph/logs/debug.log
6
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_141622-dv1pefph/logs/debug-internal.log
7
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [wandb_init.py:init():471] starting backend
11
+ 2022-03-01 14:16:22,500 INFO MainThread:244420 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-01 14:16:22,561 INFO MainThread:244420 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-01 14:16:22,621 INFO MainThread:244420 [backend.py:ensure_launched():224] started backend process with pid: 244519
14
+ 2022-03-01 14:16:22,623 INFO MainThread:244420 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-01 14:16:22,633 INFO MainThread:244420 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-01 14:16:22,776 INFO MainThread:244420 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-01 14:16:23,472 INFO MainThread:244420 [wandb_init.py:init():586] got version response
18
+ 2022-03-01 14:16:23,472 INFO MainThread:244420 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
19
+ 2022-03-01 14:16:23,581 INFO MainThread:244420 [wandb_init.py:init():624] starting run threads in backend
20
+ 2022-03-01 14:16:23,687 INFO MainThread:244420 [wandb_run.py:_console_start():1827] atexit reg
21
+ 2022-03-01 14:16:23,687 INFO MainThread:244420 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
22
+ 2022-03-01 14:16:23,688 INFO MainThread:244420 [wandb_run.py:_redirect():1706] Redirecting console.
23
+ 2022-03-01 14:16:23,689 INFO MainThread:244420 [wandb_run.py:_redirect():1762] Redirects installed.
24
+ 2022-03-01 14:16:23,689 INFO MainThread:244420 [wandb_init.py:init():651] run started, returning control to user process
25
+ 2022-03-01 14:16:23,692 INFO MainThread:244420 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 3e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar01_14-15-41_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
26
+ 2022-03-01 14:16:23,695 INFO MainThread:244420 [wandb_watch.py:watch():43] Watching
27
+ 2022-03-01 14:55:32,877 INFO MainThread:244420 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
28
+ 2022-03-01 14:55:32,878 INFO MainThread:244420 [wandb_run.py:_restore():1769] restore
29
+ 2022-03-01 14:55:35,747 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
30
+ wandb_count: 1
31
+ }
32
+ pusher_stats {
33
+ uploaded_bytes: 2095
34
+ total_bytes: 2095
35
+ }
36
+
37
+ 2022-03-01 14:55:35,952 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
38
+ wandb_count: 1
39
+ }
40
+ pusher_stats {
41
+ uploaded_bytes: 2095
42
+ total_bytes: 2095
43
+ }
44
+
45
+ 2022-03-01 14:55:36,779 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
46
+ wandb_count: 1
47
+ }
48
+ pusher_stats {
49
+ uploaded_bytes: 2095
50
+ total_bytes: 2095
51
+ }
52
+
53
+ 2022-03-01 14:55:37,033 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
54
+ wandb_count: 4
55
+ }
56
+ pusher_stats {
57
+ uploaded_bytes: 2095
58
+ total_bytes: 1788026
59
+ }
60
+
61
+ 2022-03-01 14:55:37,135 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
62
+ wandb_count: 5
63
+ }
64
+ pusher_stats {
65
+ uploaded_bytes: 5332
66
+ total_bytes: 2085601
67
+ }
68
+
69
+ 2022-03-01 14:55:37,236 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
70
+ wandb_count: 5
71
+ }
72
+ pusher_stats {
73
+ uploaded_bytes: 2085601
74
+ total_bytes: 2085601
75
+ }
76
+
77
+ 2022-03-01 14:55:37,338 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
78
+ wandb_count: 5
79
+ }
80
+ pusher_stats {
81
+ uploaded_bytes: 2085601
82
+ total_bytes: 2085601
83
+ }
84
+
85
+ 2022-03-01 14:55:37,439 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
86
+ wandb_count: 5
87
+ }
88
+ pusher_stats {
89
+ uploaded_bytes: 2085601
90
+ total_bytes: 2085601
91
+ }
92
+
93
+ 2022-03-01 14:55:37,540 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
94
+ wandb_count: 5
95
+ }
96
+ pusher_stats {
97
+ uploaded_bytes: 2085601
98
+ total_bytes: 2085601
99
+ }
100
+
101
+ 2022-03-01 14:55:39,068 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
102
+ wandb_count: 5
103
+ }
104
+ pusher_stats {
105
+ uploaded_bytes: 2085601
106
+ total_bytes: 2085601
107
+ }
108
+
109
+ 2022-03-01 14:55:39,238 INFO MainThread:244420 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
110
+ exit_result {
111
+ }
112
+ file_counts {
113
+ wandb_count: 5
114
+ }
115
+ pusher_stats {
116
+ uploaded_bytes: 2085601
117
+ total_bytes: 2085601
118
+ }
119
+ local_info {
120
+ }
121
+
122
+ 2022-03-01 14:55:40,392 INFO MainThread:244420 [wandb_run.py:_append_history():2144] rendering history
123
+ 2022-03-01 14:55:40,392 INFO MainThread:244420 [wandb_run.py:_append_summary():2102] rendering summary
124
+ 2022-03-01 14:55:40,393 INFO MainThread:244420 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220301_141622-dv1pefph/run-dv1pefph.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a677a21d8c0f090e72bfa4c09551173003265f28b85145e20e25e4964b1f988
3
+ size 17228284
wandb/run-20220301_145657-3jv2086j/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_145657-3jv2086j/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_145657-3jv2086j/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220301_145657-3jv2086j/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-01T14:56:58.613681",
5
+ "startedAt": "2022-03-01T14:56:57.535095",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=4",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "3350b717ff05e187076e5bb0722b6b02f9837858"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220301_145657-3jv2086j/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_145657-3jv2086j/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_145657-3jv2086j/logs/debug.log ADDED
@@ -0,0 +1,124 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 14:56:57,536 INFO MainThread:244706 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-01 14:56:57,536 INFO MainThread:244706 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-01 14:56:57,536 INFO MainThread:244706 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-01 14:56:57,536 INFO MainThread:244706 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-01 14:56:57,537 INFO MainThread:244706 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_145657-3jv2086j/logs/debug.log
6
+ 2022-03-01 14:56:57,537 INFO MainThread:244706 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_145657-3jv2086j/logs/debug-internal.log
7
+ 2022-03-01 14:56:57,537 INFO MainThread:244706 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-01 14:56:57,537 INFO MainThread:244706 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-01 14:56:57,537 INFO MainThread:244706 [wandb_init.py:init():471] starting backend
11
+ 2022-03-01 14:56:57,537 INFO MainThread:244706 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-01 14:56:57,597 INFO MainThread:244706 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-01 14:56:57,656 INFO MainThread:244706 [backend.py:ensure_launched():224] started backend process with pid: 244806
14
+ 2022-03-01 14:56:57,658 INFO MainThread:244706 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-01 14:56:57,669 INFO MainThread:244706 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-01 14:56:57,812 INFO MainThread:244706 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-01 14:56:58,520 INFO MainThread:244706 [wandb_init.py:init():586] got version response
18
+ 2022-03-01 14:56:58,520 INFO MainThread:244706 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
19
+ 2022-03-01 14:56:58,608 INFO MainThread:244706 [wandb_init.py:init():624] starting run threads in backend
20
+ 2022-03-01 14:56:58,718 INFO MainThread:244706 [wandb_run.py:_console_start():1827] atexit reg
21
+ 2022-03-01 14:56:58,718 INFO MainThread:244706 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
22
+ 2022-03-01 14:56:58,719 INFO MainThread:244706 [wandb_run.py:_redirect():1706] Redirecting console.
23
+ 2022-03-01 14:56:58,720 INFO MainThread:244706 [wandb_run.py:_redirect():1762] Redirects installed.
24
+ 2022-03-01 14:56:58,720 INFO MainThread:244706 [wandb_init.py:init():651] run started, returning control to user process
25
+ 2022-03-01 14:56:58,723 INFO MainThread:244706 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': 'None', 'learning_rate': 3e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar01_14-56-15_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
26
+ 2022-03-01 14:56:58,726 INFO MainThread:244706 [wandb_watch.py:watch():43] Watching
27
+ 2022-03-01 16:05:53,456 INFO MainThread:244706 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
28
+ 2022-03-01 16:05:53,457 INFO MainThread:244706 [wandb_run.py:_restore():1769] restore
29
+ 2022-03-01 16:05:56,456 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
30
+ wandb_count: 1
31
+ }
32
+ pusher_stats {
33
+ uploaded_bytes: 2095
34
+ total_bytes: 2095
35
+ }
36
+
37
+ 2022-03-01 16:05:56,664 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
38
+ wandb_count: 1
39
+ }
40
+ pusher_stats {
41
+ uploaded_bytes: 2095
42
+ total_bytes: 2095
43
+ }
44
+
45
+ 2022-03-01 16:05:57,497 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
46
+ wandb_count: 1
47
+ }
48
+ pusher_stats {
49
+ uploaded_bytes: 2095
50
+ total_bytes: 2095
51
+ }
52
+
53
+ 2022-03-01 16:05:58,122 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
54
+ wandb_count: 5
55
+ }
56
+ pusher_stats {
57
+ uploaded_bytes: 2095
58
+ total_bytes: 2447422
59
+ }
60
+
61
+ 2022-03-01 16:05:58,223 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
62
+ wandb_count: 5
63
+ }
64
+ pusher_stats {
65
+ uploaded_bytes: 1783627
66
+ total_bytes: 2447422
67
+ }
68
+
69
+ 2022-03-01 16:05:58,325 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
70
+ wandb_count: 5
71
+ }
72
+ pusher_stats {
73
+ uploaded_bytes: 2447422
74
+ total_bytes: 2447422
75
+ }
76
+
77
+ 2022-03-01 16:05:58,426 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
78
+ wandb_count: 5
79
+ }
80
+ pusher_stats {
81
+ uploaded_bytes: 2447422
82
+ total_bytes: 2447422
83
+ }
84
+
85
+ 2022-03-01 16:05:58,528 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
86
+ wandb_count: 5
87
+ }
88
+ pusher_stats {
89
+ uploaded_bytes: 2447422
90
+ total_bytes: 2447422
91
+ }
92
+
93
+ 2022-03-01 16:05:58,629 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
94
+ wandb_count: 5
95
+ }
96
+ pusher_stats {
97
+ uploaded_bytes: 2447422
98
+ total_bytes: 2447422
99
+ }
100
+
101
+ 2022-03-01 16:06:00,819 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
102
+ wandb_count: 5
103
+ }
104
+ pusher_stats {
105
+ uploaded_bytes: 2447422
106
+ total_bytes: 2447422
107
+ }
108
+
109
+ 2022-03-01 16:06:00,974 INFO MainThread:244706 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
110
+ exit_result {
111
+ }
112
+ file_counts {
113
+ wandb_count: 5
114
+ }
115
+ pusher_stats {
116
+ uploaded_bytes: 2447422
117
+ total_bytes: 2447422
118
+ }
119
+ local_info {
120
+ }
121
+
122
+ 2022-03-01 16:06:02,118 INFO MainThread:244706 [wandb_run.py:_append_history():2144] rendering history
123
+ 2022-03-01 16:06:02,119 INFO MainThread:244706 [wandb_run.py:_append_summary():2102] rendering summary
124
+ 2022-03-01 16:06:02,120 INFO MainThread:244706 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220301_145657-3jv2086j/run-3jv2086j.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a34a2f771a40312059d1c182471c3abad7e8dd47ecf3da7834feca3207f35ba
3
+ size 30267010
wandb/run-20220301_160718-1tlgvk9e/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_160718-1tlgvk9e/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_160718-1tlgvk9e/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220301_160718-1tlgvk9e/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-01T16:07:19.857648",
5
+ "startedAt": "2022-03-01T16:07:18.747130",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=14",
24
+ "--per_device_eval_batch_size=14",
25
+ "--gradient_accumulation_steps=8",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "3350b717ff05e187076e5bb0722b6b02f9837858"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220301_160718-1tlgvk9e/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_160718-1tlgvk9e/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220301_160718-1tlgvk9e/logs/debug.log ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 16:07:18,748 INFO MainThread:244920 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-01 16:07:18,749 INFO MainThread:244920 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-01 16:07:18,749 INFO MainThread:244920 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-01 16:07:18,749 INFO MainThread:244920 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-01 16:07:18,749 INFO MainThread:244920 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_160718-1tlgvk9e/logs/debug.log
6
+ 2022-03-01 16:07:18,749 INFO MainThread:244920 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_160718-1tlgvk9e/logs/debug-internal.log
7
+ 2022-03-01 16:07:18,749 INFO MainThread:244920 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-01 16:07:18,749 INFO MainThread:244920 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-01 16:07:18,749 INFO MainThread:244920 [wandb_init.py:init():471] starting backend
11
+ 2022-03-01 16:07:18,749 INFO MainThread:244920 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-01 16:07:18,810 INFO MainThread:244920 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-01 16:07:18,870 INFO MainThread:244920 [backend.py:ensure_launched():224] started backend process with pid: 245019
14
+ 2022-03-01 16:07:18,873 INFO MainThread:244920 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-01 16:07:18,883 INFO MainThread:244920 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-01 16:07:19,025 INFO MainThread:244920 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-01 16:07:19,744 INFO MainThread:244920 [wandb_init.py:init():586] got version response
18
+ 2022-03-01 16:07:19,745 INFO MainThread:244920 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
19
+ 2022-03-01 16:07:19,852 INFO MainThread:244920 [wandb_init.py:init():624] starting run threads in backend
20
+ 2022-03-01 16:07:19,961 INFO MainThread:244920 [wandb_run.py:_console_start():1827] atexit reg
21
+ 2022-03-01 16:07:19,961 INFO MainThread:244920 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
22
+ 2022-03-01 16:07:19,962 INFO MainThread:244920 [wandb_run.py:_redirect():1706] Redirecting console.
23
+ 2022-03-01 16:07:19,963 INFO MainThread:244920 [wandb_run.py:_redirect():1762] Redirects installed.
24
+ 2022-03-01 16:07:19,963 INFO MainThread:244920 [wandb_init.py:init():651] run started, returning control to user process
25
+ 2022-03-01 16:07:19,966 INFO MainThread:244920 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 14, 'per_device_eval_batch_size': 14, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': 'None', 'learning_rate': 3e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar01_16-06-37_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 14, 'eval_batch_size': 14}
26
+ 2022-03-01 16:07:19,968 INFO MainThread:244920 [wandb_watch.py:watch():43] Watching
27
+ 2022-03-01 17:09:06,822 INFO MainThread:244920 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
28
+ 2022-03-01 17:09:06,824 INFO MainThread:244920 [wandb_run.py:_restore():1769] restore
29
+ 2022-03-01 17:09:09,501 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
30
+ wandb_count: 1
31
+ }
32
+ pusher_stats {
33
+ uploaded_bytes: 2095
34
+ total_bytes: 2095
35
+ }
36
+
37
+ 2022-03-01 17:09:09,696 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
38
+ wandb_count: 1
39
+ }
40
+ pusher_stats {
41
+ uploaded_bytes: 2095
42
+ total_bytes: 2095
43
+ }
44
+
45
+ 2022-03-01 17:09:10,577 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
46
+ wandb_count: 1
47
+ }
48
+ pusher_stats {
49
+ uploaded_bytes: 2095
50
+ total_bytes: 2095
51
+ }
52
+
53
+ 2022-03-01 17:09:11,559 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
54
+ wandb_count: 4
55
+ }
56
+ pusher_stats {
57
+ uploaded_bytes: 2095
58
+ total_bytes: 2035546
59
+ }
60
+
61
+ 2022-03-01 17:09:11,660 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
62
+ wandb_count: 5
63
+ }
64
+ pusher_stats {
65
+ uploaded_bytes: 2095
66
+ total_bytes: 2333121
67
+ }
68
+
69
+ 2022-03-01 17:09:11,762 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
70
+ wandb_count: 5
71
+ }
72
+ pusher_stats {
73
+ uploaded_bytes: 2333121
74
+ total_bytes: 2333121
75
+ }
76
+
77
+ 2022-03-01 17:09:11,864 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
78
+ wandb_count: 5
79
+ }
80
+ pusher_stats {
81
+ uploaded_bytes: 2333121
82
+ total_bytes: 2333121
83
+ }
84
+
85
+ 2022-03-01 17:09:11,965 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
86
+ wandb_count: 5
87
+ }
88
+ pusher_stats {
89
+ uploaded_bytes: 2333121
90
+ total_bytes: 2333121
91
+ }
92
+
93
+ 2022-03-01 17:09:12,067 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
94
+ wandb_count: 5
95
+ }
96
+ pusher_stats {
97
+ uploaded_bytes: 2333121
98
+ total_bytes: 2333121
99
+ }
100
+
101
+ 2022-03-01 17:09:12,168 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
102
+ wandb_count: 5
103
+ }
104
+ pusher_stats {
105
+ uploaded_bytes: 2333121
106
+ total_bytes: 2333121
107
+ }
108
+
109
+ 2022-03-01 17:09:13,926 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
110
+ wandb_count: 5
111
+ }
112
+ pusher_stats {
113
+ uploaded_bytes: 2333121
114
+ total_bytes: 2333121
115
+ }
116
+
117
+ 2022-03-01 17:09:14,087 INFO MainThread:244920 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
118
+ exit_result {
119
+ }
120
+ file_counts {
121
+ wandb_count: 5
122
+ }
123
+ pusher_stats {
124
+ uploaded_bytes: 2333121
125
+ total_bytes: 2333121
126
+ }
127
+ local_info {
128
+ }
129
+
130
+ 2022-03-01 17:09:15,234 INFO MainThread:244920 [wandb_run.py:_append_history():2144] rendering history
131
+ 2022-03-01 17:09:15,235 INFO MainThread:244920 [wandb_run.py:_append_summary():2102] rendering summary
132
+ 2022-03-01 17:09:15,236 INFO MainThread:244920 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220301_160718-1tlgvk9e/run-1tlgvk9e.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18c948ff1175ae7fcc6f9c6944c32ea74e9cd195fda7f8f3101e80d61cdb60d9
3
+ size 26937836
wandb/run-20220301_171032-6le9s370/files/config.yaml ADDED
@@ -0,0 +1,713 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: ./
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.10
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.9.5
22
+ start_time: 1646154632
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.9.5
37
+ 5: 0.12.10
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ adafactor:
42
+ desc: null
43
+ value: false
44
+ adam_beta1:
45
+ desc: null
46
+ value: 0.9
47
+ adam_beta2:
48
+ desc: null
49
+ value: 0.999
50
+ adam_epsilon:
51
+ desc: null
52
+ value: 1.0e-08
53
+ add_cross_attention:
54
+ desc: null
55
+ value: false
56
+ architectures:
57
+ desc: null
58
+ value:
59
+ - SpeechEncoderDecoderModel
60
+ bad_words_ids:
61
+ desc: null
62
+ value: null
63
+ bf16:
64
+ desc: null
65
+ value: false
66
+ bf16_full_eval:
67
+ desc: null
68
+ value: false
69
+ bos_token_id:
70
+ desc: null
71
+ value: null
72
+ chunk_size_feed_forward:
73
+ desc: null
74
+ value: 0
75
+ cross_attention_hidden_size:
76
+ desc: null
77
+ value: null
78
+ dataloader_drop_last:
79
+ desc: null
80
+ value: false
81
+ dataloader_num_workers:
82
+ desc: null
83
+ value: 0
84
+ dataloader_pin_memory:
85
+ desc: null
86
+ value: true
87
+ ddp_bucket_cap_mb:
88
+ desc: null
89
+ value: None
90
+ ddp_find_unused_parameters:
91
+ desc: null
92
+ value: None
93
+ debug:
94
+ desc: null
95
+ value: '[]'
96
+ decoder:
97
+ desc: null
98
+ value:
99
+ _name_or_path: gpt2-medium
100
+ activation_function: gelu_new
101
+ add_cross_attention: true
102
+ architectures:
103
+ - GPT2LMHeadModel
104
+ attn_pdrop: 0.0
105
+ bad_words_ids: null
106
+ bos_token_id: 50256
107
+ chunk_size_feed_forward: 0
108
+ cross_attention_hidden_size: null
109
+ decoder_start_token_id: null
110
+ diversity_penalty: 0.0
111
+ do_sample: false
112
+ early_stopping: false
113
+ embd_pdrop: 0.0
114
+ encoder_no_repeat_ngram_size: 0
115
+ eos_token_id: 50256
116
+ finetuning_task: null
117
+ forced_bos_token_id: null
118
+ forced_eos_token_id: null
119
+ id2label:
120
+ '0': LABEL_0
121
+ '1': LABEL_1
122
+ initializer_range: 0.02
123
+ is_decoder: true
124
+ is_encoder_decoder: false
125
+ label2id:
126
+ LABEL_0: 0
127
+ LABEL_1: 1
128
+ layer_norm_epsilon: 1.0e-05
129
+ length_penalty: 1.0
130
+ max_length: 20
131
+ min_length: 0
132
+ model_type: gpt2
133
+ n_ctx: 1024
134
+ n_embd: 1024
135
+ n_head: 16
136
+ n_inner: null
137
+ n_layer: 24
138
+ n_positions: 1024
139
+ n_special: 0
140
+ no_repeat_ngram_size: 0
141
+ num_beam_groups: 1
142
+ num_beams: 1
143
+ num_return_sequences: 1
144
+ output_attentions: false
145
+ output_hidden_states: false
146
+ output_scores: false
147
+ pad_token_id: null
148
+ predict_special_tokens: true
149
+ prefix: null
150
+ problem_type: null
151
+ pruned_heads: {}
152
+ remove_invalid_values: false
153
+ reorder_and_upcast_attn: false
154
+ repetition_penalty: 1.0
155
+ resid_pdrop: 0.0
156
+ return_dict: true
157
+ return_dict_in_generate: false
158
+ scale_attn_by_inverse_layer_idx: false
159
+ scale_attn_weights: true
160
+ sep_token_id: null
161
+ summary_activation: null
162
+ summary_first_dropout: 0.0
163
+ summary_proj_to_labels: true
164
+ summary_type: cls_index
165
+ summary_use_proj: true
166
+ task_specific_params:
167
+ text-generation:
168
+ do_sample: true
169
+ max_length: 50
170
+ temperature: 1.0
171
+ tie_encoder_decoder: false
172
+ tie_word_embeddings: true
173
+ tokenizer_class: null
174
+ top_k: 50
175
+ top_p: 1.0
176
+ torch_dtype: null
177
+ torchscript: false
178
+ transformers_version: 4.17.0.dev0
179
+ use_bfloat16: false
180
+ use_cache: false
181
+ vocab_size: 50257
182
+ decoder_start_token_id:
183
+ desc: null
184
+ value: 50256
185
+ deepspeed:
186
+ desc: null
187
+ value: None
188
+ disable_tqdm:
189
+ desc: null
190
+ value: false
191
+ diversity_penalty:
192
+ desc: null
193
+ value: 0.0
194
+ do_eval:
195
+ desc: null
196
+ value: true
197
+ do_predict:
198
+ desc: null
199
+ value: false
200
+ do_sample:
201
+ desc: null
202
+ value: false
203
+ do_train:
204
+ desc: null
205
+ value: true
206
+ early_stopping:
207
+ desc: null
208
+ value: false
209
+ encoder:
210
+ desc: null
211
+ value:
212
+ _name_or_path: facebook/wav2vec2-large-lv60
213
+ activation_dropout: 0.0
214
+ adapter_kernel_size: 3
215
+ adapter_stride: 2
216
+ add_adapter: true
217
+ add_cross_attention: false
218
+ apply_spec_augment: false
219
+ architectures:
220
+ - Wav2Vec2ForPreTraining
221
+ attention_dropout: 0.0
222
+ bad_words_ids: null
223
+ bos_token_id: 1
224
+ chunk_size_feed_forward: 0
225
+ classifier_proj_size: 256
226
+ codevector_dim: 768
227
+ contrastive_logits_temperature: 0.1
228
+ conv_bias: true
229
+ conv_dim:
230
+ - 512
231
+ - 512
232
+ - 512
233
+ - 512
234
+ - 512
235
+ - 512
236
+ - 512
237
+ conv_kernel:
238
+ - 10
239
+ - 3
240
+ - 3
241
+ - 3
242
+ - 3
243
+ - 2
244
+ - 2
245
+ conv_stride:
246
+ - 5
247
+ - 2
248
+ - 2
249
+ - 2
250
+ - 2
251
+ - 2
252
+ - 2
253
+ cross_attention_hidden_size: null
254
+ ctc_loss_reduction: sum
255
+ ctc_zero_infinity: false
256
+ decoder_start_token_id: null
257
+ diversity_loss_weight: 0.1
258
+ diversity_penalty: 0.0
259
+ do_sample: false
260
+ do_stable_layer_norm: true
261
+ early_stopping: false
262
+ encoder_no_repeat_ngram_size: 0
263
+ eos_token_id: 2
264
+ feat_extract_activation: gelu
265
+ feat_extract_dropout: 0.0
266
+ feat_extract_norm: layer
267
+ feat_proj_dropout: 0.0
268
+ feat_quantizer_dropout: 0.0
269
+ final_dropout: 0.0
270
+ finetuning_task: null
271
+ forced_bos_token_id: null
272
+ forced_eos_token_id: null
273
+ gradient_checkpointing: false
274
+ hidden_act: gelu
275
+ hidden_dropout: 0.0
276
+ hidden_dropout_prob: 0.0
277
+ hidden_size: 1024
278
+ id2label:
279
+ '0': LABEL_0
280
+ '1': LABEL_1
281
+ initializer_range: 0.02
282
+ intermediate_size: 4096
283
+ is_decoder: false
284
+ is_encoder_decoder: false
285
+ label2id:
286
+ LABEL_0: 0
287
+ LABEL_1: 1
288
+ layer_norm_eps: 1.0e-05
289
+ layerdrop: 0.0
290
+ length_penalty: 1.0
291
+ mask_feature_length: 10
292
+ mask_feature_min_masks: 0
293
+ mask_feature_prob: 0.0
294
+ mask_time_length: 10
295
+ mask_time_min_masks: 2
296
+ mask_time_prob: 0.0
297
+ max_length: 20
298
+ min_length: 0
299
+ model_type: wav2vec2
300
+ no_repeat_ngram_size: 0
301
+ num_adapter_layers: 3
302
+ num_attention_heads: 16
303
+ num_beam_groups: 1
304
+ num_beams: 1
305
+ num_codevector_groups: 2
306
+ num_codevectors_per_group: 320
307
+ num_conv_pos_embedding_groups: 16
308
+ num_conv_pos_embeddings: 128
309
+ num_feat_extract_layers: 7
310
+ num_hidden_layers: 24
311
+ num_negatives: 100
312
+ num_return_sequences: 1
313
+ output_attentions: false
314
+ output_hidden_size: 1024
315
+ output_hidden_states: false
316
+ output_scores: false
317
+ pad_token_id: 0
318
+ prefix: null
319
+ problem_type: null
320
+ proj_codevector_dim: 768
321
+ pruned_heads: {}
322
+ remove_invalid_values: false
323
+ repetition_penalty: 1.0
324
+ return_dict: true
325
+ return_dict_in_generate: false
326
+ sep_token_id: null
327
+ task_specific_params: null
328
+ tdnn_dilation:
329
+ - 1
330
+ - 2
331
+ - 3
332
+ - 1
333
+ - 1
334
+ tdnn_dim:
335
+ - 512
336
+ - 512
337
+ - 512
338
+ - 512
339
+ - 1500
340
+ tdnn_kernel:
341
+ - 5
342
+ - 3
343
+ - 3
344
+ - 1
345
+ - 1
346
+ temperature: 1.0
347
+ tie_encoder_decoder: false
348
+ tie_word_embeddings: true
349
+ tokenizer_class: null
350
+ top_k: 50
351
+ top_p: 1.0
352
+ torch_dtype: null
353
+ torchscript: false
354
+ transformers_version: 4.17.0.dev0
355
+ use_bfloat16: false
356
+ use_weighted_layer_sum: false
357
+ vocab_size: 32
358
+ xvector_output_dim: 512
359
+ encoder_no_repeat_ngram_size:
360
+ desc: null
361
+ value: 0
362
+ eos_token_id:
363
+ desc: null
364
+ value: 50256
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: None
368
+ eval_batch_size:
369
+ desc: null
370
+ value: 16
371
+ eval_steps:
372
+ desc: null
373
+ value: 500
374
+ evaluation_strategy:
375
+ desc: null
376
+ value: steps
377
+ finetuning_task:
378
+ desc: null
379
+ value: null
380
+ forced_bos_token_id:
381
+ desc: null
382
+ value: null
383
+ forced_eos_token_id:
384
+ desc: null
385
+ value: null
386
+ fp16:
387
+ desc: null
388
+ value: true
389
+ fp16_backend:
390
+ desc: null
391
+ value: auto
392
+ fp16_full_eval:
393
+ desc: null
394
+ value: false
395
+ fp16_opt_level:
396
+ desc: null
397
+ value: O1
398
+ generation_max_length:
399
+ desc: null
400
+ value: 40
401
+ generation_num_beams:
402
+ desc: null
403
+ value: 1
404
+ gradient_accumulation_steps:
405
+ desc: null
406
+ value: 2
407
+ gradient_checkpointing:
408
+ desc: null
409
+ value: true
410
+ greater_is_better:
411
+ desc: null
412
+ value: None
413
+ group_by_length:
414
+ desc: null
415
+ value: true
416
+ half_precision_backend:
417
+ desc: null
418
+ value: amp
419
+ hub_model_id:
420
+ desc: null
421
+ value: None
422
+ hub_strategy:
423
+ desc: null
424
+ value: every_save
425
+ hub_token:
426
+ desc: null
427
+ value: <HUB_TOKEN>
428
+ id2label:
429
+ desc: null
430
+ value:
431
+ '0': LABEL_0
432
+ '1': LABEL_1
433
+ ignore_data_skip:
434
+ desc: null
435
+ value: false
436
+ is_decoder:
437
+ desc: null
438
+ value: false
439
+ is_encoder_decoder:
440
+ desc: null
441
+ value: true
442
+ label2id:
443
+ desc: null
444
+ value:
445
+ LABEL_0: 0
446
+ LABEL_1: 1
447
+ label_names:
448
+ desc: null
449
+ value: None
450
+ label_smoothing_factor:
451
+ desc: null
452
+ value: 0.0
453
+ learning_rate:
454
+ desc: null
455
+ value: 3.0e-05
456
+ length_column_name:
457
+ desc: null
458
+ value: input_length
459
+ length_penalty:
460
+ desc: null
461
+ value: 1.0
462
+ load_best_model_at_end:
463
+ desc: null
464
+ value: false
465
+ local_rank:
466
+ desc: null
467
+ value: -1
468
+ log_level:
469
+ desc: null
470
+ value: -1
471
+ log_level_replica:
472
+ desc: null
473
+ value: -1
474
+ log_on_each_node:
475
+ desc: null
476
+ value: true
477
+ logging_dir:
478
+ desc: null
479
+ value: ./runs/Mar01_17-09-50_sanchit--v100
480
+ logging_first_step:
481
+ desc: null
482
+ value: false
483
+ logging_nan_inf_filter:
484
+ desc: null
485
+ value: true
486
+ logging_steps:
487
+ desc: null
488
+ value: 1
489
+ logging_strategy:
490
+ desc: null
491
+ value: steps
492
+ lr_scheduler_type:
493
+ desc: null
494
+ value: linear
495
+ max_grad_norm:
496
+ desc: null
497
+ value: 1.0
498
+ max_length:
499
+ desc: null
500
+ value: 50
501
+ max_steps:
502
+ desc: null
503
+ value: -1
504
+ metric_for_best_model:
505
+ desc: null
506
+ value: None
507
+ min_length:
508
+ desc: null
509
+ value: 0
510
+ model_type:
511
+ desc: null
512
+ value: speech-encoder-decoder
513
+ mp_parameters:
514
+ desc: null
515
+ value: ''
516
+ no_cuda:
517
+ desc: null
518
+ value: false
519
+ no_repeat_ngram_size:
520
+ desc: null
521
+ value: 0
522
+ num_beam_groups:
523
+ desc: null
524
+ value: 1
525
+ num_beams:
526
+ desc: null
527
+ value: 1
528
+ num_return_sequences:
529
+ desc: null
530
+ value: 1
531
+ num_train_epochs:
532
+ desc: null
533
+ value: 1.0
534
+ optim:
535
+ desc: null
536
+ value: adamw_hf
537
+ output_attentions:
538
+ desc: null
539
+ value: false
540
+ output_dir:
541
+ desc: null
542
+ value: ./
543
+ output_hidden_states:
544
+ desc: null
545
+ value: false
546
+ output_scores:
547
+ desc: null
548
+ value: false
549
+ overwrite_output_dir:
550
+ desc: null
551
+ value: true
552
+ pad_token_id:
553
+ desc: null
554
+ value: 50256
555
+ past_index:
556
+ desc: null
557
+ value: -1
558
+ per_device_eval_batch_size:
559
+ desc: null
560
+ value: 16
561
+ per_device_train_batch_size:
562
+ desc: null
563
+ value: 16
564
+ per_gpu_eval_batch_size:
565
+ desc: null
566
+ value: None
567
+ per_gpu_train_batch_size:
568
+ desc: null
569
+ value: None
570
+ predict_with_generate:
571
+ desc: null
572
+ value: true
573
+ prediction_loss_only:
574
+ desc: null
575
+ value: false
576
+ prefix:
577
+ desc: null
578
+ value: null
579
+ problem_type:
580
+ desc: null
581
+ value: null
582
+ processor_class:
583
+ desc: null
584
+ value: Wav2Vec2Processor
585
+ pruned_heads:
586
+ desc: null
587
+ value: {}
588
+ push_to_hub:
589
+ desc: null
590
+ value: true
591
+ push_to_hub_model_id:
592
+ desc: null
593
+ value: None
594
+ push_to_hub_organization:
595
+ desc: null
596
+ value: None
597
+ push_to_hub_token:
598
+ desc: null
599
+ value: <PUSH_TO_HUB_TOKEN>
600
+ remove_invalid_values:
601
+ desc: null
602
+ value: false
603
+ remove_unused_columns:
604
+ desc: null
605
+ value: true
606
+ repetition_penalty:
607
+ desc: null
608
+ value: 1.0
609
+ report_to:
610
+ desc: null
611
+ value: '[''wandb'']'
612
+ resume_from_checkpoint:
613
+ desc: null
614
+ value: None
615
+ return_dict:
616
+ desc: null
617
+ value: true
618
+ return_dict_in_generate:
619
+ desc: null
620
+ value: false
621
+ run_name:
622
+ desc: null
623
+ value: ./
624
+ save_on_each_node:
625
+ desc: null
626
+ value: false
627
+ save_steps:
628
+ desc: null
629
+ value: 500
630
+ save_strategy:
631
+ desc: null
632
+ value: steps
633
+ save_total_limit:
634
+ desc: null
635
+ value: 1
636
+ seed:
637
+ desc: null
638
+ value: 42
639
+ sep_token_id:
640
+ desc: null
641
+ value: null
642
+ sharded_ddp:
643
+ desc: null
644
+ value: '[]'
645
+ skip_memory_metrics:
646
+ desc: null
647
+ value: true
648
+ sortish_sampler:
649
+ desc: null
650
+ value: false
651
+ task_specific_params:
652
+ desc: null
653
+ value: null
654
+ temperature:
655
+ desc: null
656
+ value: 1.0
657
+ tf32:
658
+ desc: null
659
+ value: None
660
+ tie_encoder_decoder:
661
+ desc: null
662
+ value: false
663
+ tie_word_embeddings:
664
+ desc: null
665
+ value: false
666
+ tokenizer_class:
667
+ desc: null
668
+ value: null
669
+ top_k:
670
+ desc: null
671
+ value: 50
672
+ top_p:
673
+ desc: null
674
+ value: 1.0
675
+ torch_dtype:
676
+ desc: null
677
+ value: torch.float32
678
+ torchscript:
679
+ desc: null
680
+ value: false
681
+ tpu_metrics_debug:
682
+ desc: null
683
+ value: false
684
+ tpu_num_cores:
685
+ desc: null
686
+ value: None
687
+ train_batch_size:
688
+ desc: null
689
+ value: 16
690
+ transformers_version:
691
+ desc: null
692
+ value: null
693
+ use_bfloat16:
694
+ desc: null
695
+ value: false
696
+ use_cache:
697
+ desc: null
698
+ value: false
699
+ use_legacy_prediction_loop:
700
+ desc: null
701
+ value: false
702
+ warmup_ratio:
703
+ desc: null
704
+ value: 0.0
705
+ warmup_steps:
706
+ desc: null
707
+ value: 500
708
+ weight_decay:
709
+ desc: null
710
+ value: 0.0
711
+ xpu_backend:
712
+ desc: null
713
+ value: None
wandb/run-20220301_171032-6le9s370/files/output.log ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ 0%| | 0/892 [00:00<?, ?it/s][WARNING|modeling_utils.py:388] 2022-03-01 17:10:37,650 >> Could not estimate the number of tokens of the input, floating-point operations will not be computed
4
+ Traceback (most recent call last):
5
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 539, in <module>
6
+ main()
7
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 491, in main
8
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
9
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1384, in train
10
+ tr_loss_step = self.training_step(model, inputs)
11
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1959, in training_step
12
+ loss = self.compute_loss(model, inputs)
13
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1991, in compute_loss
14
+ outputs = model(**inputs)
15
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
16
+ return forward_call(*input, **kwargs)
17
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py", line 503, in forward
18
+ encoder_outputs = self.encoder(
19
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
20
+ return forward_call(*input, **kwargs)
21
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1346, in forward
22
+ extract_features = self.feature_extractor(input_values)
23
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
24
+ return forward_call(*input, **kwargs)
25
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 514, in forward
26
+ hidden_states = conv_layer(hidden_states)
27
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
28
+ return forward_call(*input, **kwargs)
29
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 389, in forward
30
+ hidden_states = self.layer_norm(hidden_states)
31
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
32
+ return forward_call(*input, **kwargs)
33
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/normalization.py", line 189, in forward
34
+ return F.layer_norm(
35
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/functional.py", line 2347, in layer_norm
36
+ return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
37
+ RuntimeError: CUDA out of memory. Tried to allocate 1.63 GiB (GPU 0; 15.78 GiB total capacity; 10.07 GiB already allocated; 707.31 MiB free; 13.39 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
wandb/run-20220301_171032-6le9s370/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220301_171032-6le9s370/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-01T17:10:33.270278",
5
+ "startedAt": "2022-03-01T17:10:32.188812",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=16",
24
+ "--per_device_eval_batch_size=16",
25
+ "--gradient_accumulation_steps=2",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "3350b717ff05e187076e5bb0722b6b02f9837858"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220301_171032-6le9s370/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
1
+ {"_wandb": {"runtime": 8}}
wandb/run-20220301_171032-6le9s370/logs/debug-internal.log ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 17:10:33,078 INFO MainThread:245236 [internal.py:wandb_internal():89] W&B internal server running at pid: 245236, started at: 2022-03-01 17:10:33.078632
2
+ 2022-03-01 17:10:33,080 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: check_version
3
+ 2022-03-01 17:10:33,081 INFO WriterThread:245236 [datastore.py:open_for_write():77] open: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/run-6le9s370.wandb
4
+ 2022-03-01 17:10:33,082 DEBUG SenderThread:245236 [sender.py:send():235] send: header
5
+ 2022-03-01 17:10:33,082 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: check_version
6
+ 2022-03-01 17:10:33,155 DEBUG SenderThread:245236 [sender.py:send():235] send: run
7
+ 2022-03-01 17:10:33,265 INFO SenderThread:245236 [dir_watcher.py:__init__():169] watching files in: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files
8
+ 2022-03-01 17:10:33,265 INFO SenderThread:245236 [sender.py:_start_run_threads():809] run started: 6le9s370 with start time 1646154632
9
+ 2022-03-01 17:10:33,265 DEBUG SenderThread:245236 [sender.py:send():235] send: summary
10
+ 2022-03-01 17:10:33,265 INFO SenderThread:245236 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
11
+ 2022-03-01 17:10:33,266 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: run_start
12
+ 2022-03-01 17:10:33,270 DEBUG HandlerThread:245236 [meta.py:__init__():36] meta init
13
+ 2022-03-01 17:10:33,270 DEBUG HandlerThread:245236 [meta.py:__init__():50] meta init done
14
+ 2022-03-01 17:10:33,270 DEBUG HandlerThread:245236 [meta.py:probe():210] probe
15
+ 2022-03-01 17:10:33,276 DEBUG HandlerThread:245236 [meta.py:_setup_git():200] setup git
16
+ 2022-03-01 17:10:33,290 DEBUG HandlerThread:245236 [meta.py:_setup_git():207] setup git done
17
+ 2022-03-01 17:10:33,290 DEBUG HandlerThread:245236 [meta.py:_save_pip():54] save pip
18
+ 2022-03-01 17:10:33,291 DEBUG HandlerThread:245236 [meta.py:_save_pip():68] save pip done
19
+ 2022-03-01 17:10:33,291 DEBUG HandlerThread:245236 [meta.py:probe():248] probe done
20
+ 2022-03-01 17:10:33,365 DEBUG SenderThread:245236 [sender.py:send():235] send: files
21
+ 2022-03-01 17:10:33,365 INFO SenderThread:245236 [sender.py:_save_file():944] saving file wandb-metadata.json with policy now
22
+ 2022-03-01 17:10:33,370 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: stop_status
23
+ 2022-03-01 17:10:33,370 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: stop_status
24
+ 2022-03-01 17:10:33,410 DEBUG SenderThread:245236 [sender.py:send():235] send: config
25
+ 2022-03-01 17:10:33,411 DEBUG SenderThread:245236 [sender.py:send():235] send: metric
26
+ 2022-03-01 17:10:33,411 DEBUG SenderThread:245236 [sender.py:send():235] send: metric
27
+ 2022-03-01 17:10:33,411 WARNING SenderThread:245236 [sender.py:send_metric():902] Seen metric with glob (shouldnt happen)
28
+ 2022-03-01 17:10:33,654 INFO Thread-11 :245236 [upload_job.py:push():137] Uploaded file /tmp/tmpimoy9n47wandb/3840sgse-wandb-metadata.json
29
+ 2022-03-01 17:10:34,267 INFO Thread-8 :245236 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/wandb-summary.json
30
+ 2022-03-01 17:10:34,267 INFO Thread-8 :245236 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/requirements.txt
31
+ 2022-03-01 17:10:34,268 INFO Thread-8 :245236 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/output.log
32
+ 2022-03-01 17:10:34,268 INFO Thread-8 :245236 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/wandb-metadata.json
33
+ 2022-03-01 17:10:36,266 INFO Thread-8 :245236 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/output.log
34
+ 2022-03-01 17:10:40,268 INFO Thread-8 :245236 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/output.log
35
+ 2022-03-01 17:10:42,126 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: poll_exit
36
+ 2022-03-01 17:10:42,126 DEBUG SenderThread:245236 [sender.py:send():235] send: telemetry
37
+ 2022-03-01 17:10:42,127 DEBUG SenderThread:245236 [sender.py:send():235] send: exit
38
+ 2022-03-01 17:10:42,127 INFO SenderThread:245236 [sender.py:send_exit():371] handling exit code: 1
39
+ 2022-03-01 17:10:42,127 INFO SenderThread:245236 [sender.py:send_exit():373] handling runtime: 8
40
+ 2022-03-01 17:10:42,127 INFO SenderThread:245236 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
41
+ 2022-03-01 17:10:42,127 INFO SenderThread:245236 [sender.py:send_exit():379] send defer
42
+ 2022-03-01 17:10:42,127 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: poll_exit
43
+ 2022-03-01 17:10:42,128 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
44
+ 2022-03-01 17:10:42,128 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 0
45
+ 2022-03-01 17:10:42,128 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
46
+ 2022-03-01 17:10:42,128 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 0
47
+ 2022-03-01 17:10:42,128 INFO SenderThread:245236 [sender.py:transition_state():392] send defer: 1
48
+ 2022-03-01 17:10:42,128 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
49
+ 2022-03-01 17:10:42,128 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 1
50
+ 2022-03-01 17:10:42,190 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
51
+ 2022-03-01 17:10:42,190 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 1
52
+ 2022-03-01 17:10:42,190 INFO SenderThread:245236 [sender.py:transition_state():392] send defer: 2
53
+ 2022-03-01 17:10:42,190 DEBUG SenderThread:245236 [sender.py:send():235] send: stats
54
+ 2022-03-01 17:10:42,191 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
55
+ 2022-03-01 17:10:42,191 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 2
56
+ 2022-03-01 17:10:42,191 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
57
+ 2022-03-01 17:10:42,191 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 2
58
+ 2022-03-01 17:10:42,191 INFO SenderThread:245236 [sender.py:transition_state():392] send defer: 3
59
+ 2022-03-01 17:10:42,191 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
60
+ 2022-03-01 17:10:42,191 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 3
61
+ 2022-03-01 17:10:42,191 DEBUG SenderThread:245236 [sender.py:send():235] send: summary
62
+ 2022-03-01 17:10:42,192 INFO SenderThread:245236 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
63
+ 2022-03-01 17:10:42,192 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
64
+ 2022-03-01 17:10:42,192 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 3
65
+ 2022-03-01 17:10:42,192 INFO SenderThread:245236 [sender.py:transition_state():392] send defer: 4
66
+ 2022-03-01 17:10:42,192 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
67
+ 2022-03-01 17:10:42,192 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 4
68
+ 2022-03-01 17:10:42,192 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
69
+ 2022-03-01 17:10:42,193 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 4
70
+ 2022-03-01 17:10:42,229 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: poll_exit
71
+ 2022-03-01 17:10:42,301 INFO Thread-8 :245236 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/wandb-summary.json
72
+ 2022-03-01 17:10:42,301 INFO Thread-8 :245236 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/output.log
73
+ 2022-03-01 17:10:42,302 INFO SenderThread:245236 [sender.py:transition_state():392] send defer: 5
74
+ 2022-03-01 17:10:42,302 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: poll_exit
75
+ 2022-03-01 17:10:42,302 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
76
+ 2022-03-01 17:10:42,302 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 5
77
+ 2022-03-01 17:10:42,303 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
78
+ 2022-03-01 17:10:42,303 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 5
79
+ 2022-03-01 17:10:42,303 INFO SenderThread:245236 [dir_watcher.py:finish():283] shutting down directory watcher
80
+ 2022-03-01 17:10:42,403 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: poll_exit
81
+ 2022-03-01 17:10:43,302 INFO Thread-8 :245236 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/config.yaml
82
+ 2022-03-01 17:10:43,302 INFO SenderThread:245236 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files
83
+ 2022-03-01 17:10:43,303 INFO SenderThread:245236 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/wandb-metadata.json wandb-metadata.json
84
+ 2022-03-01 17:10:43,303 INFO SenderThread:245236 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/output.log output.log
85
+ 2022-03-01 17:10:43,303 INFO SenderThread:245236 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/wandb-summary.json wandb-summary.json
86
+ 2022-03-01 17:10:43,303 INFO SenderThread:245236 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/requirements.txt requirements.txt
87
+ 2022-03-01 17:10:43,303 INFO SenderThread:245236 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/config.yaml config.yaml
88
+ 2022-03-01 17:10:43,312 INFO SenderThread:245236 [sender.py:transition_state():392] send defer: 6
89
+ 2022-03-01 17:10:43,312 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: poll_exit
90
+ 2022-03-01 17:10:43,316 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
91
+ 2022-03-01 17:10:43,316 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 6
92
+ 2022-03-01 17:10:43,316 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
93
+ 2022-03-01 17:10:43,316 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 6
94
+ 2022-03-01 17:10:43,317 INFO SenderThread:245236 [file_pusher.py:finish():177] shutting down file pusher
95
+ 2022-03-01 17:10:43,414 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: poll_exit
96
+ 2022-03-01 17:10:43,414 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: poll_exit
97
+ 2022-03-01 17:10:43,516 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: poll_exit
98
+ 2022-03-01 17:10:43,516 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: poll_exit
99
+ 2022-03-01 17:10:43,561 INFO Thread-12 :245236 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/output.log
100
+ 2022-03-01 17:10:43,580 INFO Thread-13 :245236 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/wandb-summary.json
101
+ 2022-03-01 17:10:43,581 INFO Thread-14 :245236 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/requirements.txt
102
+ 2022-03-01 17:10:43,598 INFO Thread-15 :245236 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/files/config.yaml
103
+ 2022-03-01 17:10:43,617 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: poll_exit
104
+ 2022-03-01 17:10:43,617 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: poll_exit
105
+ 2022-03-01 17:10:43,719 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: poll_exit
106
+ 2022-03-01 17:10:43,719 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: poll_exit
107
+ 2022-03-01 17:10:43,799 INFO Thread-7 :245236 [sender.py:transition_state():392] send defer: 7
108
+ 2022-03-01 17:10:43,799 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
109
+ 2022-03-01 17:10:43,799 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 7
110
+ 2022-03-01 17:10:43,799 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
111
+ 2022-03-01 17:10:43,799 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 7
112
+ 2022-03-01 17:10:43,820 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: poll_exit
113
+ 2022-03-01 17:10:43,839 INFO SenderThread:245236 [sender.py:transition_state():392] send defer: 8
114
+ 2022-03-01 17:10:43,840 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: poll_exit
115
+ 2022-03-01 17:10:43,840 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
116
+ 2022-03-01 17:10:43,840 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 8
117
+ 2022-03-01 17:10:43,840 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
118
+ 2022-03-01 17:10:43,840 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 8
119
+ 2022-03-01 17:10:43,841 INFO SenderThread:245236 [sender.py:transition_state():392] send defer: 9
120
+ 2022-03-01 17:10:43,841 DEBUG SenderThread:245236 [sender.py:send():235] send: final
121
+ 2022-03-01 17:10:43,841 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: defer
122
+ 2022-03-01 17:10:43,842 DEBUG SenderThread:245236 [sender.py:send():235] send: footer
123
+ 2022-03-01 17:10:43,842 INFO HandlerThread:245236 [handler.py:handle_request_defer():154] handle defer: 9
124
+ 2022-03-01 17:10:43,842 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: defer
125
+ 2022-03-01 17:10:43,842 INFO SenderThread:245236 [sender.py:send_request_defer():388] handle sender defer: 9
126
+ 2022-03-01 17:10:43,941 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: poll_exit
127
+ 2022-03-01 17:10:43,942 DEBUG SenderThread:245236 [sender.py:send_request():249] send_request: poll_exit
128
+ 2022-03-01 17:10:43,942 INFO SenderThread:245236 [file_pusher.py:join():182] waiting for file pusher
129
+ 2022-03-01 17:10:43,995 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: get_summary
130
+ 2022-03-01 17:10:43,996 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: sampled_history
131
+ 2022-03-01 17:10:43,997 DEBUG HandlerThread:245236 [handler.py:handle_request():131] handle_request: shutdown
132
+ 2022-03-01 17:10:43,997 INFO HandlerThread:245236 [handler.py:finish():739] shutting down handler
133
+ 2022-03-01 17:10:44,842 INFO WriterThread:245236 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/run-6le9s370.wandb
134
+ 2022-03-01 17:10:44,994 INFO SenderThread:245236 [sender.py:finish():1075] shutting down sender
135
+ 2022-03-01 17:10:44,995 INFO SenderThread:245236 [file_pusher.py:finish():177] shutting down file pusher
136
+ 2022-03-01 17:10:44,995 INFO SenderThread:245236 [file_pusher.py:join():182] waiting for file pusher
137
+ 2022-03-01 17:10:44,997 INFO MainThread:245236 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220301_171032-6le9s370/logs/debug.log ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 17:10:32,190 INFO MainThread:245136 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-03-01 17:10:32,190 INFO MainThread:245136 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/settings
3
+ 2022-03-01 17:10:32,190 INFO MainThread:245136 [wandb_setup.py:_flush():75] Loading settings from environment variables: {}
4
+ 2022-03-01 17:10:32,190 INFO MainThread:245136 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_speech_recognition_seq2seq.py', 'program': '/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py'}
5
+ 2022-03-01 17:10:32,190 INFO MainThread:245136 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/logs/debug.log
6
+ 2022-03-01 17:10:32,190 INFO MainThread:245136 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171032-6le9s370/logs/debug-internal.log
7
+ 2022-03-01 17:10:32,190 INFO MainThread:245136 [wandb_init.py:init():420] calling init triggers
8
+ 2022-03-01 17:10:32,190 INFO MainThread:245136 [wandb_init.py:init():425] wandb.init called with sweep_config: {}
9
+ config: {}
10
+ 2022-03-01 17:10:32,191 INFO MainThread:245136 [wandb_init.py:init():471] starting backend
11
+ 2022-03-01 17:10:32,191 INFO MainThread:245136 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-03-01 17:10:32,250 INFO MainThread:245136 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-03-01 17:10:32,309 INFO MainThread:245136 [backend.py:ensure_launched():224] started backend process with pid: 245236
14
+ 2022-03-01 17:10:32,311 INFO MainThread:245136 [wandb_init.py:init():480] backend started and connected
15
+ 2022-03-01 17:10:32,321 INFO MainThread:245136 [wandb_init.py:init():550] updated telemetry
16
+ 2022-03-01 17:10:32,461 INFO MainThread:245136 [wandb_init.py:init():581] communicating current version
17
+ 2022-03-01 17:10:33,154 INFO MainThread:245136 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.11 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
18
+
19
+ 2022-03-01 17:10:33,154 INFO MainThread:245136 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
20
+ 2022-03-01 17:10:33,265 INFO MainThread:245136 [wandb_init.py:init():624] starting run threads in backend
21
+ 2022-03-01 17:10:33,369 INFO MainThread:245136 [wandb_run.py:_console_start():1827] atexit reg
22
+ 2022-03-01 17:10:33,370 INFO MainThread:245136 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
23
+ 2022-03-01 17:10:33,370 INFO MainThread:245136 [wandb_run.py:_redirect():1706] Redirecting console.
24
+ 2022-03-01 17:10:33,372 INFO MainThread:245136 [wandb_run.py:_redirect():1762] Redirects installed.
25
+ 2022-03-01 17:10:33,372 INFO MainThread:245136 [wandb_init.py:init():651] run started, returning control to user process
26
+ 2022-03-01 17:10:33,375 INFO MainThread:245136 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 50, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 50256, 'eos_token_id': 50256, 'sep_token_id': None, 'decoder_start_token_id': 50256, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50257, 'n_positions': 1024, 'n_embd': 1024, 'n_layer': 24, 'n_head': 16, 'n_inner': None, 'activation_function': 'gelu_new', 'resid_pdrop': 0.0, 'embd_pdrop': 0.0, 'attn_pdrop': 0.0, 'layer_norm_epsilon': 1e-05, 'initializer_range': 0.02, 'summary_type': 'cls_index', 'summary_use_proj': True, 'summary_activation': None, 'summary_first_dropout': 0.0, 'summary_proj_to_labels': True, 'scale_attn_weights': True, 'use_cache': False, 'scale_attn_by_inverse_layer_idx': False, 'reorder_and_upcast_attn': False, 'bos_token_id': 50256, 'eos_token_id': 50256, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['GPT2LMHeadModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'pad_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': {'text-generation': {'do_sample': True, 'max_length': 50}}, 'problem_type': None, '_name_or_path': 'gpt2-medium', 'transformers_version': '4.17.0.dev0', 'n_ctx': 1024, 'n_special': 0, 'predict_special_tokens': True, 'model_type': 'gpt2'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-large-lv60', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'hidden_dropout_prob': 0.0, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.0, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 'None', 'learning_rate': 3e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Mar01_17-09-50_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 1, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': False, 'metric_for_best_model': 'None', 'greater_is_better': 'None', 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 40, 'generation_num_beams': 1, 'train_batch_size': 16, 'eval_batch_size': 16}
27
+ 2022-03-01 17:10:33,378 INFO MainThread:245136 [wandb_watch.py:watch():43] Watching
28
+ 2022-03-01 17:10:39,684 INFO MainThread:245136 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
29
+ 2022-03-01 17:10:39,686 INFO MainThread:245136 [wandb_run.py:_restore():1769] restore
30
+ 2022-03-01 17:10:42,128 INFO MainThread:245136 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
31
+ wandb_count: 1
32
+ }
33
+ pusher_stats {
34
+ uploaded_bytes: 2095
35
+ total_bytes: 2095
36
+ }
37
+
38
+ 2022-03-01 17:10:42,303 INFO MainThread:245136 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
39
+ wandb_count: 1
40
+ }
41
+ pusher_stats {
42
+ uploaded_bytes: 2095
43
+ total_bytes: 2095
44
+ }
45
+
46
+ 2022-03-01 17:10:43,313 INFO MainThread:245136 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
47
+ wandb_count: 4
48
+ }
49
+ pusher_stats {
50
+ uploaded_bytes: 2095
51
+ total_bytes: 8718
52
+ }
53
+
54
+ 2022-03-01 17:10:43,415 INFO MainThread:245136 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
55
+ wandb_count: 5
56
+ }
57
+ pusher_stats {
58
+ uploaded_bytes: 21067
59
+ total_bytes: 21067
60
+ }
61
+
62
+ 2022-03-01 17:10:43,516 INFO MainThread:245136 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
63
+ wandb_count: 5
64
+ }
65
+ pusher_stats {
66
+ uploaded_bytes: 21067
67
+ total_bytes: 21067
68
+ }
69
+
70
+ 2022-03-01 17:10:43,618 INFO MainThread:245136 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
71
+ wandb_count: 5
72
+ }
73
+ pusher_stats {
74
+ uploaded_bytes: 21067
75
+ total_bytes: 21067
76
+ }
77
+
78
+ 2022-03-01 17:10:43,719 INFO MainThread:245136 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
79
+ wandb_count: 5
80
+ }
81
+ pusher_stats {
82
+ uploaded_bytes: 21067
83
+ total_bytes: 21067
84
+ }
85
+
86
+ 2022-03-01 17:10:43,841 INFO MainThread:245136 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
87
+ wandb_count: 5
88
+ }
89
+ pusher_stats {
90
+ uploaded_bytes: 21067
91
+ total_bytes: 21067
92
+ }
93
+
94
+ 2022-03-01 17:10:43,995 INFO MainThread:245136 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
95
+ exit_result {
96
+ }
97
+ file_counts {
98
+ wandb_count: 5
99
+ }
100
+ pusher_stats {
101
+ uploaded_bytes: 21067
102
+ total_bytes: 21067
103
+ }
104
+ local_info {
105
+ }
106
+
107
+ 2022-03-01 17:10:45,126 INFO MainThread:245136 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220301_171032-6le9s370/run-6le9s370.wandb ADDED
Binary file (13.7 kB). View file
wandb/run-20220301_171200-3h52ha0p/files/config.yaml ADDED
@@ -0,0 +1,713 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: ./
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.10
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.9.5
22
+ start_time: 1646154720
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.9.5
37
+ 5: 0.12.10
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ adafactor:
42
+ desc: null
43
+ value: false
44
+ adam_beta1:
45
+ desc: null
46
+ value: 0.9
47
+ adam_beta2:
48
+ desc: null
49
+ value: 0.999
50
+ adam_epsilon:
51
+ desc: null
52
+ value: 1.0e-08
53
+ add_cross_attention:
54
+ desc: null
55
+ value: false
56
+ architectures:
57
+ desc: null
58
+ value:
59
+ - SpeechEncoderDecoderModel
60
+ bad_words_ids:
61
+ desc: null
62
+ value: null
63
+ bf16:
64
+ desc: null
65
+ value: false
66
+ bf16_full_eval:
67
+ desc: null
68
+ value: false
69
+ bos_token_id:
70
+ desc: null
71
+ value: null
72
+ chunk_size_feed_forward:
73
+ desc: null
74
+ value: 0
75
+ cross_attention_hidden_size:
76
+ desc: null
77
+ value: null
78
+ dataloader_drop_last:
79
+ desc: null
80
+ value: false
81
+ dataloader_num_workers:
82
+ desc: null
83
+ value: 0
84
+ dataloader_pin_memory:
85
+ desc: null
86
+ value: true
87
+ ddp_bucket_cap_mb:
88
+ desc: null
89
+ value: None
90
+ ddp_find_unused_parameters:
91
+ desc: null
92
+ value: None
93
+ debug:
94
+ desc: null
95
+ value: '[]'
96
+ decoder:
97
+ desc: null
98
+ value:
99
+ _name_or_path: gpt2-medium
100
+ activation_function: gelu_new
101
+ add_cross_attention: true
102
+ architectures:
103
+ - GPT2LMHeadModel
104
+ attn_pdrop: 0.0
105
+ bad_words_ids: null
106
+ bos_token_id: 50256
107
+ chunk_size_feed_forward: 0
108
+ cross_attention_hidden_size: null
109
+ decoder_start_token_id: null
110
+ diversity_penalty: 0.0
111
+ do_sample: false
112
+ early_stopping: false
113
+ embd_pdrop: 0.0
114
+ encoder_no_repeat_ngram_size: 0
115
+ eos_token_id: 50256
116
+ finetuning_task: null
117
+ forced_bos_token_id: null
118
+ forced_eos_token_id: null
119
+ id2label:
120
+ '0': LABEL_0
121
+ '1': LABEL_1
122
+ initializer_range: 0.02
123
+ is_decoder: true
124
+ is_encoder_decoder: false
125
+ label2id:
126
+ LABEL_0: 0
127
+ LABEL_1: 1
128
+ layer_norm_epsilon: 1.0e-05
129
+ length_penalty: 1.0
130
+ max_length: 20
131
+ min_length: 0
132
+ model_type: gpt2
133
+ n_ctx: 1024
134
+ n_embd: 1024
135
+ n_head: 16
136
+ n_inner: null
137
+ n_layer: 24
138
+ n_positions: 1024
139
+ n_special: 0
140
+ no_repeat_ngram_size: 0
141
+ num_beam_groups: 1
142
+ num_beams: 1
143
+ num_return_sequences: 1
144
+ output_attentions: false
145
+ output_hidden_states: false
146
+ output_scores: false
147
+ pad_token_id: null
148
+ predict_special_tokens: true
149
+ prefix: null
150
+ problem_type: null
151
+ pruned_heads: {}
152
+ remove_invalid_values: false
153
+ reorder_and_upcast_attn: false
154
+ repetition_penalty: 1.0
155
+ resid_pdrop: 0.0
156
+ return_dict: true
157
+ return_dict_in_generate: false
158
+ scale_attn_by_inverse_layer_idx: false
159
+ scale_attn_weights: true
160
+ sep_token_id: null
161
+ summary_activation: null
162
+ summary_first_dropout: 0.0
163
+ summary_proj_to_labels: true
164
+ summary_type: cls_index
165
+ summary_use_proj: true
166
+ task_specific_params:
167
+ text-generation:
168
+ do_sample: true
169
+ max_length: 50
170
+ temperature: 1.0
171
+ tie_encoder_decoder: false
172
+ tie_word_embeddings: true
173
+ tokenizer_class: null
174
+ top_k: 50
175
+ top_p: 1.0
176
+ torch_dtype: null
177
+ torchscript: false
178
+ transformers_version: 4.17.0.dev0
179
+ use_bfloat16: false
180
+ use_cache: false
181
+ vocab_size: 50257
182
+ decoder_start_token_id:
183
+ desc: null
184
+ value: 50256
185
+ deepspeed:
186
+ desc: null
187
+ value: None
188
+ disable_tqdm:
189
+ desc: null
190
+ value: false
191
+ diversity_penalty:
192
+ desc: null
193
+ value: 0.0
194
+ do_eval:
195
+ desc: null
196
+ value: true
197
+ do_predict:
198
+ desc: null
199
+ value: false
200
+ do_sample:
201
+ desc: null
202
+ value: false
203
+ do_train:
204
+ desc: null
205
+ value: true
206
+ early_stopping:
207
+ desc: null
208
+ value: false
209
+ encoder:
210
+ desc: null
211
+ value:
212
+ _name_or_path: facebook/wav2vec2-large-lv60
213
+ activation_dropout: 0.0
214
+ adapter_kernel_size: 3
215
+ adapter_stride: 2
216
+ add_adapter: true
217
+ add_cross_attention: false
218
+ apply_spec_augment: false
219
+ architectures:
220
+ - Wav2Vec2ForPreTraining
221
+ attention_dropout: 0.0
222
+ bad_words_ids: null
223
+ bos_token_id: 1
224
+ chunk_size_feed_forward: 0
225
+ classifier_proj_size: 256
226
+ codevector_dim: 768
227
+ contrastive_logits_temperature: 0.1
228
+ conv_bias: true
229
+ conv_dim:
230
+ - 512
231
+ - 512
232
+ - 512
233
+ - 512
234
+ - 512
235
+ - 512
236
+ - 512
237
+ conv_kernel:
238
+ - 10
239
+ - 3
240
+ - 3
241
+ - 3
242
+ - 3
243
+ - 2
244
+ - 2
245
+ conv_stride:
246
+ - 5
247
+ - 2
248
+ - 2
249
+ - 2
250
+ - 2
251
+ - 2
252
+ - 2
253
+ cross_attention_hidden_size: null
254
+ ctc_loss_reduction: sum
255
+ ctc_zero_infinity: false
256
+ decoder_start_token_id: null
257
+ diversity_loss_weight: 0.1
258
+ diversity_penalty: 0.0
259
+ do_sample: false
260
+ do_stable_layer_norm: true
261
+ early_stopping: false
262
+ encoder_no_repeat_ngram_size: 0
263
+ eos_token_id: 2
264
+ feat_extract_activation: gelu
265
+ feat_extract_dropout: 0.0
266
+ feat_extract_norm: layer
267
+ feat_proj_dropout: 0.0
268
+ feat_quantizer_dropout: 0.0
269
+ final_dropout: 0.0
270
+ finetuning_task: null
271
+ forced_bos_token_id: null
272
+ forced_eos_token_id: null
273
+ gradient_checkpointing: false
274
+ hidden_act: gelu
275
+ hidden_dropout: 0.0
276
+ hidden_dropout_prob: 0.0
277
+ hidden_size: 1024
278
+ id2label:
279
+ '0': LABEL_0
280
+ '1': LABEL_1
281
+ initializer_range: 0.02
282
+ intermediate_size: 4096
283
+ is_decoder: false
284
+ is_encoder_decoder: false
285
+ label2id:
286
+ LABEL_0: 0
287
+ LABEL_1: 1
288
+ layer_norm_eps: 1.0e-05
289
+ layerdrop: 0.0
290
+ length_penalty: 1.0
291
+ mask_feature_length: 10
292
+ mask_feature_min_masks: 0
293
+ mask_feature_prob: 0.0
294
+ mask_time_length: 10
295
+ mask_time_min_masks: 2
296
+ mask_time_prob: 0.0
297
+ max_length: 20
298
+ min_length: 0
299
+ model_type: wav2vec2
300
+ no_repeat_ngram_size: 0
301
+ num_adapter_layers: 3
302
+ num_attention_heads: 16
303
+ num_beam_groups: 1
304
+ num_beams: 1
305
+ num_codevector_groups: 2
306
+ num_codevectors_per_group: 320
307
+ num_conv_pos_embedding_groups: 16
308
+ num_conv_pos_embeddings: 128
309
+ num_feat_extract_layers: 7
310
+ num_hidden_layers: 24
311
+ num_negatives: 100
312
+ num_return_sequences: 1
313
+ output_attentions: false
314
+ output_hidden_size: 1024
315
+ output_hidden_states: false
316
+ output_scores: false
317
+ pad_token_id: 0
318
+ prefix: null
319
+ problem_type: null
320
+ proj_codevector_dim: 768
321
+ pruned_heads: {}
322
+ remove_invalid_values: false
323
+ repetition_penalty: 1.0
324
+ return_dict: true
325
+ return_dict_in_generate: false
326
+ sep_token_id: null
327
+ task_specific_params: null
328
+ tdnn_dilation:
329
+ - 1
330
+ - 2
331
+ - 3
332
+ - 1
333
+ - 1
334
+ tdnn_dim:
335
+ - 512
336
+ - 512
337
+ - 512
338
+ - 512
339
+ - 1500
340
+ tdnn_kernel:
341
+ - 5
342
+ - 3
343
+ - 3
344
+ - 1
345
+ - 1
346
+ temperature: 1.0
347
+ tie_encoder_decoder: false
348
+ tie_word_embeddings: true
349
+ tokenizer_class: null
350
+ top_k: 50
351
+ top_p: 1.0
352
+ torch_dtype: null
353
+ torchscript: false
354
+ transformers_version: 4.17.0.dev0
355
+ use_bfloat16: false
356
+ use_weighted_layer_sum: false
357
+ vocab_size: 32
358
+ xvector_output_dim: 512
359
+ encoder_no_repeat_ngram_size:
360
+ desc: null
361
+ value: 0
362
+ eos_token_id:
363
+ desc: null
364
+ value: 50256
365
+ eval_accumulation_steps:
366
+ desc: null
367
+ value: None
368
+ eval_batch_size:
369
+ desc: null
370
+ value: 16
371
+ eval_steps:
372
+ desc: null
373
+ value: 500
374
+ evaluation_strategy:
375
+ desc: null
376
+ value: steps
377
+ finetuning_task:
378
+ desc: null
379
+ value: null
380
+ forced_bos_token_id:
381
+ desc: null
382
+ value: null
383
+ forced_eos_token_id:
384
+ desc: null
385
+ value: null
386
+ fp16:
387
+ desc: null
388
+ value: true
389
+ fp16_backend:
390
+ desc: null
391
+ value: auto
392
+ fp16_full_eval:
393
+ desc: null
394
+ value: false
395
+ fp16_opt_level:
396
+ desc: null
397
+ value: O1
398
+ generation_max_length:
399
+ desc: null
400
+ value: 40
401
+ generation_num_beams:
402
+ desc: null
403
+ value: 1
404
+ gradient_accumulation_steps:
405
+ desc: null
406
+ value: 4
407
+ gradient_checkpointing:
408
+ desc: null
409
+ value: true
410
+ greater_is_better:
411
+ desc: null
412
+ value: None
413
+ group_by_length:
414
+ desc: null
415
+ value: true
416
+ half_precision_backend:
417
+ desc: null
418
+ value: amp
419
+ hub_model_id:
420
+ desc: null
421
+ value: None
422
+ hub_strategy:
423
+ desc: null
424
+ value: every_save
425
+ hub_token:
426
+ desc: null
427
+ value: <HUB_TOKEN>
428
+ id2label:
429
+ desc: null
430
+ value:
431
+ '0': LABEL_0
432
+ '1': LABEL_1
433
+ ignore_data_skip:
434
+ desc: null
435
+ value: false
436
+ is_decoder:
437
+ desc: null
438
+ value: false
439
+ is_encoder_decoder:
440
+ desc: null
441
+ value: true
442
+ label2id:
443
+ desc: null
444
+ value:
445
+ LABEL_0: 0
446
+ LABEL_1: 1
447
+ label_names:
448
+ desc: null
449
+ value: None
450
+ label_smoothing_factor:
451
+ desc: null
452
+ value: 0.0
453
+ learning_rate:
454
+ desc: null
455
+ value: 3.0e-05
456
+ length_column_name:
457
+ desc: null
458
+ value: input_length
459
+ length_penalty:
460
+ desc: null
461
+ value: 1.0
462
+ load_best_model_at_end:
463
+ desc: null
464
+ value: false
465
+ local_rank:
466
+ desc: null
467
+ value: -1
468
+ log_level:
469
+ desc: null
470
+ value: -1
471
+ log_level_replica:
472
+ desc: null
473
+ value: -1
474
+ log_on_each_node:
475
+ desc: null
476
+ value: true
477
+ logging_dir:
478
+ desc: null
479
+ value: ./runs/Mar01_17-11-19_sanchit--v100
480
+ logging_first_step:
481
+ desc: null
482
+ value: false
483
+ logging_nan_inf_filter:
484
+ desc: null
485
+ value: true
486
+ logging_steps:
487
+ desc: null
488
+ value: 1
489
+ logging_strategy:
490
+ desc: null
491
+ value: steps
492
+ lr_scheduler_type:
493
+ desc: null
494
+ value: linear
495
+ max_grad_norm:
496
+ desc: null
497
+ value: 1.0
498
+ max_length:
499
+ desc: null
500
+ value: 50
501
+ max_steps:
502
+ desc: null
503
+ value: -1
504
+ metric_for_best_model:
505
+ desc: null
506
+ value: None
507
+ min_length:
508
+ desc: null
509
+ value: 0
510
+ model_type:
511
+ desc: null
512
+ value: speech-encoder-decoder
513
+ mp_parameters:
514
+ desc: null
515
+ value: ''
516
+ no_cuda:
517
+ desc: null
518
+ value: false
519
+ no_repeat_ngram_size:
520
+ desc: null
521
+ value: 0
522
+ num_beam_groups:
523
+ desc: null
524
+ value: 1
525
+ num_beams:
526
+ desc: null
527
+ value: 1
528
+ num_return_sequences:
529
+ desc: null
530
+ value: 1
531
+ num_train_epochs:
532
+ desc: null
533
+ value: 1.0
534
+ optim:
535
+ desc: null
536
+ value: adamw_hf
537
+ output_attentions:
538
+ desc: null
539
+ value: false
540
+ output_dir:
541
+ desc: null
542
+ value: ./
543
+ output_hidden_states:
544
+ desc: null
545
+ value: false
546
+ output_scores:
547
+ desc: null
548
+ value: false
549
+ overwrite_output_dir:
550
+ desc: null
551
+ value: true
552
+ pad_token_id:
553
+ desc: null
554
+ value: 50256
555
+ past_index:
556
+ desc: null
557
+ value: -1
558
+ per_device_eval_batch_size:
559
+ desc: null
560
+ value: 16
561
+ per_device_train_batch_size:
562
+ desc: null
563
+ value: 16
564
+ per_gpu_eval_batch_size:
565
+ desc: null
566
+ value: None
567
+ per_gpu_train_batch_size:
568
+ desc: null
569
+ value: None
570
+ predict_with_generate:
571
+ desc: null
572
+ value: true
573
+ prediction_loss_only:
574
+ desc: null
575
+ value: false
576
+ prefix:
577
+ desc: null
578
+ value: null
579
+ problem_type:
580
+ desc: null
581
+ value: null
582
+ processor_class:
583
+ desc: null
584
+ value: Wav2Vec2Processor
585
+ pruned_heads:
586
+ desc: null
587
+ value: {}
588
+ push_to_hub:
589
+ desc: null
590
+ value: true
591
+ push_to_hub_model_id:
592
+ desc: null
593
+ value: None
594
+ push_to_hub_organization:
595
+ desc: null
596
+ value: None
597
+ push_to_hub_token:
598
+ desc: null
599
+ value: <PUSH_TO_HUB_TOKEN>
600
+ remove_invalid_values:
601
+ desc: null
602
+ value: false
603
+ remove_unused_columns:
604
+ desc: null
605
+ value: true
606
+ repetition_penalty:
607
+ desc: null
608
+ value: 1.0
609
+ report_to:
610
+ desc: null
611
+ value: '[''wandb'']'
612
+ resume_from_checkpoint:
613
+ desc: null
614
+ value: None
615
+ return_dict:
616
+ desc: null
617
+ value: true
618
+ return_dict_in_generate:
619
+ desc: null
620
+ value: false
621
+ run_name:
622
+ desc: null
623
+ value: ./
624
+ save_on_each_node:
625
+ desc: null
626
+ value: false
627
+ save_steps:
628
+ desc: null
629
+ value: 500
630
+ save_strategy:
631
+ desc: null
632
+ value: steps
633
+ save_total_limit:
634
+ desc: null
635
+ value: 1
636
+ seed:
637
+ desc: null
638
+ value: 42
639
+ sep_token_id:
640
+ desc: null
641
+ value: null
642
+ sharded_ddp:
643
+ desc: null
644
+ value: '[]'
645
+ skip_memory_metrics:
646
+ desc: null
647
+ value: true
648
+ sortish_sampler:
649
+ desc: null
650
+ value: false
651
+ task_specific_params:
652
+ desc: null
653
+ value: null
654
+ temperature:
655
+ desc: null
656
+ value: 1.0
657
+ tf32:
658
+ desc: null
659
+ value: None
660
+ tie_encoder_decoder:
661
+ desc: null
662
+ value: false
663
+ tie_word_embeddings:
664
+ desc: null
665
+ value: false
666
+ tokenizer_class:
667
+ desc: null
668
+ value: null
669
+ top_k:
670
+ desc: null
671
+ value: 50
672
+ top_p:
673
+ desc: null
674
+ value: 1.0
675
+ torch_dtype:
676
+ desc: null
677
+ value: torch.float32
678
+ torchscript:
679
+ desc: null
680
+ value: false
681
+ tpu_metrics_debug:
682
+ desc: null
683
+ value: false
684
+ tpu_num_cores:
685
+ desc: null
686
+ value: None
687
+ train_batch_size:
688
+ desc: null
689
+ value: 16
690
+ transformers_version:
691
+ desc: null
692
+ value: null
693
+ use_bfloat16:
694
+ desc: null
695
+ value: false
696
+ use_cache:
697
+ desc: null
698
+ value: false
699
+ use_legacy_prediction_loop:
700
+ desc: null
701
+ value: false
702
+ warmup_ratio:
703
+ desc: null
704
+ value: 0.0
705
+ warmup_steps:
706
+ desc: null
707
+ value: 500
708
+ weight_decay:
709
+ desc: null
710
+ value: 0.0
711
+ xpu_backend:
712
+ desc: null
713
+ value: None
wandb/run-20220301_171200-3h52ha0p/files/output.log ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ 0%| | 0/446 [00:00<?, ?it/s][WARNING|modeling_utils.py:388] 2022-03-01 17:12:05,828 >> Could not estimate the number of tokens of the input, floating-point operations will not be computed
4
+ Traceback (most recent call last):
5
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 539, in <module>
6
+ main()
7
+ File "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py", line 491, in main
8
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
9
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1384, in train
10
+ tr_loss_step = self.training_step(model, inputs)
11
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1959, in training_step
12
+ loss = self.compute_loss(model, inputs)
13
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 1991, in compute_loss
14
+ outputs = model(**inputs)
15
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
16
+ return forward_call(*input, **kwargs)
17
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/speech_encoder_decoder/modeling_speech_encoder_decoder.py", line 503, in forward
18
+ encoder_outputs = self.encoder(
19
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
20
+ return forward_call(*input, **kwargs)
21
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1346, in forward
22
+ extract_features = self.feature_extractor(input_values)
23
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
24
+ return forward_call(*input, **kwargs)
25
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 514, in forward
26
+ hidden_states = conv_layer(hidden_states)
27
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
28
+ return forward_call(*input, **kwargs)
29
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/models/wav2vec2/modeling_wav2vec2.py", line 389, in forward
30
+ hidden_states = self.layer_norm(hidden_states)
31
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
32
+ return forward_call(*input, **kwargs)
33
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/modules/normalization.py", line 189, in forward
34
+ return F.layer_norm(
35
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/torch/nn/functional.py", line 2347, in layer_norm
36
+ return torch.layer_norm(input, normalized_shape, weight, bias, eps, torch.backends.cudnn.enabled)
37
+ RuntimeError: CUDA out of memory. Tried to allocate 1.64 GiB (GPU 0; 15.78 GiB total capacity; 10.11 GiB already allocated; 707.31 MiB free; 13.39 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
wandb/run-20220301_171200-3h52ha0p/files/requirements.txt ADDED
@@ -0,0 +1,184 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ async-timeout==4.0.2
10
+ attrs==21.4.0
11
+ audioread==2.1.9
12
+ babel==2.9.1
13
+ backcall==0.2.0
14
+ bitsandbytes-cuda113==0.26.0
15
+ black==22.1.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.11
21
+ chex==0.1.0
22
+ click==8.0.3
23
+ clldutils==3.10.1
24
+ colorlog==6.6.0
25
+ csvw==1.11.0
26
+ cycler==0.11.0
27
+ datasets==1.18.3
28
+ debugpy==1.5.1
29
+ decorator==5.1.1
30
+ defusedxml==0.7.1
31
+ dill==0.3.4
32
+ dlinfo==1.2.1
33
+ dm-tree==0.1.6
34
+ docker-pycreds==0.4.0
35
+ entrypoints==0.4
36
+ executing==0.8.2
37
+ filelock==3.4.2
38
+ flatbuffers==2.0
39
+ flax==0.4.0
40
+ fonttools==4.29.1
41
+ frozenlist==1.3.0
42
+ fsspec==2022.1.0
43
+ gitdb==4.0.9
44
+ gitpython==3.1.27
45
+ google-auth-oauthlib==0.4.6
46
+ google-auth==2.6.0
47
+ grpcio==1.43.0
48
+ huggingface-hub==0.4.0
49
+ hypothesis==6.36.1
50
+ idna==3.3
51
+ importlib-metadata==4.10.1
52
+ ipykernel==6.8.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.0.1
55
+ ipywidgets==7.6.5
56
+ isodate==0.6.1
57
+ jax==0.2.28
58
+ jaxlib==0.1.76+cuda11.cudnn82
59
+ jedi==0.18.1
60
+ jinja2==3.0.3
61
+ jiwer==2.3.0
62
+ joblib==1.1.0
63
+ json5==0.9.6
64
+ jsonschema==4.4.0
65
+ jupyter-client==7.1.2
66
+ jupyter-console==6.4.0
67
+ jupyter-core==4.9.1
68
+ jupyter-server==1.13.5
69
+ jupyter==1.0.0
70
+ jupyterlab-pygments==0.1.2
71
+ jupyterlab-server==2.10.3
72
+ jupyterlab-widgets==1.0.2
73
+ jupyterlab==3.2.9
74
+ kiwisolver==1.3.2
75
+ librosa==0.8.1
76
+ llvmlite==0.38.0
77
+ markdown==3.3.6
78
+ markupsafe==2.0.1
79
+ matplotlib-inline==0.1.3
80
+ matplotlib==3.5.1
81
+ mistune==0.8.4
82
+ msgpack==1.0.3
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nbclassic==0.3.5
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.21.5
94
+ oauthlib==3.2.0
95
+ opt-einsum==3.3.0
96
+ optax==0.1.0
97
+ packaging==21.3
98
+ pandas==1.4.0
99
+ pandocfilters==1.5.0
100
+ parso==0.8.3
101
+ pathspec==0.9.0
102
+ pathtools==0.1.2
103
+ pexpect==4.8.0
104
+ phonemizer==3.0.1
105
+ pickleshare==0.7.5
106
+ pillow==9.0.0
107
+ pip==22.0.2
108
+ pkg-resources==0.0.0
109
+ platformdirs==2.4.1
110
+ pooch==1.6.0
111
+ prometheus-client==0.13.1
112
+ promise==2.3
113
+ prompt-toolkit==3.0.26
114
+ protobuf==3.19.4
115
+ psutil==5.9.0
116
+ ptyprocess==0.7.0
117
+ pure-eval==0.2.2
118
+ pyarrow==6.0.1
119
+ pyasn1-modules==0.2.8
120
+ pyasn1==0.4.8
121
+ pycparser==2.21
122
+ pyctcdecode==0.3.0
123
+ pygments==2.11.2
124
+ pygtrie==2.4.2
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ python-dateutil==2.8.2
128
+ python-levenshtein==0.12.2
129
+ pytz==2021.3
130
+ pyyaml==6.0
131
+ pyzmq==22.3.0
132
+ qtconsole==5.2.2
133
+ qtpy==2.0.1
134
+ regex==2022.1.18
135
+ requests-oauthlib==1.3.1
136
+ requests==2.27.1
137
+ resampy==0.2.2
138
+ rfc3986==2.0.0
139
+ rsa==4.8
140
+ sacremoses==0.0.47
141
+ scikit-learn==1.0.2
142
+ scipy==1.7.3
143
+ segments==2.2.0
144
+ send2trash==1.8.0
145
+ sentry-sdk==1.5.6
146
+ setuptools==44.1.1
147
+ shortuuid==1.0.8
148
+ six==1.16.0
149
+ smmap==5.0.0
150
+ sniffio==1.2.0
151
+ sortedcontainers==2.4.0
152
+ soundfile==0.10.3.post1
153
+ stack-data==0.1.4
154
+ tabulate==0.8.9
155
+ tensorboard-data-server==0.6.1
156
+ tensorboard-plugin-wit==1.8.1
157
+ tensorboard==2.8.0
158
+ termcolor==1.1.0
159
+ terminado==0.13.1
160
+ testpath==0.5.0
161
+ threadpoolctl==3.1.0
162
+ tokenizers==0.11.4
163
+ tomli==2.0.0
164
+ toolz==0.11.2
165
+ torch==1.10.2+cu113
166
+ torchaudio==0.10.2+cu113
167
+ tornado==6.1
168
+ tqdm==4.62.3
169
+ traitlets==5.1.1
170
+ transformers==4.17.0.dev0
171
+ typing-extensions==3.10.0.2
172
+ uritemplate==4.1.1
173
+ urllib3==1.26.8
174
+ wandb==0.12.10
175
+ wcwidth==0.2.5
176
+ webencodings==0.5.1
177
+ websocket-client==1.2.3
178
+ werkzeug==2.0.2
179
+ wheel==0.37.1
180
+ widgetsnbextension==3.5.2
181
+ xxhash==2.0.2
182
+ yarl==1.7.2
183
+ yaspin==2.1.0
184
+ zipp==3.7.0
wandb/run-20220301_171200-3h52ha0p/files/wandb-metadata.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-03-01T17:12:01.663655",
5
+ "startedAt": "2022-03-01T17:12:00.569602",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=librispeech_asr",
13
+ "--model_name_or_path=./",
14
+ "--tokenizer_name=./",
15
+ "--dataset_config_name=clean",
16
+ "--train_split_name=train.100",
17
+ "--eval_split_name=validation",
18
+ "--output_dir=./",
19
+ "--preprocessing_num_workers=1",
20
+ "--length_column_name=input_length",
21
+ "--overwrite_output_dir",
22
+ "--num_train_epochs=1",
23
+ "--per_device_train_batch_size=16",
24
+ "--per_device_eval_batch_size=16",
25
+ "--gradient_accumulation_steps=4",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--learning_rate=3e-5",
29
+ "--warmup_steps=500",
30
+ "--evaluation_strategy=steps",
31
+ "--text_column_name=text",
32
+ "--save_steps=500",
33
+ "--eval_steps=500",
34
+ "--logging_steps=1",
35
+ "--save_total_limit=1",
36
+ "--freeze_feature_encoder",
37
+ "--gradient_checkpointing",
38
+ "--fp16",
39
+ "--group_by_length",
40
+ "--predict_with_generate",
41
+ "--do_lower_case",
42
+ "--do_train",
43
+ "--do_eval",
44
+ "--report_to=wandb",
45
+ "--push_to_hub",
46
+ "--use_auth_token"
47
+ ],
48
+ "state": "running",
49
+ "program": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/run_speech_recognition_seq2seq.py",
50
+ "codePath": "run_speech_recognition_seq2seq.py",
51
+ "git": {
52
+ "remote": "https://huggingface.co/sanchit-gandhi/wav2vec2-gpt2-wandb-grid-search",
53
+ "commit": "3350b717ff05e187076e5bb0722b6b02f9837858"
54
+ },
55
+ "email": "sanchit@huggingface.co",
56
+ "root": "/home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search",
57
+ "host": "sanchit--v100",
58
+ "username": "sanchit_huggingface_co",
59
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python"
60
+ }
wandb/run-20220301_171200-3h52ha0p/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
1
+ {"_wandb": {"runtime": 8}}
wandb/run-20220301_171200-3h52ha0p/logs/debug-internal.log ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-03-01 17:12:01,466 INFO MainThread:245412 [internal.py:wandb_internal():89] W&B internal server running at pid: 245412, started at: 2022-03-01 17:12:01.466387
2
+ 2022-03-01 17:12:01,468 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: check_version
3
+ 2022-03-01 17:12:01,469 INFO WriterThread:245412 [datastore.py:open_for_write():77] open: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/run-3h52ha0p.wandb
4
+ 2022-03-01 17:12:01,470 DEBUG SenderThread:245412 [sender.py:send():235] send: header
5
+ 2022-03-01 17:12:01,470 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: check_version
6
+ 2022-03-01 17:12:01,535 DEBUG SenderThread:245412 [sender.py:send():235] send: run
7
+ 2022-03-01 17:12:01,658 INFO SenderThread:245412 [dir_watcher.py:__init__():169] watching files in: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files
8
+ 2022-03-01 17:12:01,658 INFO SenderThread:245412 [sender.py:_start_run_threads():809] run started: 3h52ha0p with start time 1646154720
9
+ 2022-03-01 17:12:01,658 DEBUG SenderThread:245412 [sender.py:send():235] send: summary
10
+ 2022-03-01 17:12:01,658 INFO SenderThread:245412 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
11
+ 2022-03-01 17:12:01,659 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: run_start
12
+ 2022-03-01 17:12:01,663 DEBUG HandlerThread:245412 [meta.py:__init__():36] meta init
13
+ 2022-03-01 17:12:01,663 DEBUG HandlerThread:245412 [meta.py:__init__():50] meta init done
14
+ 2022-03-01 17:12:01,663 DEBUG HandlerThread:245412 [meta.py:probe():210] probe
15
+ 2022-03-01 17:12:01,669 DEBUG HandlerThread:245412 [meta.py:_setup_git():200] setup git
16
+ 2022-03-01 17:12:01,684 DEBUG HandlerThread:245412 [meta.py:_setup_git():207] setup git done
17
+ 2022-03-01 17:12:01,684 DEBUG HandlerThread:245412 [meta.py:_save_pip():54] save pip
18
+ 2022-03-01 17:12:01,684 DEBUG HandlerThread:245412 [meta.py:_save_pip():68] save pip done
19
+ 2022-03-01 17:12:01,685 DEBUG HandlerThread:245412 [meta.py:probe():248] probe done
20
+ 2022-03-01 17:12:01,758 DEBUG SenderThread:245412 [sender.py:send():235] send: files
21
+ 2022-03-01 17:12:01,758 INFO SenderThread:245412 [sender.py:_save_file():944] saving file wandb-metadata.json with policy now
22
+ 2022-03-01 17:12:01,762 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: stop_status
23
+ 2022-03-01 17:12:01,763 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: stop_status
24
+ 2022-03-01 17:12:01,796 DEBUG SenderThread:245412 [sender.py:send():235] send: config
25
+ 2022-03-01 17:12:01,797 DEBUG SenderThread:245412 [sender.py:send():235] send: metric
26
+ 2022-03-01 17:12:01,797 DEBUG SenderThread:245412 [sender.py:send():235] send: metric
27
+ 2022-03-01 17:12:01,797 WARNING SenderThread:245412 [sender.py:send_metric():902] Seen metric with glob (shouldnt happen)
28
+ 2022-03-01 17:12:02,036 INFO Thread-11 :245412 [upload_job.py:push():137] Uploaded file /tmp/tmpt38fg238wandb/2am5wtm4-wandb-metadata.json
29
+ 2022-03-01 17:12:02,659 INFO Thread-8 :245412 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/requirements.txt
30
+ 2022-03-01 17:12:02,660 INFO Thread-8 :245412 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/output.log
31
+ 2022-03-01 17:12:02,660 INFO Thread-8 :245412 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/wandb-metadata.json
32
+ 2022-03-01 17:12:02,660 INFO Thread-8 :245412 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/wandb-summary.json
33
+ 2022-03-01 17:12:04,659 INFO Thread-8 :245412 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/output.log
34
+ 2022-03-01 17:12:08,660 INFO Thread-8 :245412 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/output.log
35
+ 2022-03-01 17:12:10,572 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
36
+ 2022-03-01 17:12:10,572 DEBUG SenderThread:245412 [sender.py:send():235] send: telemetry
37
+ 2022-03-01 17:12:10,573 DEBUG SenderThread:245412 [sender.py:send():235] send: exit
38
+ 2022-03-01 17:12:10,573 INFO SenderThread:245412 [sender.py:send_exit():371] handling exit code: 1
39
+ 2022-03-01 17:12:10,573 INFO SenderThread:245412 [sender.py:send_exit():373] handling runtime: 8
40
+ 2022-03-01 17:12:10,573 INFO SenderThread:245412 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
41
+ 2022-03-01 17:12:10,573 INFO SenderThread:245412 [sender.py:send_exit():379] send defer
42
+ 2022-03-01 17:12:10,574 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
43
+ 2022-03-01 17:12:10,574 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
44
+ 2022-03-01 17:12:10,574 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 0
45
+ 2022-03-01 17:12:10,574 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
46
+ 2022-03-01 17:12:10,574 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 0
47
+ 2022-03-01 17:12:10,575 INFO SenderThread:245412 [sender.py:transition_state():392] send defer: 1
48
+ 2022-03-01 17:12:10,575 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
49
+ 2022-03-01 17:12:10,575 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 1
50
+ 2022-03-01 17:12:10,661 INFO Thread-8 :245412 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/output.log
51
+ 2022-03-01 17:12:10,661 INFO Thread-8 :245412 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/wandb-summary.json
52
+ 2022-03-01 17:12:10,684 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
53
+ 2022-03-01 17:12:10,684 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 1
54
+ 2022-03-01 17:12:10,684 INFO SenderThread:245412 [sender.py:transition_state():392] send defer: 2
55
+ 2022-03-01 17:12:10,684 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
56
+ 2022-03-01 17:12:10,685 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
57
+ 2022-03-01 17:12:10,685 DEBUG SenderThread:245412 [sender.py:send():235] send: stats
58
+ 2022-03-01 17:12:10,685 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
59
+ 2022-03-01 17:12:10,686 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 2
60
+ 2022-03-01 17:12:10,686 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
61
+ 2022-03-01 17:12:10,686 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 2
62
+ 2022-03-01 17:12:10,686 INFO SenderThread:245412 [sender.py:transition_state():392] send defer: 3
63
+ 2022-03-01 17:12:10,686 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
64
+ 2022-03-01 17:12:10,686 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 3
65
+ 2022-03-01 17:12:10,686 DEBUG SenderThread:245412 [sender.py:send():235] send: summary
66
+ 2022-03-01 17:12:10,687 INFO SenderThread:245412 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
67
+ 2022-03-01 17:12:10,687 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
68
+ 2022-03-01 17:12:10,687 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 3
69
+ 2022-03-01 17:12:10,687 INFO SenderThread:245412 [sender.py:transition_state():392] send defer: 4
70
+ 2022-03-01 17:12:10,687 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
71
+ 2022-03-01 17:12:10,687 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 4
72
+ 2022-03-01 17:12:10,687 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
73
+ 2022-03-01 17:12:10,688 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 4
74
+ 2022-03-01 17:12:10,791 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
75
+ 2022-03-01 17:12:10,821 INFO SenderThread:245412 [sender.py:transition_state():392] send defer: 5
76
+ 2022-03-01 17:12:10,821 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
77
+ 2022-03-01 17:12:10,822 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
78
+ 2022-03-01 17:12:10,822 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 5
79
+ 2022-03-01 17:12:10,822 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
80
+ 2022-03-01 17:12:10,822 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 5
81
+ 2022-03-01 17:12:10,822 INFO SenderThread:245412 [dir_watcher.py:finish():283] shutting down directory watcher
82
+ 2022-03-01 17:12:10,923 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
83
+ 2022-03-01 17:12:11,662 INFO Thread-8 :245412 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/config.yaml
84
+ 2022-03-01 17:12:11,662 INFO SenderThread:245412 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/wandb-summary.json
85
+ 2022-03-01 17:12:11,662 INFO SenderThread:245412 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files
86
+ 2022-03-01 17:12:11,663 INFO SenderThread:245412 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/wandb-metadata.json wandb-metadata.json
87
+ 2022-03-01 17:12:11,663 INFO SenderThread:245412 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/output.log output.log
88
+ 2022-03-01 17:12:11,663 INFO SenderThread:245412 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/wandb-summary.json wandb-summary.json
89
+ 2022-03-01 17:12:11,663 INFO SenderThread:245412 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/requirements.txt requirements.txt
90
+ 2022-03-01 17:12:11,663 INFO SenderThread:245412 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/config.yaml config.yaml
91
+ 2022-03-01 17:12:11,666 INFO SenderThread:245412 [sender.py:transition_state():392] send defer: 6
92
+ 2022-03-01 17:12:11,666 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
93
+ 2022-03-01 17:12:11,672 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
94
+ 2022-03-01 17:12:11,673 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 6
95
+ 2022-03-01 17:12:11,673 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
96
+ 2022-03-01 17:12:11,673 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 6
97
+ 2022-03-01 17:12:11,674 INFO SenderThread:245412 [file_pusher.py:finish():177] shutting down file pusher
98
+ 2022-03-01 17:12:11,768 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
99
+ 2022-03-01 17:12:11,768 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
100
+ 2022-03-01 17:12:11,869 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
101
+ 2022-03-01 17:12:11,869 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
102
+ 2022-03-01 17:12:11,967 INFO Thread-12 :245412 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/output.log
103
+ 2022-03-01 17:12:11,970 INFO Thread-14 :245412 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/requirements.txt
104
+ 2022-03-01 17:12:11,971 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
105
+ 2022-03-01 17:12:11,971 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
106
+ 2022-03-01 17:12:11,976 INFO Thread-13 :245412 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/wandb-summary.json
107
+ 2022-03-01 17:12:11,995 INFO Thread-15 :245412 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/files/config.yaml
108
+ 2022-03-01 17:12:12,072 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
109
+ 2022-03-01 17:12:12,073 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
110
+ 2022-03-01 17:12:12,174 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
111
+ 2022-03-01 17:12:12,174 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
112
+ 2022-03-01 17:12:12,196 INFO Thread-7 :245412 [sender.py:transition_state():392] send defer: 7
113
+ 2022-03-01 17:12:12,196 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
114
+ 2022-03-01 17:12:12,196 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 7
115
+ 2022-03-01 17:12:12,196 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
116
+ 2022-03-01 17:12:12,196 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 7
117
+ 2022-03-01 17:12:12,243 INFO SenderThread:245412 [sender.py:transition_state():392] send defer: 8
118
+ 2022-03-01 17:12:12,244 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
119
+ 2022-03-01 17:12:12,244 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 8
120
+ 2022-03-01 17:12:12,244 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
121
+ 2022-03-01 17:12:12,244 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 8
122
+ 2022-03-01 17:12:12,244 INFO SenderThread:245412 [sender.py:transition_state():392] send defer: 9
123
+ 2022-03-01 17:12:12,245 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: defer
124
+ 2022-03-01 17:12:12,245 INFO HandlerThread:245412 [handler.py:handle_request_defer():154] handle defer: 9
125
+ 2022-03-01 17:12:12,245 DEBUG SenderThread:245412 [sender.py:send():235] send: final
126
+ 2022-03-01 17:12:12,245 DEBUG SenderThread:245412 [sender.py:send():235] send: footer
127
+ 2022-03-01 17:12:12,245 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: defer
128
+ 2022-03-01 17:12:12,245 INFO SenderThread:245412 [sender.py:send_request_defer():388] handle sender defer: 9
129
+ 2022-03-01 17:12:12,275 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: poll_exit
130
+ 2022-03-01 17:12:12,275 DEBUG SenderThread:245412 [sender.py:send_request():249] send_request: poll_exit
131
+ 2022-03-01 17:12:12,276 INFO SenderThread:245412 [file_pusher.py:join():182] waiting for file pusher
132
+ 2022-03-01 17:12:12,332 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: get_summary
133
+ 2022-03-01 17:12:12,333 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: sampled_history
134
+ 2022-03-01 17:12:12,334 DEBUG HandlerThread:245412 [handler.py:handle_request():131] handle_request: shutdown
135
+ 2022-03-01 17:12:12,334 INFO HandlerThread:245412 [handler.py:finish():739] shutting down handler
136
+ 2022-03-01 17:12:13,246 INFO WriterThread:245412 [datastore.py:close():281] close: /home/sanchit_huggingface_co/wav2vec2-gpt2-wandb-grid-search/wandb/run-20220301_171200-3h52ha0p/run-3h52ha0p.wandb
137
+ 2022-03-01 17:12:13,331 INFO SenderThread:245412 [sender.py:finish():1075] shutting down sender
138
+ 2022-03-01 17:12:13,331 INFO SenderThread:245412 [file_pusher.py:finish():177] shutting down file pusher
139
+ 2022-03-01 17:12:13,331 INFO SenderThread:245412 [file_pusher.py:join():182] waiting for file pusher
140
+ 2022-03-01 17:12:13,333 INFO MainThread:245412 [internal.py:handle_exit():79] Internal process exited