sanchit-gandhi HF staff commited on
Commit
21aa7a2
1 Parent(s): 9638619

Training in progress, step 500

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. config.json +1 -1
  3. emissions.csv +1 -0
  4. pytorch_model.bin +1 -1
  5. runs/May03_15-33-55_sanchit--v100/events.out.tfevents.1651592261.sanchit--v100.40140.0 +0 -3
  6. runs/May03_15-58-13_sanchit--v100/1651596054.9030492/events.out.tfevents.1651596054.sanchit--v100.40903.1 +0 -3
  7. runs/May03_15-58-13_sanchit--v100/events.out.tfevents.1651596054.sanchit--v100.40903.0 +0 -3
  8. runs/May03_17-15-22_sanchit--v100/1651598399.7088904/events.out.tfevents.1651598399.sanchit--v100.42111.1 +0 -3
  9. runs/May03_17-15-22_sanchit--v100/events.out.tfevents.1651598399.sanchit--v100.42111.0 +0 -3
  10. runs/May04_08-29-27_sanchit--v100/1651653030.564084/events.out.tfevents.1651653030.sanchit--v100.48541.1 +0 -3
  11. runs/May04_08-29-27_sanchit--v100/events.out.tfevents.1651653030.sanchit--v100.48541.0 +0 -3
  12. runs/May04_13-30-37_sanchit--v100/1651674088.8879716/events.out.tfevents.1651674088.sanchit--v100.50375.1 +0 -3
  13. runs/May04_13-30-37_sanchit--v100/events.out.tfevents.1651674088.sanchit--v100.50375.0 +0 -3
  14. runs/May05_16-32-27_sanchit--v100/1651768642.7722313/events.out.tfevents.1651768642.sanchit--v100.65172.1 +0 -3
  15. runs/May05_16-32-27_sanchit--v100/events.out.tfevents.1651768642.sanchit--v100.65172.0 +0 -3
  16. runs/May05_16-41-38_sanchit--v100/1651768958.3876698/events.out.tfevents.1651768958.sanchit--v100.65949.1 +0 -3
  17. runs/May05_16-41-38_sanchit--v100/events.out.tfevents.1651768958.sanchit--v100.65949.0 +0 -3
  18. runs/May05_16-44-56_sanchit--v100/1651769149.9394004/events.out.tfevents.1651769149.sanchit--v100.66677.1 +0 -3
  19. runs/May05_16-44-56_sanchit--v100/events.out.tfevents.1651769149.sanchit--v100.66677.0 +0 -3
  20. runs/May05_16-48-07_sanchit--v100/1651769342.0456421/events.out.tfevents.1651769342.sanchit--v100.67410.1 +0 -3
  21. runs/May05_16-48-07_sanchit--v100/events.out.tfevents.1651769342.sanchit--v100.67410.0 +0 -3
  22. runs/{May03_12-39-14_sanchit--v100/1651584502.8814862/events.out.tfevents.1651584502.sanchit--v100.38894.1 → May05_17-36-54_sanchit--v100/1651772268.485843/events.out.tfevents.1651772268.sanchit--v100.74387.1} +1 -1
  23. runs/{May03_12-39-14_sanchit--v100/events.out.tfevents.1651584502.sanchit--v100.38894.0 → May05_17-36-54_sanchit--v100/events.out.tfevents.1651772268.sanchit--v100.74387.0} +2 -2
  24. sweep.yaml +6 -8
  25. train_results.json +0 -8
  26. trainer_state.json +0 -0
  27. training_args.bin +1 -1
  28. wandb/debug-cli.log +132 -0
  29. wandb/debug-internal.log +1 -1
  30. wandb/debug.log +1 -1
  31. wandb/latest-run +1 -1
  32. wandb/run-20220505_164902-lr7oonrp/files/config.yaml +9 -0
  33. wandb/run-20220505_164902-lr7oonrp/files/output.log +22 -0
  34. wandb/run-20220505_164902-lr7oonrp/files/wandb-summary.json +1 -1
  35. wandb/run-20220505_164902-lr7oonrp/logs/debug-internal.log +105 -0
  36. wandb/run-20220505_164902-lr7oonrp/logs/debug.log +82 -0
  37. wandb/run-20220505_164902-lr7oonrp/run-lr7oonrp.wandb +0 -0
  38. wandb/run-20220505_173213-ykkm1chu/files/config.yaml +788 -0
  39. wandb/run-20220505_173213-ykkm1chu/files/output.log +29 -0
  40. wandb/run-20220505_173213-ykkm1chu/files/requirements.txt +287 -0
  41. wandb/run-20220505_173213-ykkm1chu/files/wandb-metadata.json +57 -0
  42. wandb/run-20220505_173213-ykkm1chu/files/wandb-summary.json +1 -0
  43. wandb/run-20220505_173213-ykkm1chu/logs/debug-internal.log +76 -0
  44. wandb/run-20220505_173213-ykkm1chu/logs/debug.log +28 -0
  45. wandb/run-20220505_173213-ykkm1chu/run-ykkm1chu.wandb +0 -0
  46. wandb/run-20220505_173748-b097rk18/files/config.yaml +0 -0
  47. wandb/run-20220505_173748-b097rk18/files/output.log +0 -0
  48. wandb/run-20220505_173748-b097rk18/files/requirements.txt +287 -0
  49. wandb/run-20220505_173748-b097rk18/files/wandb-metadata.json +57 -0
  50. wandb/run-20220505_173748-b097rk18/files/wandb-summary.json +0 -0
.gitattributes CHANGED
@@ -29,3 +29,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
29
  wandb/run-20220504_142129-1tmxz74i/run-1tmxz74i.wandb filter=lfs diff=lfs merge=lfs -text
30
  wandb/run-20220504_142129-1tmxz74i/logs/debug-internal.log filter=lfs diff=lfs merge=lfs -text
31
  wandb/run-20220504_142129-1tmxz74i/files/output.log filter=lfs diff=lfs merge=lfs -text
 
29
  wandb/run-20220504_142129-1tmxz74i/run-1tmxz74i.wandb filter=lfs diff=lfs merge=lfs -text
30
  wandb/run-20220504_142129-1tmxz74i/logs/debug-internal.log filter=lfs diff=lfs merge=lfs -text
31
  wandb/run-20220504_142129-1tmxz74i/files/output.log filter=lfs diff=lfs merge=lfs -text
32
+ wandb/run-20220505_173748-b097rk18/run-b097rk18.wandb filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -182,7 +182,7 @@
182
  "forced_eos_token_id": null,
183
  "gradient_checkpointing": false,
184
  "hidden_act": "gelu",
185
- "hidden_dropout": 0.16627274455555233,
186
  "hidden_size": 1024,
187
  "id2label": {
188
  "0": "LABEL_0",
182
  "forced_eos_token_id": null,
183
  "gradient_checkpointing": false,
184
  "hidden_act": "gelu",
185
+ "hidden_dropout": 0.2,
186
  "hidden_size": 1024,
187
  "id2label": {
188
  "0": "LABEL_0",
emissions.csv CHANGED
@@ -4,3 +4,4 @@ timestamp,experiment_id,project_name,duration,emissions,energy_consumed,country_
4
  2022-05-05T16:42:47,89692b38-e983-4664-a932-8df61002625d,codecarbon,4.291411638259888,0.00019785712188713343,0.00034938569995962115,USA,USA,Iowa,Y,gcp,us-central1
5
  2022-05-05T16:45:58,ef3cf220-dd7d-4101-8f2e-5b58125d0f6a,codecarbon,4.533631801605225,0.00011988175208065735,0.0002116930109141045,USA,USA,Iowa,Y,gcp,us-central1
6
  2022-05-05T16:49:10,d2831db0-f95d-4aec-bbd0-5241d2505f54,codecarbon,4.5968592166900635,0.00010660969900888212,0.00018825657603546201,USA,USA,Iowa,Y,gcp,us-central1
 
4
  2022-05-05T16:42:47,89692b38-e983-4664-a932-8df61002625d,codecarbon,4.291411638259888,0.00019785712188713343,0.00034938569995962115,USA,USA,Iowa,Y,gcp,us-central1
5
  2022-05-05T16:45:58,ef3cf220-dd7d-4101-8f2e-5b58125d0f6a,codecarbon,4.533631801605225,0.00011988175208065735,0.0002116930109141045,USA,USA,Iowa,Y,gcp,us-central1
6
  2022-05-05T16:49:10,d2831db0-f95d-4aec-bbd0-5241d2505f54,codecarbon,4.5968592166900635,0.00010660969900888212,0.00018825657603546201,USA,USA,Iowa,Y,gcp,us-central1
7
+ 2022-05-05T17:32:21,595a622e-ed46-407f-95ef-aee33f1bdf6d,codecarbon,4.391920328140259,9.467203556285354e-05,0.0001671764710627822,USA,USA,Iowa,Y,gcp,us-central1
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee59259a1d3d50f94ca31b73c03ef96aea7289bbdb17676e503a799ca664af3f
3
  size 2353867057
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e68181435e477b5116db4b7759d8834b718e949730cecc7dd1c173661c0b5b6c
3
  size 2353867057
runs/May03_15-33-55_sanchit--v100/events.out.tfevents.1651592261.sanchit--v100.40140.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c965accf9ea2be03062bfecc4c45cdb54eb6b4b444ea37a2c9533c8db4fed044
3
- size 87941
 
 
 
runs/May03_15-58-13_sanchit--v100/1651596054.9030492/events.out.tfevents.1651596054.sanchit--v100.40903.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:76e39c8c39b9be3c4531fd72d066671f7365b550e8079fee72b6060a36be59f6
3
- size 5184
 
 
 
runs/May03_15-58-13_sanchit--v100/events.out.tfevents.1651596054.sanchit--v100.40903.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:410298c8949435d86a040fb00c0e0a7bde2f0445544abeadadae587ba569bfba
3
- size 73654
 
 
 
runs/May03_17-15-22_sanchit--v100/1651598399.7088904/events.out.tfevents.1651598399.sanchit--v100.42111.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:61b94cd769015be4bbc81b78b8074c963d947a90c91e7cf517abaad8c3ff3078
3
- size 5184
 
 
 
runs/May03_17-15-22_sanchit--v100/events.out.tfevents.1651598399.sanchit--v100.42111.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:06ebac1e97e241e13d6c8a536ad57e3ef5b58c510633a0824a10536aae4662a4
3
- size 797661
 
 
 
runs/May04_08-29-27_sanchit--v100/1651653030.564084/events.out.tfevents.1651653030.sanchit--v100.48541.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9f818ffc5324af0a19b67e9e869654bb112292ade31d00f48263ec9cf177206c
3
- size 5184
 
 
 
runs/May04_08-29-27_sanchit--v100/events.out.tfevents.1651653030.sanchit--v100.48541.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6120cfa0b8dd7cc6afe852da88002229769bc8f3efa93f730fa0661088c42d36
3
- size 88290
 
 
 
runs/May04_13-30-37_sanchit--v100/1651674088.8879716/events.out.tfevents.1651674088.sanchit--v100.50375.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:aef34e54eaac03ee713590a0eb81b34a1eb839a803375bb6e9e3a6e439991c2b
3
- size 5184
 
 
 
runs/May04_13-30-37_sanchit--v100/events.out.tfevents.1651674088.sanchit--v100.50375.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:baefc16420fc1dd3035ed23659f51947934cd5d60be8002b0e1d45b1b6244ef1
3
- size 1541927
 
 
 
runs/May05_16-32-27_sanchit--v100/1651768642.7722313/events.out.tfevents.1651768642.sanchit--v100.65172.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1fe9010d25b5426ae74e1154e5be2727f4b9d3dc3d3ac8423b4e36ab102ca5b1
3
- size 5184
 
 
 
runs/May05_16-32-27_sanchit--v100/events.out.tfevents.1651768642.sanchit--v100.65172.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:dffb33c2552c28d9cdf40229f277c5ce3d0f56c41496a3ad6c1fcdd0e08ec2d4
3
- size 10206
 
 
 
runs/May05_16-41-38_sanchit--v100/1651768958.3876698/events.out.tfevents.1651768958.sanchit--v100.65949.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:db04a69bbb2713a70173d62a1e8ee86184b8325a2e3bde42e89647c96cbd1027
3
- size 5184
 
 
 
runs/May05_16-41-38_sanchit--v100/events.out.tfevents.1651768958.sanchit--v100.65949.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d852cd4da235fc0dd1135a7957f543279a67523fdbc8cdd26c32a2b17c3da93
3
- size 10206
 
 
 
runs/May05_16-44-56_sanchit--v100/1651769149.9394004/events.out.tfevents.1651769149.sanchit--v100.66677.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf44e1ee4066c7afb2c0126abffcdc62708cfaf679d440482c5b10bdfa09dfed
3
- size 5184
 
 
 
runs/May05_16-44-56_sanchit--v100/events.out.tfevents.1651769149.sanchit--v100.66677.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6cf35e831e3ed592f3e7d3c82817cc2768013f4c0662d3bfcd1223cc883fbf3f
3
- size 10206
 
 
 
runs/May05_16-48-07_sanchit--v100/1651769342.0456421/events.out.tfevents.1651769342.sanchit--v100.67410.1 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:736259815069e83ea3029a853d47a1d4e7feceac2e28a92a58b2bf2531c6944e
3
- size 5184
 
 
 
runs/May05_16-48-07_sanchit--v100/events.out.tfevents.1651769342.sanchit--v100.67410.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c214fc3b70e657543c5ebfda93915ca0349eaee975294c9f561c46df5eedc39b
3
- size 10207
 
 
 
runs/{May03_12-39-14_sanchit--v100/1651584502.8814862/events.out.tfevents.1651584502.sanchit--v100.38894.1 → May05_17-36-54_sanchit--v100/1651772268.485843/events.out.tfevents.1651772268.sanchit--v100.74387.1} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a5e50bb326f116678bb0d2af94ee25e92d82ce3268121a68b288dd930e13470
3
  size 5184
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:151c1d4351b970fd9dd0b6408ef78ef7693e783a530f7635197797b147fdbdd8
3
  size 5184
runs/{May03_12-39-14_sanchit--v100/events.out.tfevents.1651584502.sanchit--v100.38894.0 → May05_17-36-54_sanchit--v100/events.out.tfevents.1651772268.sanchit--v100.74387.0} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65a28e37ffdeeb562c734e7a4815296918b93b818763a73a70ab20d29627105c
3
- size 10592
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac2d6fb8bada68aa8ed66f0a7a406bf11f62cf5dd66fb0accc7d789bec6c30f0
3
+ size 88259
sweep.yaml CHANGED
@@ -13,7 +13,7 @@ command:
13
  - --push_to_hub
14
  - --use_auth_token
15
  - ${args}
16
- method: random
17
  metric:
18
  goal: maximize
19
  name: eval/bleu
@@ -35,19 +35,17 @@ parameters:
35
  per_device_eval_batch_size:
36
  value: 8
37
  gradient_accumulation_steps:
38
- value: 8
39
  generation_max_length:
40
  value: 40
41
  generation_num_beams:
42
  value: 1
43
  learning_rate:
44
- distribution: log_uniform
45
- max: -6.9
46
- min: -9.2
47
  hidden_dropout:
48
- distribution: log_uniform
49
- max: -1.6
50
- min: -3.4
51
  warmup_steps:
52
  value: 500
53
  evaluation_strategy:
13
  - --push_to_hub
14
  - --use_auth_token
15
  - ${args}
16
+ method: grid
17
  metric:
18
  goal: maximize
19
  name: eval/bleu
35
  per_device_eval_batch_size:
36
  value: 8
37
  gradient_accumulation_steps:
38
+ value: 16
39
  generation_max_length:
40
  value: 40
41
  generation_num_beams:
42
  value: 1
43
  learning_rate:
44
+ values:
45
+ - 3e-5
46
+ - 3e-4
47
  hidden_dropout:
48
+ value: 0.2
 
 
49
  warmup_steps:
50
  value: 500
51
  evaluation_strategy:
train_results.json DELETED
@@ -1,8 +0,0 @@
1
- {
2
- "epoch": 3.0,
3
- "train_loss": 0.0,
4
- "train_runtime": 8.6541,
5
- "train_samples": 207374,
6
- "train_samples_per_second": 71887.743,
7
- "train_steps_per_second": 1123.17
8
- }
 
 
 
 
 
 
 
 
trainer_state.json DELETED
The diff for this file is too large to render. See raw diff
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0d48663471605abd96c1360ba2f44538ceb5703ea64f57354b42be8d949c099
3
  size 3247
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d4c647b971e03f3437cfa5c377bc5c5dd4ecd573f7c676852a13d5a49d1c8db
3
  size 3247
wandb/debug-cli.log CHANGED
@@ -184,3 +184,135 @@
184
  warmup_steps: 500
185
  2022-05-05 16:48:04 INFO About to run command: python3 run_xtreme_s.py --overwrite_output_dir --freeze_feature_encoder --gradient_checkpointing --predict_with_generate --fp16 --group_by_length --do_train --do_eval --load_best_model_at_end --push_to_hub --use_auth_token --eval_split_name=test --eval_steps=500 --evaluation_strategy=steps --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=8 --greater_is_better=True --hidden_dropout=0.16627274455555233 --language=fr.en --learning_rate=0.00022154311254852488 --logging_steps=1 --max_duration_in_seconds=20 --metric_for_best_model=bleu --model_name_or_path=./ --num_train_epochs=3 --output_dir=./ --per_device_eval_batch_size=8 --per_device_train_batch_size=8 --save_steps=500 --task=covost2 --warmup_steps=500
186
  2022-05-05 16:48:09 INFO Running runs: ['lr7oonrp']
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
184
  warmup_steps: 500
185
  2022-05-05 16:48:04 INFO About to run command: python3 run_xtreme_s.py --overwrite_output_dir --freeze_feature_encoder --gradient_checkpointing --predict_with_generate --fp16 --group_by_length --do_train --do_eval --load_best_model_at_end --push_to_hub --use_auth_token --eval_split_name=test --eval_steps=500 --evaluation_strategy=steps --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=8 --greater_is_better=True --hidden_dropout=0.16627274455555233 --language=fr.en --learning_rate=0.00022154311254852488 --logging_steps=1 --max_duration_in_seconds=20 --metric_for_best_model=bleu --model_name_or_path=./ --num_train_epochs=3 --output_dir=./ --per_device_eval_batch_size=8 --per_device_train_batch_size=8 --save_steps=500 --task=covost2 --warmup_steps=500
186
  2022-05-05 16:48:09 INFO Running runs: ['lr7oonrp']
187
+ 2022-05-05 16:51:25 ERROR Detected 5 failed runs in a row, shutting down.
188
+ 2022-05-05 16:51:25 INFO To change this value set WANDB_AGENT_MAX_INITIAL_FAILURES=val
189
+ 2022-05-05 17:29:38 INFO Running runs: []
190
+ 2022-05-05 17:29:38 INFO Agent received command: run
191
+ 2022-05-05 17:29:38 INFO Agent starting run with config:
192
+ eval_split_name: test
193
+ eval_steps: 500
194
+ evaluation_strategy: steps
195
+ generation_max_length: 40
196
+ generation_num_beams: 1
197
+ gradient_accumulation_steps: 16
198
+ greater_is_better: True
199
+ hidden_dropout: 0.2
200
+ language: fr.en
201
+ learning_rate: 0.0007057712331944904
202
+ logging_steps: 1
203
+ max_duration_in_seconds: 20
204
+ metric_for_best_model: bleu
205
+ model_name_or_path: ./
206
+ num_train_epochs: 3
207
+ output_dir: ./
208
+ per_device_eval_batch_size: 8
209
+ per_device_train_batch_size: 8
210
+ save_steps: 500
211
+ task: covost2
212
+ warmup_steps: 500
213
+ 2022-05-05 17:29:38 INFO About to run command: python3 run_xtreme_s.py --overwrite_output_dir --freeze_feature_encoder --gradient_checkpointing --predict_with_generate --fp16 --group_by_length --do_train --do_eval --load_best_model_at_end --push_to_hub --use_auth_token --eval_split_name=test --eval_steps=500 --evaluation_strategy=steps --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=16 --greater_is_better=True --hidden_dropout=0.2 --language=fr.en --learning_rate=0.0007057712331944904 --logging_steps=1 --max_duration_in_seconds=20 --metric_for_best_model=bleu --model_name_or_path=./ --num_train_epochs=3 --output_dir=./ --per_device_eval_batch_size=8 --per_device_train_batch_size=8 --save_steps=500 --task=covost2 --warmup_steps=500
214
+ 2022-05-05 17:29:43 INFO Running runs: ['rwbnpkt2']
215
+ 2022-05-05 17:29:50 INFO Running runs: []
216
+ 2022-05-05 17:29:51 INFO Agent received command: run
217
+ 2022-05-05 17:29:51 INFO Agent starting run with config:
218
+ eval_split_name: test
219
+ eval_steps: 500
220
+ evaluation_strategy: steps
221
+ generation_max_length: 40
222
+ generation_num_beams: 1
223
+ gradient_accumulation_steps: 16
224
+ greater_is_better: True
225
+ hidden_dropout: 0.2
226
+ language: fr.en
227
+ learning_rate: 0.0005587128574267087
228
+ logging_steps: 1
229
+ max_duration_in_seconds: 20
230
+ metric_for_best_model: bleu
231
+ model_name_or_path: ./
232
+ num_train_epochs: 3
233
+ output_dir: ./
234
+ per_device_eval_batch_size: 8
235
+ per_device_train_batch_size: 8
236
+ save_steps: 500
237
+ task: covost2
238
+ warmup_steps: 500
239
+ 2022-05-05 17:29:51 INFO About to run command: python3 run_xtreme_s.py --overwrite_output_dir --freeze_feature_encoder --gradient_checkpointing --predict_with_generate --fp16 --group_by_length --do_train --do_eval --load_best_model_at_end --push_to_hub --use_auth_token --eval_split_name=test --eval_steps=500 --evaluation_strategy=steps --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=16 --greater_is_better=True --hidden_dropout=0.2 --language=fr.en --learning_rate=0.0005587128574267087 --logging_steps=1 --max_duration_in_seconds=20 --metric_for_best_model=bleu --model_name_or_path=./ --num_train_epochs=3 --output_dir=./ --per_device_eval_batch_size=8 --per_device_train_batch_size=8 --save_steps=500 --task=covost2 --warmup_steps=500
240
+ 2022-05-05 17:29:56 INFO Running runs: ['ydgnpqx2']
241
+ 2022-05-05 17:30:19 INFO Cleaning up finished run: rwbnpkt2
242
+ 2022-05-05 17:30:19 INFO Agent received command: run
243
+ 2022-05-05 17:30:19 INFO Agent starting run with config:
244
+ eval_split_name: test
245
+ eval_steps: 500
246
+ evaluation_strategy: steps
247
+ generation_max_length: 40
248
+ generation_num_beams: 1
249
+ gradient_accumulation_steps: 16
250
+ greater_is_better: True
251
+ hidden_dropout: 0.2
252
+ language: fr.en
253
+ learning_rate: 0.0003851276453057612
254
+ logging_steps: 1
255
+ max_duration_in_seconds: 20
256
+ metric_for_best_model: bleu
257
+ model_name_or_path: ./
258
+ num_train_epochs: 3
259
+ output_dir: ./
260
+ per_device_eval_batch_size: 8
261
+ per_device_train_batch_size: 8
262
+ save_steps: 500
263
+ task: covost2
264
+ warmup_steps: 500
265
+ 2022-05-05 17:30:19 INFO About to run command: python3 run_xtreme_s.py --overwrite_output_dir --freeze_feature_encoder --gradient_checkpointing --predict_with_generate --fp16 --group_by_length --do_train --do_eval --load_best_model_at_end --push_to_hub --use_auth_token --eval_split_name=test --eval_steps=500 --evaluation_strategy=steps --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=16 --greater_is_better=True --hidden_dropout=0.2 --language=fr.en --learning_rate=0.0003851276453057612 --logging_steps=1 --max_duration_in_seconds=20 --metric_for_best_model=bleu --model_name_or_path=./ --num_train_epochs=3 --output_dir=./ --per_device_eval_batch_size=8 --per_device_train_batch_size=8 --save_steps=500 --task=covost2 --warmup_steps=500
266
+ 2022-05-05 17:30:24 INFO Running runs: ['xbi4p92m']
267
+ 2022-05-05 17:31:15 INFO Running runs: []
268
+ 2022-05-05 17:31:16 INFO Agent received command: run
269
+ 2022-05-05 17:31:16 INFO Agent starting run with config:
270
+ eval_split_name: test
271
+ eval_steps: 500
272
+ evaluation_strategy: steps
273
+ generation_max_length: 40
274
+ generation_num_beams: 1
275
+ gradient_accumulation_steps: 16
276
+ greater_is_better: True
277
+ hidden_dropout: 0.2
278
+ language: fr.en
279
+ learning_rate: 0.0003287457929573604
280
+ logging_steps: 1
281
+ max_duration_in_seconds: 20
282
+ metric_for_best_model: bleu
283
+ model_name_or_path: ./
284
+ num_train_epochs: 3
285
+ output_dir: ./
286
+ per_device_eval_batch_size: 8
287
+ per_device_train_batch_size: 8
288
+ save_steps: 500
289
+ task: covost2
290
+ warmup_steps: 500
291
+ 2022-05-05 17:31:16 INFO About to run command: python3 run_xtreme_s.py --overwrite_output_dir --freeze_feature_encoder --gradient_checkpointing --predict_with_generate --fp16 --group_by_length --do_train --do_eval --load_best_model_at_end --push_to_hub --use_auth_token --eval_split_name=test --eval_steps=500 --evaluation_strategy=steps --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=16 --greater_is_better=True --hidden_dropout=0.2 --language=fr.en --learning_rate=0.0003287457929573604 --logging_steps=1 --max_duration_in_seconds=20 --metric_for_best_model=bleu --model_name_or_path=./ --num_train_epochs=3 --output_dir=./ --per_device_eval_batch_size=8 --per_device_train_batch_size=8 --save_steps=500 --task=covost2 --warmup_steps=500
292
+ 2022-05-05 17:31:21 INFO Running runs: ['ykkm1chu']
293
+ 2022-05-05 17:36:50 INFO Running runs: []
294
+ 2022-05-05 17:36:50 INFO Agent received command: run
295
+ 2022-05-05 17:36:50 INFO Agent starting run with config:
296
+ eval_split_name: test
297
+ eval_steps: 500
298
+ evaluation_strategy: steps
299
+ generation_max_length: 40
300
+ generation_num_beams: 1
301
+ gradient_accumulation_steps: 16
302
+ greater_is_better: True
303
+ hidden_dropout: 0.2
304
+ language: fr.en
305
+ learning_rate: 3e-05
306
+ logging_steps: 1
307
+ max_duration_in_seconds: 20
308
+ metric_for_best_model: bleu
309
+ model_name_or_path: ./
310
+ num_train_epochs: 3
311
+ output_dir: ./
312
+ per_device_eval_batch_size: 8
313
+ per_device_train_batch_size: 8
314
+ save_steps: 500
315
+ task: covost2
316
+ warmup_steps: 500
317
+ 2022-05-05 17:36:50 INFO About to run command: python3 run_xtreme_s.py --overwrite_output_dir --freeze_feature_encoder --gradient_checkpointing --predict_with_generate --fp16 --group_by_length --do_train --do_eval --load_best_model_at_end --push_to_hub --use_auth_token --eval_split_name=test --eval_steps=500 --evaluation_strategy=steps --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=16 --greater_is_better=True --hidden_dropout=0.2 --language=fr.en --learning_rate=3e-05 --logging_steps=1 --max_duration_in_seconds=20 --metric_for_best_model=bleu --model_name_or_path=./ --num_train_epochs=3 --output_dir=./ --per_device_eval_batch_size=8 --per_device_train_batch_size=8 --save_steps=500 --task=covost2 --warmup_steps=500
318
+ 2022-05-05 17:36:55 INFO Running runs: ['b097rk18']
wandb/debug-internal.log CHANGED
@@ -1 +1 @@
1
- run-20220505_164902-lr7oonrp/logs/debug-internal.log
1
+ run-20220505_173748-b097rk18/logs/debug-internal.log
wandb/debug.log CHANGED
@@ -1 +1 @@
1
- run-20220505_164902-lr7oonrp/logs/debug.log
1
+ run-20220505_173748-b097rk18/logs/debug.log
wandb/latest-run CHANGED
@@ -1 +1 @@
1
- run-20220505_164902-lr7oonrp
1
+ run-20220505_173748-b097rk18
wandb/run-20220505_164902-lr7oonrp/files/config.yaml CHANGED
@@ -52,7 +52,16 @@ _wandb:
52
  - 5
53
  - 11
54
  - 12
 
 
 
 
 
 
 
55
  3:
 
 
56
  - 13
57
  4: 3.9.5
58
  5: 0.12.10
52
  - 5
53
  - 11
54
  - 12
55
+ 2:
56
+ - 1
57
+ - 2
58
+ - 3
59
+ - 5
60
+ - 11
61
+ - 12
62
  3:
63
+ - 1
64
+ - 7
65
  - 13
66
  4: 3.9.5
67
  5: 0.12.10
wandb/run-20220505_164902-lr7oonrp/files/output.log CHANGED
@@ -51,3 +51,25 @@ To https://huggingface.co/sanchit-gandhi/xtreme_s_xlsr_2_bart_covost2_fr_en
51
  Saving model checkpoint to ./
52
  Configuration saved in ./config.json
53
  Model weights saved in ./pytorch_model.bin
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
51
  Saving model checkpoint to ./
52
  Configuration saved in ./config.json
53
  Model weights saved in ./pytorch_model.bin
54
+ Feature extractor saved in ./preprocessor_config.json
55
+ 05/05/2022 16:51:11 - WARNING - huggingface_hub.repository - To https://huggingface.co/sanchit-gandhi/xtreme_s_xlsr_2_bart_covost2_fr_en
56
+ 9ad50a4..9638619 main -> main
57
+ To https://huggingface.co/sanchit-gandhi/xtreme_s_xlsr_2_bart_covost2_fr_en
58
+ 9ad50a4..9638619 main -> main
59
+ Traceback (most recent call last):
60
+ File "/home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/run_xtreme_s.py", line 972, in <module>
61
+ main()
62
+ File "/home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/run_xtreme_s.py", line 964, in main
63
+ trainer.push_to_hub(**kwargs)
64
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 2944, in push_to_hub
65
+ self.create_model_card(model_name=model_name, **kwargs)
66
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/trainer.py", line 2841, in create_model_card
67
+ training_summary = TrainingSummary.from_trainer(
68
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/modelcard.py", line 611, in from_trainer
69
+ return cls(
70
+ File "<string>", line 16, in __init__
71
+ File "/home/sanchit_huggingface_co/transformers/src/transformers/modelcard.py", line 401, in __post_init__
72
+ info = model_info(self.finetuned_from)
73
+ File "/home/sanchit_huggingface_co/gcp/lib/python3.9/site-packages/huggingface_hub/hf_api.py", line 870, in model_info
74
+ return ModelInfo(**d)
75
+ TypeError: huggingface_hub.hf_api.ModelInfo() argument after ** must be a mapping, not list
wandb/run-20220505_164902-lr7oonrp/files/wandb-summary.json CHANGED
@@ -1 +1 @@
1
- {"train/train_runtime": 8.6541, "train/train_samples_per_second": 71887.743, "train/train_steps_per_second": 1123.17, "train/total_flos": 0.0, "train/train_loss": 0.0, "train/epoch": 3.0, "train/global_step": 9720, "_runtime": 8, "_timestamp": 1651769350, "_step": 0}
1
+ {"train/train_runtime": 8.6541, "train/train_samples_per_second": 71887.743, "train/train_steps_per_second": 1123.17, "train/total_flos": 0.0, "train/train_loss": 0.0, "train/epoch": 3.0, "train/global_step": 9720, "_runtime": 8, "_timestamp": 1651769350, "_step": 0, "_wandb": {"runtime": 134}}
wandb/run-20220505_164902-lr7oonrp/logs/debug-internal.log CHANGED
@@ -71,3 +71,108 @@
71
  2022-05-05 16:51:05,586 DEBUG SenderThread:67587 [sender.py:send():235] send: stats
72
  2022-05-05 16:51:06,466 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: stop_status
73
  2022-05-05 16:51:06,467 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: stop_status
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
71
  2022-05-05 16:51:05,586 DEBUG SenderThread:67587 [sender.py:send():235] send: stats
72
  2022-05-05 16:51:06,466 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: stop_status
73
  2022-05-05 16:51:06,467 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: stop_status
74
+ 2022-05-05 16:51:12,618 INFO Thread-8 :67587 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/output.log
75
+ 2022-05-05 16:51:16,619 INFO Thread-8 :67587 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/output.log
76
+ 2022-05-05 16:51:18,214 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: poll_exit
77
+ 2022-05-05 16:51:18,214 DEBUG SenderThread:67587 [sender.py:send():235] send: telemetry
78
+ 2022-05-05 16:51:18,214 DEBUG SenderThread:67587 [sender.py:send():235] send: exit
79
+ 2022-05-05 16:51:18,215 INFO SenderThread:67587 [sender.py:send_exit():371] handling exit code: 1
80
+ 2022-05-05 16:51:18,215 INFO SenderThread:67587 [sender.py:send_exit():373] handling runtime: 134
81
+ 2022-05-05 16:51:18,215 INFO SenderThread:67587 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
82
+ 2022-05-05 16:51:18,215 INFO SenderThread:67587 [sender.py:send_exit():379] send defer
83
+ 2022-05-05 16:51:18,216 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: poll_exit
84
+ 2022-05-05 16:51:18,216 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
85
+ 2022-05-05 16:51:18,216 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 0
86
+ 2022-05-05 16:51:18,217 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
87
+ 2022-05-05 16:51:18,217 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 0
88
+ 2022-05-05 16:51:18,217 INFO SenderThread:67587 [sender.py:transition_state():392] send defer: 1
89
+ 2022-05-05 16:51:18,217 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
90
+ 2022-05-05 16:51:18,217 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 1
91
+ 2022-05-05 16:51:18,266 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
92
+ 2022-05-05 16:51:18,266 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 1
93
+ 2022-05-05 16:51:18,266 INFO SenderThread:67587 [sender.py:transition_state():392] send defer: 2
94
+ 2022-05-05 16:51:18,266 DEBUG SenderThread:67587 [sender.py:send():235] send: stats
95
+ 2022-05-05 16:51:18,267 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
96
+ 2022-05-05 16:51:18,267 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 2
97
+ 2022-05-05 16:51:18,267 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
98
+ 2022-05-05 16:51:18,268 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 2
99
+ 2022-05-05 16:51:18,268 INFO SenderThread:67587 [sender.py:transition_state():392] send defer: 3
100
+ 2022-05-05 16:51:18,268 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
101
+ 2022-05-05 16:51:18,268 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 3
102
+ 2022-05-05 16:51:18,268 DEBUG SenderThread:67587 [sender.py:send():235] send: summary
103
+ 2022-05-05 16:51:18,269 INFO SenderThread:67587 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
104
+ 2022-05-05 16:51:18,269 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
105
+ 2022-05-05 16:51:18,270 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 3
106
+ 2022-05-05 16:51:18,270 INFO SenderThread:67587 [sender.py:transition_state():392] send defer: 4
107
+ 2022-05-05 16:51:18,270 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
108
+ 2022-05-05 16:51:18,270 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 4
109
+ 2022-05-05 16:51:18,270 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
110
+ 2022-05-05 16:51:18,270 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 4
111
+ 2022-05-05 16:51:18,318 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: poll_exit
112
+ 2022-05-05 16:51:18,620 INFO Thread-8 :67587 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/output.log
113
+ 2022-05-05 16:51:18,620 INFO Thread-8 :67587 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/wandb-summary.json
114
+ 2022-05-05 16:51:18,806 INFO SenderThread:67587 [sender.py:transition_state():392] send defer: 5
115
+ 2022-05-05 16:51:18,807 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: poll_exit
116
+ 2022-05-05 16:51:18,807 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
117
+ 2022-05-05 16:51:18,808 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 5
118
+ 2022-05-05 16:51:18,808 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
119
+ 2022-05-05 16:51:18,808 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 5
120
+ 2022-05-05 16:51:18,808 INFO SenderThread:67587 [dir_watcher.py:finish():283] shutting down directory watcher
121
+ 2022-05-05 16:51:18,909 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: poll_exit
122
+ 2022-05-05 16:51:19,621 INFO Thread-8 :67587 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/config.yaml
123
+ 2022-05-05 16:51:19,621 INFO SenderThread:67587 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files
124
+ 2022-05-05 16:51:19,621 INFO SenderThread:67587 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/wandb-metadata.json wandb-metadata.json
125
+ 2022-05-05 16:51:19,621 INFO SenderThread:67587 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/output.log output.log
126
+ 2022-05-05 16:51:19,622 INFO SenderThread:67587 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/wandb-summary.json wandb-summary.json
127
+ 2022-05-05 16:51:19,622 INFO SenderThread:67587 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/requirements.txt requirements.txt
128
+ 2022-05-05 16:51:19,625 INFO SenderThread:67587 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/config.yaml config.yaml
129
+ 2022-05-05 16:51:19,625 INFO SenderThread:67587 [sender.py:transition_state():392] send defer: 6
130
+ 2022-05-05 16:51:19,626 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: poll_exit
131
+ 2022-05-05 16:51:19,629 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
132
+ 2022-05-05 16:51:19,629 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 6
133
+ 2022-05-05 16:51:19,629 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
134
+ 2022-05-05 16:51:19,629 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 6
135
+ 2022-05-05 16:51:19,629 INFO SenderThread:67587 [file_pusher.py:finish():177] shutting down file pusher
136
+ 2022-05-05 16:51:19,727 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: poll_exit
137
+ 2022-05-05 16:51:19,728 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: poll_exit
138
+ 2022-05-05 16:51:19,830 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: poll_exit
139
+ 2022-05-05 16:51:19,830 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: poll_exit
140
+ 2022-05-05 16:51:19,891 INFO Thread-14 :67587 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/requirements.txt
141
+ 2022-05-05 16:51:19,902 INFO Thread-12 :67587 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/output.log
142
+ 2022-05-05 16:51:19,909 INFO Thread-13 :67587 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/wandb-summary.json
143
+ 2022-05-05 16:51:19,933 INFO Thread-15 :67587 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/files/config.yaml
144
+ 2022-05-05 16:51:19,934 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: poll_exit
145
+ 2022-05-05 16:51:19,934 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: poll_exit
146
+ 2022-05-05 16:51:20,036 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: poll_exit
147
+ 2022-05-05 16:51:20,037 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: poll_exit
148
+ 2022-05-05 16:51:20,134 INFO Thread-7 :67587 [sender.py:transition_state():392] send defer: 7
149
+ 2022-05-05 16:51:20,135 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
150
+ 2022-05-05 16:51:20,135 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 7
151
+ 2022-05-05 16:51:20,135 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
152
+ 2022-05-05 16:51:20,135 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 7
153
+ 2022-05-05 16:51:20,139 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: poll_exit
154
+ 2022-05-05 16:51:20,296 INFO SenderThread:67587 [sender.py:transition_state():392] send defer: 8
155
+ 2022-05-05 16:51:20,296 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: poll_exit
156
+ 2022-05-05 16:51:20,297 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
157
+ 2022-05-05 16:51:20,297 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 8
158
+ 2022-05-05 16:51:20,297 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
159
+ 2022-05-05 16:51:20,297 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 8
160
+ 2022-05-05 16:51:20,297 INFO SenderThread:67587 [sender.py:transition_state():392] send defer: 9
161
+ 2022-05-05 16:51:20,298 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: defer
162
+ 2022-05-05 16:51:20,298 DEBUG SenderThread:67587 [sender.py:send():235] send: final
163
+ 2022-05-05 16:51:20,298 INFO HandlerThread:67587 [handler.py:handle_request_defer():154] handle defer: 9
164
+ 2022-05-05 16:51:20,298 DEBUG SenderThread:67587 [sender.py:send():235] send: footer
165
+ 2022-05-05 16:51:20,299 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: defer
166
+ 2022-05-05 16:51:20,299 INFO SenderThread:67587 [sender.py:send_request_defer():388] handle sender defer: 9
167
+ 2022-05-05 16:51:20,398 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: poll_exit
168
+ 2022-05-05 16:51:20,398 DEBUG SenderThread:67587 [sender.py:send_request():249] send_request: poll_exit
169
+ 2022-05-05 16:51:20,398 INFO SenderThread:67587 [file_pusher.py:join():182] waiting for file pusher
170
+ 2022-05-05 16:51:21,569 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: get_summary
171
+ 2022-05-05 16:51:21,570 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: sampled_history
172
+ 2022-05-05 16:51:21,571 DEBUG HandlerThread:67587 [handler.py:handle_request():131] handle_request: shutdown
173
+ 2022-05-05 16:51:21,571 INFO HandlerThread:67587 [handler.py:finish():739] shutting down handler
174
+ 2022-05-05 16:51:22,298 INFO WriterThread:67587 [datastore.py:close():281] close: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_164902-lr7oonrp/run-lr7oonrp.wandb
175
+ 2022-05-05 16:51:22,568 INFO SenderThread:67587 [sender.py:finish():1075] shutting down sender
176
+ 2022-05-05 16:51:22,568 INFO SenderThread:67587 [file_pusher.py:finish():177] shutting down file pusher
177
+ 2022-05-05 16:51:22,568 INFO SenderThread:67587 [file_pusher.py:join():182] waiting for file pusher
178
+ 2022-05-05 16:51:22,570 INFO MainThread:67587 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220505_164902-lr7oonrp/logs/debug.log CHANGED
@@ -26,3 +26,85 @@ config: {}
26
  2022-05-05 16:49:06,070 INFO MainThread:67410 [wandb_init.py:init():651] run started, returning control to user process
27
  2022-05-05 16:49:06,073 INFO MainThread:67410 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 40, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50265, 'max_position_embeddings': 1024, 'd_model': 1024, 'encoder_ffn_dim': 4096, 'encoder_layers': 12, 'encoder_attention_heads': 16, 'decoder_ffn_dim': 4096, 'decoder_layers': 12, 'decoder_attention_heads': 16, 'dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.1, 'activation_function': 'gelu', 'init_std': 0.02, 'encoder_layerdrop': 0.0, 'decoder_layerdrop': 0.0, 'classifier_dropout': 0.0, 'use_cache': True, 'num_hidden_layers': 12, 'scale_embedding': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': True, 'num_beams': 4, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 3, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': 0, 'forced_eos_token_id': 2, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['BartModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1', 2: 'LABEL_2'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': 2, 'task_specific_params': {'summarization': {'length_penalty': 1.0, 'max_length': 128, 'min_length': 12, 'num_beams': 4}, 'summarization_cnn': {'length_penalty': 2.0, 'max_length': 142, 'min_length': 56, 'num_beams': 4}, 'summarization_xsum': {'length_penalty': 1.0, 'max_length': 62, 'min_length': 11, 'num_beams': 6}}, 'problem_type': None, '_name_or_path': 'facebook/bart-large', 'transformers_version': '4.19.0.dev0', 'add_bias_logits': False, 'add_final_layer_norm': False, 'classif_dropout': 0.1, 'gradient_checkpointing': False, 'normalize_before': False, 'model_type': 'bart'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.19.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.16627274455555233, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.1, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'prediction_loss_only': False, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'eval_accumulation_steps': 'None', 'eval_delay': 0, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/May05_16-48-07_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_total_limit': 'None', 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'data_seed': 'None', 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': "['tensorboard', 'wandb', 'codecarbon']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'gradient_checkpointing': True, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'train_batch_size': 8, 'eval_batch_size': 8}
28
  2022-05-05 16:49:06,076 INFO MainThread:67410 [wandb_watch.py:watch():43] Watching
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  2022-05-05 16:49:06,070 INFO MainThread:67410 [wandb_init.py:init():651] run started, returning control to user process
27
  2022-05-05 16:49:06,073 INFO MainThread:67410 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 40, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50265, 'max_position_embeddings': 1024, 'd_model': 1024, 'encoder_ffn_dim': 4096, 'encoder_layers': 12, 'encoder_attention_heads': 16, 'decoder_ffn_dim': 4096, 'decoder_layers': 12, 'decoder_attention_heads': 16, 'dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.1, 'activation_function': 'gelu', 'init_std': 0.02, 'encoder_layerdrop': 0.0, 'decoder_layerdrop': 0.0, 'classifier_dropout': 0.0, 'use_cache': True, 'num_hidden_layers': 12, 'scale_embedding': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': True, 'num_beams': 4, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 3, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': 0, 'forced_eos_token_id': 2, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['BartModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1', 2: 'LABEL_2'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': 2, 'task_specific_params': {'summarization': {'length_penalty': 1.0, 'max_length': 128, 'min_length': 12, 'num_beams': 4}, 'summarization_cnn': {'length_penalty': 2.0, 'max_length': 142, 'min_length': 56, 'num_beams': 4}, 'summarization_xsum': {'length_penalty': 1.0, 'max_length': 62, 'min_length': 11, 'num_beams': 6}}, 'problem_type': None, '_name_or_path': 'facebook/bart-large', 'transformers_version': '4.19.0.dev0', 'add_bias_logits': False, 'add_final_layer_norm': False, 'classif_dropout': 0.1, 'gradient_checkpointing': False, 'normalize_before': False, 'model_type': 'bart'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.19.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.16627274455555233, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.1, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'prediction_loss_only': False, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'eval_accumulation_steps': 'None', 'eval_delay': 0, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/May05_16-48-07_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_total_limit': 'None', 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'data_seed': 'None', 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': "['tensorboard', 'wandb', 'codecarbon']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'gradient_checkpointing': True, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'train_batch_size': 8, 'eval_batch_size': 8}
28
  2022-05-05 16:49:06,076 INFO MainThread:67410 [wandb_watch.py:watch():43] Watching
29
+ 2022-05-05 16:51:15,383 INFO MainThread:67410 [wandb_run.py:_atexit_cleanup():1797] got exitcode: 1
30
+ 2022-05-05 16:51:15,388 INFO MainThread:67410 [wandb_run.py:_restore():1769] restore
31
+ 2022-05-05 16:51:18,216 INFO MainThread:67410 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
32
+ wandb_count: 1
33
+ }
34
+ pusher_stats {
35
+ uploaded_bytes: 1972
36
+ total_bytes: 1972
37
+ }
38
+
39
+ 2022-05-05 16:51:18,807 INFO MainThread:67410 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
40
+ wandb_count: 1
41
+ }
42
+ pusher_stats {
43
+ uploaded_bytes: 1972
44
+ total_bytes: 1972
45
+ }
46
+
47
+ 2022-05-05 16:51:19,626 INFO MainThread:67410 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
48
+ wandb_count: 5
49
+ }
50
+ pusher_stats {
51
+ uploaded_bytes: 1972
52
+ total_bytes: 27373
53
+ }
54
+
55
+ 2022-05-05 16:51:19,729 INFO MainThread:67410 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
56
+ wandb_count: 5
57
+ }
58
+ pusher_stats {
59
+ uploaded_bytes: 27373
60
+ total_bytes: 27373
61
+ }
62
+
63
+ 2022-05-05 16:51:19,831 INFO MainThread:67410 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
64
+ wandb_count: 5
65
+ }
66
+ pusher_stats {
67
+ uploaded_bytes: 27373
68
+ total_bytes: 27373
69
+ }
70
+
71
+ 2022-05-05 16:51:19,935 INFO MainThread:67410 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
72
+ wandb_count: 5
73
+ }
74
+ pusher_stats {
75
+ uploaded_bytes: 27373
76
+ total_bytes: 27373
77
+ }
78
+
79
+ 2022-05-05 16:51:20,038 INFO MainThread:67410 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
80
+ wandb_count: 5
81
+ }
82
+ pusher_stats {
83
+ uploaded_bytes: 27373
84
+ total_bytes: 27373
85
+ }
86
+
87
+ 2022-05-05 16:51:20,297 INFO MainThread:67410 [wandb_run.py:_wait_for_finish():1929] got exit ret: file_counts {
88
+ wandb_count: 5
89
+ }
90
+ pusher_stats {
91
+ uploaded_bytes: 27373
92
+ total_bytes: 27373
93
+ }
94
+
95
+ 2022-05-05 16:51:21,568 INFO MainThread:67410 [wandb_run.py:_wait_for_finish():1929] got exit ret: done: true
96
+ exit_result {
97
+ }
98
+ file_counts {
99
+ wandb_count: 5
100
+ }
101
+ pusher_stats {
102
+ uploaded_bytes: 27373
103
+ total_bytes: 27373
104
+ }
105
+ local_info {
106
+ }
107
+
108
+ 2022-05-05 16:51:23,149 INFO MainThread:67410 [wandb_run.py:_append_history():2144] rendering history
109
+ 2022-05-05 16:51:23,149 INFO MainThread:67410 [wandb_run.py:_append_summary():2102] rendering summary
110
+ 2022-05-05 16:51:23,149 INFO MainThread:67410 [wandb_run.py:_append_files():2194] logging synced files
wandb/run-20220505_164902-lr7oonrp/run-lr7oonrp.wandb CHANGED
Binary files a/wandb/run-20220505_164902-lr7oonrp/run-lr7oonrp.wandb and b/wandb/run-20220505_164902-lr7oonrp/run-lr7oonrp.wandb differ
wandb/run-20220505_173213-ykkm1chu/files/config.yaml ADDED
@@ -0,0 +1,788 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: ./
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.10
13
+ framework: huggingface
14
+ huggingface_version: 4.19.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ - 1: train/train_runtime
22
+ 5: 1
23
+ 6:
24
+ - 1
25
+ - 1: train/train_samples_per_second
26
+ 5: 1
27
+ 6:
28
+ - 1
29
+ - 1: train/train_steps_per_second
30
+ 5: 1
31
+ 6:
32
+ - 1
33
+ - 1: train/total_flos
34
+ 5: 1
35
+ 6:
36
+ - 1
37
+ - 1: train/train_loss
38
+ 5: 1
39
+ 6:
40
+ - 1
41
+ - 1: train/epoch
42
+ 5: 1
43
+ 6:
44
+ - 1
45
+ python_version: 3.9.5
46
+ start_time: 1651771933
47
+ t:
48
+ 1:
49
+ - 1
50
+ - 2
51
+ - 3
52
+ - 5
53
+ - 11
54
+ - 12
55
+ 3:
56
+ - 13
57
+ 4: 3.9.5
58
+ 5: 0.12.10
59
+ 6: 4.19.0.dev0
60
+ 8:
61
+ - 5
62
+ adafactor:
63
+ desc: null
64
+ value: false
65
+ adam_beta1:
66
+ desc: null
67
+ value: 0.9
68
+ adam_beta2:
69
+ desc: null
70
+ value: 0.999
71
+ adam_epsilon:
72
+ desc: null
73
+ value: 1.0e-08
74
+ add_cross_attention:
75
+ desc: null
76
+ value: false
77
+ architectures:
78
+ desc: null
79
+ value:
80
+ - SpeechEncoderDecoderModel
81
+ bad_words_ids:
82
+ desc: null
83
+ value: null
84
+ bf16:
85
+ desc: null
86
+ value: false
87
+ bf16_full_eval:
88
+ desc: null
89
+ value: false
90
+ bos_token_id:
91
+ desc: null
92
+ value: null
93
+ chunk_size_feed_forward:
94
+ desc: null
95
+ value: 0
96
+ cross_attention_hidden_size:
97
+ desc: null
98
+ value: null
99
+ data_seed:
100
+ desc: null
101
+ value: None
102
+ dataloader_drop_last:
103
+ desc: null
104
+ value: false
105
+ dataloader_num_workers:
106
+ desc: null
107
+ value: 0
108
+ dataloader_pin_memory:
109
+ desc: null
110
+ value: true
111
+ ddp_bucket_cap_mb:
112
+ desc: null
113
+ value: None
114
+ ddp_find_unused_parameters:
115
+ desc: null
116
+ value: None
117
+ debug:
118
+ desc: null
119
+ value: '[]'
120
+ decoder:
121
+ desc: null
122
+ value:
123
+ _name_or_path: facebook/bart-large
124
+ activation_dropout: 0.1
125
+ activation_function: gelu
126
+ add_bias_logits: false
127
+ add_cross_attention: true
128
+ add_final_layer_norm: false
129
+ architectures:
130
+ - BartModel
131
+ attention_dropout: 0.1
132
+ bad_words_ids: null
133
+ bos_token_id: 0
134
+ chunk_size_feed_forward: 0
135
+ classif_dropout: 0.1
136
+ classifier_dropout: 0.0
137
+ cross_attention_hidden_size: null
138
+ d_model: 1024
139
+ decoder_attention_heads: 16
140
+ decoder_ffn_dim: 4096
141
+ decoder_layerdrop: 0.0
142
+ decoder_layers: 12
143
+ decoder_start_token_id: 2
144
+ diversity_penalty: 0.0
145
+ do_sample: false
146
+ dropout: 0.1
147
+ early_stopping: true
148
+ encoder_attention_heads: 16
149
+ encoder_ffn_dim: 4096
150
+ encoder_layerdrop: 0.0
151
+ encoder_layers: 12
152
+ encoder_no_repeat_ngram_size: 0
153
+ eos_token_id: 2
154
+ exponential_decay_length_penalty: null
155
+ finetuning_task: null
156
+ forced_bos_token_id: 0
157
+ forced_eos_token_id: 2
158
+ gradient_checkpointing: false
159
+ id2label:
160
+ '0': LABEL_0
161
+ '1': LABEL_1
162
+ '2': LABEL_2
163
+ init_std: 0.02
164
+ is_decoder: true
165
+ is_encoder_decoder: false
166
+ label2id:
167
+ LABEL_0: 0
168
+ LABEL_1: 1
169
+ LABEL_2: 2
170
+ length_penalty: 1.0
171
+ max_length: 20
172
+ max_position_embeddings: 1024
173
+ min_length: 0
174
+ model_type: bart
175
+ no_repeat_ngram_size: 3
176
+ normalize_before: false
177
+ num_beam_groups: 1
178
+ num_beams: 4
179
+ num_hidden_layers: 12
180
+ num_return_sequences: 1
181
+ output_attentions: false
182
+ output_hidden_states: false
183
+ output_scores: false
184
+ pad_token_id: 1
185
+ prefix: null
186
+ problem_type: null
187
+ pruned_heads: {}
188
+ remove_invalid_values: false
189
+ repetition_penalty: 1.0
190
+ return_dict: true
191
+ return_dict_in_generate: false
192
+ scale_embedding: false
193
+ sep_token_id: null
194
+ task_specific_params:
195
+ summarization:
196
+ length_penalty: 1.0
197
+ max_length: 128
198
+ min_length: 12
199
+ num_beams: 4
200
+ summarization_cnn:
201
+ length_penalty: 2.0
202
+ max_length: 142
203
+ min_length: 56
204
+ num_beams: 4
205
+ summarization_xsum:
206
+ length_penalty: 1.0
207
+ max_length: 62
208
+ min_length: 11
209
+ num_beams: 6
210
+ temperature: 1.0
211
+ tie_encoder_decoder: false
212
+ tie_word_embeddings: true
213
+ tokenizer_class: null
214
+ top_k: 50
215
+ top_p: 1.0
216
+ torch_dtype: null
217
+ torchscript: false
218
+ transformers_version: 4.19.0.dev0
219
+ typical_p: 1.0
220
+ use_bfloat16: false
221
+ use_cache: true
222
+ vocab_size: 50265
223
+ decoder_start_token_id:
224
+ desc: null
225
+ value: 0
226
+ deepspeed:
227
+ desc: null
228
+ value: None
229
+ disable_tqdm:
230
+ desc: null
231
+ value: false
232
+ diversity_penalty:
233
+ desc: null
234
+ value: 0.0
235
+ do_eval:
236
+ desc: null
237
+ value: true
238
+ do_predict:
239
+ desc: null
240
+ value: false
241
+ do_sample:
242
+ desc: null
243
+ value: false
244
+ do_train:
245
+ desc: null
246
+ value: true
247
+ early_stopping:
248
+ desc: null
249
+ value: false
250
+ encoder:
251
+ desc: null
252
+ value:
253
+ _name_or_path: facebook/wav2vec2-xls-r-300m
254
+ activation_dropout: 0.0
255
+ adapter_kernel_size: 3
256
+ adapter_stride: 2
257
+ add_adapter: true
258
+ add_cross_attention: false
259
+ apply_spec_augment: true
260
+ architectures:
261
+ - Wav2Vec2ForPreTraining
262
+ attention_dropout: 0.1
263
+ bad_words_ids: null
264
+ bos_token_id: 1
265
+ chunk_size_feed_forward: 0
266
+ classifier_proj_size: 256
267
+ codevector_dim: 768
268
+ contrastive_logits_temperature: 0.1
269
+ conv_bias: true
270
+ conv_dim:
271
+ - 512
272
+ - 512
273
+ - 512
274
+ - 512
275
+ - 512
276
+ - 512
277
+ - 512
278
+ conv_kernel:
279
+ - 10
280
+ - 3
281
+ - 3
282
+ - 3
283
+ - 3
284
+ - 2
285
+ - 2
286
+ conv_stride:
287
+ - 5
288
+ - 2
289
+ - 2
290
+ - 2
291
+ - 2
292
+ - 2
293
+ - 2
294
+ cross_attention_hidden_size: null
295
+ ctc_loss_reduction: sum
296
+ ctc_zero_infinity: false
297
+ decoder_start_token_id: null
298
+ diversity_loss_weight: 0.1
299
+ diversity_penalty: 0.0
300
+ do_sample: false
301
+ do_stable_layer_norm: true
302
+ early_stopping: false
303
+ encoder_no_repeat_ngram_size: 0
304
+ eos_token_id: 2
305
+ exponential_decay_length_penalty: null
306
+ feat_extract_activation: gelu
307
+ feat_extract_dropout: 0.0
308
+ feat_extract_norm: layer
309
+ feat_proj_dropout: 0.0
310
+ feat_quantizer_dropout: 0.0
311
+ final_dropout: 0.0
312
+ finetuning_task: null
313
+ forced_bos_token_id: null
314
+ forced_eos_token_id: null
315
+ gradient_checkpointing: false
316
+ hidden_act: gelu
317
+ hidden_dropout: 0.2
318
+ hidden_size: 1024
319
+ id2label:
320
+ '0': LABEL_0
321
+ '1': LABEL_1
322
+ initializer_range: 0.02
323
+ intermediate_size: 4096
324
+ is_decoder: false
325
+ is_encoder_decoder: false
326
+ label2id:
327
+ LABEL_0: 0
328
+ LABEL_1: 1
329
+ layer_norm_eps: 1.0e-05
330
+ layerdrop: 0.0
331
+ length_penalty: 1.0
332
+ mask_feature_length: 10
333
+ mask_feature_min_masks: 0
334
+ mask_feature_prob: 0.0
335
+ mask_time_length: 10
336
+ mask_time_min_masks: 2
337
+ mask_time_prob: 0.1
338
+ max_length: 20
339
+ min_length: 0
340
+ model_type: wav2vec2
341
+ no_repeat_ngram_size: 0
342
+ num_adapter_layers: 3
343
+ num_attention_heads: 16
344
+ num_beam_groups: 1
345
+ num_beams: 1
346
+ num_codevector_groups: 2
347
+ num_codevectors_per_group: 320
348
+ num_conv_pos_embedding_groups: 16
349
+ num_conv_pos_embeddings: 128
350
+ num_feat_extract_layers: 7
351
+ num_hidden_layers: 24
352
+ num_negatives: 100
353
+ num_return_sequences: 1
354
+ output_attentions: false
355
+ output_hidden_size: 1024
356
+ output_hidden_states: false
357
+ output_scores: false
358
+ pad_token_id: 0
359
+ prefix: null
360
+ problem_type: null
361
+ proj_codevector_dim: 768
362
+ pruned_heads: {}
363
+ remove_invalid_values: false
364
+ repetition_penalty: 1.0
365
+ return_dict: true
366
+ return_dict_in_generate: false
367
+ sep_token_id: null
368
+ task_specific_params: null
369
+ tdnn_dilation:
370
+ - 1
371
+ - 2
372
+ - 3
373
+ - 1
374
+ - 1
375
+ tdnn_dim:
376
+ - 512
377
+ - 512
378
+ - 512
379
+ - 512
380
+ - 1500
381
+ tdnn_kernel:
382
+ - 5
383
+ - 3
384
+ - 3
385
+ - 1
386
+ - 1
387
+ temperature: 1.0
388
+ tie_encoder_decoder: false
389
+ tie_word_embeddings: true
390
+ tokenizer_class: null
391
+ top_k: 50
392
+ top_p: 1.0
393
+ torch_dtype: float32
394
+ torchscript: false
395
+ transformers_version: 4.19.0.dev0
396
+ typical_p: 1.0
397
+ use_bfloat16: false
398
+ use_weighted_layer_sum: false
399
+ vocab_size: 32
400
+ xvector_output_dim: 512
401
+ encoder_no_repeat_ngram_size:
402
+ desc: null
403
+ value: 0
404
+ eos_token_id:
405
+ desc: null
406
+ value: 2
407
+ eval_accumulation_steps:
408
+ desc: null
409
+ value: None
410
+ eval_batch_size:
411
+ desc: null
412
+ value: 8
413
+ eval_delay:
414
+ desc: null
415
+ value: 0
416
+ eval_split_name:
417
+ desc: null
418
+ value: test
419
+ eval_steps:
420
+ desc: null
421
+ value: 500
422
+ evaluation_strategy:
423
+ desc: null
424
+ value: steps
425
+ exponential_decay_length_penalty:
426
+ desc: null
427
+ value: null
428
+ finetuning_task:
429
+ desc: null
430
+ value: null
431
+ forced_bos_token_id:
432
+ desc: null
433
+ value: null
434
+ forced_eos_token_id:
435
+ desc: null
436
+ value: null
437
+ fp16:
438
+ desc: null
439
+ value: true
440
+ fp16_backend:
441
+ desc: null
442
+ value: auto
443
+ fp16_full_eval:
444
+ desc: null
445
+ value: false
446
+ fp16_opt_level:
447
+ desc: null
448
+ value: O1
449
+ generation_max_length:
450
+ desc: null
451
+ value: 40
452
+ generation_num_beams:
453
+ desc: null
454
+ value: 1
455
+ gradient_accumulation_steps:
456
+ desc: null
457
+ value: 16
458
+ gradient_checkpointing:
459
+ desc: null
460
+ value: true
461
+ greater_is_better:
462
+ desc: null
463
+ value: true
464
+ group_by_length:
465
+ desc: null
466
+ value: true
467
+ half_precision_backend:
468
+ desc: null
469
+ value: amp
470
+ hidden_dropout:
471
+ desc: null
472
+ value: 0.2
473
+ hub_model_id:
474
+ desc: null
475
+ value: None
476
+ hub_private_repo:
477
+ desc: null
478
+ value: false
479
+ hub_strategy:
480
+ desc: null
481
+ value: every_save
482
+ hub_token:
483
+ desc: null
484
+ value: <HUB_TOKEN>
485
+ id2label:
486
+ desc: null
487
+ value:
488
+ '0': LABEL_0
489
+ '1': LABEL_1
490
+ ignore_data_skip:
491
+ desc: null
492
+ value: false
493
+ include_inputs_for_metrics:
494
+ desc: null
495
+ value: false
496
+ is_decoder:
497
+ desc: null
498
+ value: false
499
+ is_encoder_decoder:
500
+ desc: null
501
+ value: true
502
+ label2id:
503
+ desc: null
504
+ value:
505
+ LABEL_0: 0
506
+ LABEL_1: 1
507
+ label_names:
508
+ desc: null
509
+ value: None
510
+ label_smoothing_factor:
511
+ desc: null
512
+ value: 0.0
513
+ language:
514
+ desc: null
515
+ value: fr.en
516
+ learning_rate:
517
+ desc: null
518
+ value: 0.0003287457929573604
519
+ length_column_name:
520
+ desc: null
521
+ value: length
522
+ length_penalty:
523
+ desc: null
524
+ value: 1.0
525
+ load_best_model_at_end:
526
+ desc: null
527
+ value: true
528
+ local_rank:
529
+ desc: null
530
+ value: -1
531
+ log_level:
532
+ desc: null
533
+ value: -1
534
+ log_level_replica:
535
+ desc: null
536
+ value: -1
537
+ log_on_each_node:
538
+ desc: null
539
+ value: true
540
+ logging_dir:
541
+ desc: null
542
+ value: ./runs/May05_17-31-19_sanchit--v100
543
+ logging_first_step:
544
+ desc: null
545
+ value: false
546
+ logging_nan_inf_filter:
547
+ desc: null
548
+ value: true
549
+ logging_steps:
550
+ desc: null
551
+ value: 1
552
+ logging_strategy:
553
+ desc: null
554
+ value: steps
555
+ lr_scheduler_type:
556
+ desc: null
557
+ value: linear
558
+ max_duration_in_seconds:
559
+ desc: null
560
+ value: 20
561
+ max_grad_norm:
562
+ desc: null
563
+ value: 1.0
564
+ max_length:
565
+ desc: null
566
+ value: 40
567
+ max_steps:
568
+ desc: null
569
+ value: -1
570
+ metric_for_best_model:
571
+ desc: null
572
+ value: bleu
573
+ min_length:
574
+ desc: null
575
+ value: 0
576
+ model_name_or_path:
577
+ desc: null
578
+ value: ./
579
+ model_type:
580
+ desc: null
581
+ value: speech-encoder-decoder
582
+ mp_parameters:
583
+ desc: null
584
+ value: ''
585
+ no_cuda:
586
+ desc: null
587
+ value: false
588
+ no_repeat_ngram_size:
589
+ desc: null
590
+ value: 0
591
+ num_beam_groups:
592
+ desc: null
593
+ value: 1
594
+ num_beams:
595
+ desc: null
596
+ value: 1
597
+ num_return_sequences:
598
+ desc: null
599
+ value: 1
600
+ num_train_epochs:
601
+ desc: null
602
+ value: 3
603
+ optim:
604
+ desc: null
605
+ value: adamw_hf
606
+ output_attentions:
607
+ desc: null
608
+ value: false
609
+ output_dir:
610
+ desc: null
611
+ value: ./
612
+ output_hidden_states:
613
+ desc: null
614
+ value: false
615
+ output_scores:
616
+ desc: null
617
+ value: false
618
+ overwrite_output_dir:
619
+ desc: null
620
+ value: true
621
+ pad_token_id:
622
+ desc: null
623
+ value: 1
624
+ past_index:
625
+ desc: null
626
+ value: -1
627
+ per_device_eval_batch_size:
628
+ desc: null
629
+ value: 8
630
+ per_device_train_batch_size:
631
+ desc: null
632
+ value: 8
633
+ per_gpu_eval_batch_size:
634
+ desc: null
635
+ value: None
636
+ per_gpu_train_batch_size:
637
+ desc: null
638
+ value: None
639
+ predict_with_generate:
640
+ desc: null
641
+ value: true
642
+ prediction_loss_only:
643
+ desc: null
644
+ value: false
645
+ prefix:
646
+ desc: null
647
+ value: null
648
+ problem_type:
649
+ desc: null
650
+ value: null
651
+ processor_class:
652
+ desc: null
653
+ value: Wav2Vec2Processor
654
+ pruned_heads:
655
+ desc: null
656
+ value: {}
657
+ push_to_hub:
658
+ desc: null
659
+ value: true
660
+ push_to_hub_model_id:
661
+ desc: null
662
+ value: None
663
+ push_to_hub_organization:
664
+ desc: null
665
+ value: None
666
+ push_to_hub_token:
667
+ desc: null
668
+ value: <PUSH_TO_HUB_TOKEN>
669
+ remove_invalid_values:
670
+ desc: null
671
+ value: false
672
+ remove_unused_columns:
673
+ desc: null
674
+ value: true
675
+ repetition_penalty:
676
+ desc: null
677
+ value: 1.0
678
+ report_to:
679
+ desc: null
680
+ value: '[''tensorboard'', ''wandb'', ''codecarbon'']'
681
+ resume_from_checkpoint:
682
+ desc: null
683
+ value: None
684
+ return_dict:
685
+ desc: null
686
+ value: true
687
+ return_dict_in_generate:
688
+ desc: null
689
+ value: false
690
+ run_name:
691
+ desc: null
692
+ value: ./
693
+ save_on_each_node:
694
+ desc: null
695
+ value: false
696
+ save_steps:
697
+ desc: null
698
+ value: 500
699
+ save_strategy:
700
+ desc: null
701
+ value: steps
702
+ save_total_limit:
703
+ desc: null
704
+ value: None
705
+ seed:
706
+ desc: null
707
+ value: 42
708
+ sep_token_id:
709
+ desc: null
710
+ value: null
711
+ sharded_ddp:
712
+ desc: null
713
+ value: '[]'
714
+ skip_memory_metrics:
715
+ desc: null
716
+ value: true
717
+ sortish_sampler:
718
+ desc: null
719
+ value: false
720
+ task:
721
+ desc: null
722
+ value: covost2
723
+ task_specific_params:
724
+ desc: null
725
+ value: null
726
+ temperature:
727
+ desc: null
728
+ value: 1.0
729
+ tf32:
730
+ desc: null
731
+ value: None
732
+ tie_encoder_decoder:
733
+ desc: null
734
+ value: false
735
+ tie_word_embeddings:
736
+ desc: null
737
+ value: false
738
+ tokenizer_class:
739
+ desc: null
740
+ value: null
741
+ top_k:
742
+ desc: null
743
+ value: 50
744
+ top_p:
745
+ desc: null
746
+ value: 1.0
747
+ torch_dtype:
748
+ desc: null
749
+ value: torch.float32
750
+ torchscript:
751
+ desc: null
752
+ value: false
753
+ tpu_metrics_debug:
754
+ desc: null
755
+ value: false
756
+ tpu_num_cores:
757
+ desc: null
758
+ value: None
759
+ train_batch_size:
760
+ desc: null
761
+ value: 8
762
+ transformers_version:
763
+ desc: null
764
+ value: null
765
+ typical_p:
766
+ desc: null
767
+ value: 1.0
768
+ use_bfloat16:
769
+ desc: null
770
+ value: false
771
+ use_cache:
772
+ desc: null
773
+ value: false
774
+ use_legacy_prediction_loop:
775
+ desc: null
776
+ value: false
777
+ warmup_ratio:
778
+ desc: null
779
+ value: 0.0
780
+ warmup_steps:
781
+ desc: null
782
+ value: 500
783
+ weight_decay:
784
+ desc: null
785
+ value: 0.0
786
+ xpu_backend:
787
+ desc: null
788
+ value: None
wandb/run-20220505_173213-ykkm1chu/files/output.log ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb: WARNING Config item 'output_dir' was locked by 'sweep' (ignored update).
2
+ wandb: WARNING Config item 'evaluation_strategy' was locked by 'sweep' (ignored update).
3
+ wandb: WARNING Config item 'per_device_train_batch_size' was locked by 'sweep' (ignored update).
4
+ wandb: WARNING Config item 'per_device_eval_batch_size' was locked by 'sweep' (ignored update).
5
+ wandb: WARNING Config item 'gradient_accumulation_steps' was locked by 'sweep' (ignored update).
6
+ wandb: WARNING Config item 'learning_rate' was locked by 'sweep' (ignored update).
7
+ wandb: WARNING Config item 'num_train_epochs' was locked by 'sweep' (ignored update).
8
+ wandb: WARNING Config item 'warmup_steps' was locked by 'sweep' (ignored update).
9
+ wandb: WARNING Config item 'logging_steps' was locked by 'sweep' (ignored update).
10
+ wandb: WARNING Config item 'save_steps' was locked by 'sweep' (ignored update).
11
+ wandb: WARNING Config item 'eval_steps' was locked by 'sweep' (ignored update).
12
+ wandb: WARNING Config item 'metric_for_best_model' was locked by 'sweep' (ignored update).
13
+ wandb: WARNING Config item 'greater_is_better' was locked by 'sweep' (ignored update).
14
+ wandb: WARNING Config item 'generation_max_length' was locked by 'sweep' (ignored update).
15
+ wandb: WARNING Config item 'generation_num_beams' was locked by 'sweep' (ignored update).
16
+ 0%| | 0/4860 [00:00<?, ?it/s]
17
+ {'train_runtime': 8.4243, 'train_samples_per_second': 73848.5, 'train_steps_per_second': 576.902, 'train_loss': 0.0, 'epoch': 3.0}
18
+ Training completed. Do not forget to share your model on huggingface.co/models =)
19
+ Loading best model from ./checkpoint-1000 (score: 2.4961869532998874e-13).
20
+ Could not locate the best model at ./checkpoint-1000/pytorch_model.bin, if you are running a distributed training on multiple nodes, you should activate `--save_on_each_node`.
21
+ 0%| | 0/4860 [00:04<?, ?it/s]
22
+ Skipping the first batches: : 0it [00:08, ?it/s] | 0/4860 [00:04<?, ?it/s]
23
+ Saving model checkpoint to ./
24
+ Configuration saved in ./config.json
25
+ Model weights saved in ./pytorch_model.bin
26
+ Feature extractor saved in ./preprocessor_config.json
27
+ Saving model checkpoint to ./
28
+ Configuration saved in ./config.json
29
+ Model weights saved in ./pytorch_model.bin
wandb/run-20220505_173213-ykkm1chu/files/requirements.txt ADDED
@@ -0,0 +1,287 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ alembic==1.7.7
5
+ anyio==3.5.0
6
+ appdirs==1.4.4
7
+ apscheduler==3.9.1
8
+ argon2-cffi-bindings==21.2.0
9
+ argon2-cffi==21.3.0
10
+ arrow==1.2.2
11
+ asttokens==2.0.5
12
+ astunparse==1.6.3
13
+ async-timeout==4.0.2
14
+ attrs==21.4.0
15
+ audioread==2.1.9
16
+ autopage==0.5.0
17
+ babel==2.9.1
18
+ backcall==0.2.0
19
+ backoff==1.11.1
20
+ binaryornot==0.4.4
21
+ bitsandbytes-cuda113==0.26.0
22
+ black==22.1.0
23
+ bleach==4.1.0
24
+ boto3==1.16.34
25
+ botocore==1.19.63
26
+ brotli==1.0.9
27
+ cachetools==5.0.0
28
+ certifi==2021.10.8
29
+ cffi==1.15.0
30
+ chardet==4.0.0
31
+ charset-normalizer==2.0.11
32
+ chex==0.1.0
33
+ click==8.0.3
34
+ cliff==3.10.1
35
+ clldutils==3.10.1
36
+ cmaes==0.8.2
37
+ cmd2==2.4.0
38
+ codecarbon==1.2.0
39
+ colorlog==6.6.0
40
+ cookiecutter==1.7.3
41
+ cryptography==36.0.2
42
+ csvw==1.11.0
43
+ cycler==0.11.0
44
+ dash-bootstrap-components==1.1.0
45
+ dash-core-components==2.0.0
46
+ dash-html-components==2.0.0
47
+ dash-table==5.0.0
48
+ dash==2.3.1
49
+ datasets==2.1.1.dev0
50
+ debugpy==1.5.1
51
+ decorator==5.1.1
52
+ defusedxml==0.7.1
53
+ deprecated==1.2.13
54
+ dill==0.3.4
55
+ dlinfo==1.2.1
56
+ dm-tree==0.1.6
57
+ docker-pycreds==0.4.0
58
+ docker==4.4.4
59
+ entrypoints==0.4
60
+ execnet==1.9.0
61
+ executing==0.8.2
62
+ faiss-cpu==1.7.2
63
+ filelock==3.4.2
64
+ fire==0.4.0
65
+ flake8==4.0.1
66
+ flask-compress==1.11
67
+ flask==2.1.1
68
+ flatbuffers==1.12
69
+ flax==0.4.0
70
+ fonttools==4.29.1
71
+ frozenlist==1.3.0
72
+ fsspec==2022.1.0
73
+ fugashi==1.1.2
74
+ gast==0.5.3
75
+ gitdb==4.0.9
76
+ gitpython==3.1.18
77
+ google-auth-oauthlib==0.4.6
78
+ google-auth==2.6.0
79
+ google-pasta==0.2.0
80
+ greenlet==1.1.2
81
+ grpcio==1.43.0
82
+ h5py==3.6.0
83
+ hf-doc-builder==0.2.0
84
+ huggingface-hub==0.4.0
85
+ hypothesis==6.36.1
86
+ idna==3.3
87
+ importlib-metadata==4.10.1
88
+ iniconfig==1.1.1
89
+ ipadic==1.0.0
90
+ ipdb==0.13.9
91
+ ipykernel==6.8.0
92
+ ipython-genutils==0.2.0
93
+ ipython==8.0.1
94
+ ipywidgets==7.6.5
95
+ isodate==0.6.1
96
+ isort==5.10.1
97
+ itsdangerous==2.1.2
98
+ jax==0.2.28
99
+ jaxlib==0.1.76+cuda11.cudnn82
100
+ jedi==0.18.1
101
+ jinja2-time==0.2.0
102
+ jinja2==3.0.3
103
+ jiwer==2.3.0
104
+ jmespath==0.10.0
105
+ joblib==1.1.0
106
+ json5==0.9.6
107
+ jsonschema==4.4.0
108
+ jupyter-client==7.1.2
109
+ jupyter-console==6.4.0
110
+ jupyter-core==4.9.1
111
+ jupyter-server==1.13.5
112
+ jupyter==1.0.0
113
+ jupyterlab-pygments==0.1.2
114
+ jupyterlab-server==2.10.3
115
+ jupyterlab-widgets==1.0.2
116
+ jupyterlab==3.2.9
117
+ keras-preprocessing==1.1.2
118
+ keras==2.8.0
119
+ kiwisolver==1.3.2
120
+ kubernetes==12.0.1
121
+ libclang==13.0.0
122
+ librosa==0.8.1
123
+ llvmlite==0.38.0
124
+ mako==1.2.0
125
+ markdown==3.3.6
126
+ markupsafe==2.0.1
127
+ matplotlib-inline==0.1.3
128
+ matplotlib==3.5.1
129
+ mccabe==0.6.1
130
+ mistune==0.8.4
131
+ msgpack==1.0.3
132
+ multidict==6.0.2
133
+ multiprocess==0.70.12.2
134
+ mypy-extensions==0.4.3
135
+ nbclassic==0.3.5
136
+ nbclient==0.5.10
137
+ nbconvert==6.4.1
138
+ nbformat==5.1.3
139
+ nest-asyncio==1.5.4
140
+ nltk==3.7
141
+ notebook==6.4.8
142
+ numba==0.55.1
143
+ numpy==1.21.5
144
+ oauthlib==3.2.0
145
+ onnx==1.11.0
146
+ onnxconverter-common==1.9.0
147
+ opt-einsum==3.3.0
148
+ optax==0.1.0
149
+ optuna==2.10.0
150
+ packaging==21.3
151
+ pandas==1.4.0
152
+ pandocfilters==1.5.0
153
+ parameterized==0.8.1
154
+ parso==0.8.3
155
+ pathspec==0.9.0
156
+ pathtools==0.1.2
157
+ pbr==5.8.1
158
+ pexpect==4.8.0
159
+ phonemizer==3.0.1
160
+ pickleshare==0.7.5
161
+ pillow==9.0.0
162
+ pint==0.16.1
163
+ pip==22.0.2
164
+ pkg-resources==0.0.0
165
+ plac==1.3.5
166
+ platformdirs==2.4.1
167
+ plotly==5.6.0
168
+ pluggy==1.0.0
169
+ pooch==1.6.0
170
+ portalocker==2.0.0
171
+ poyo==0.5.0
172
+ prettytable==3.2.0
173
+ prometheus-client==0.13.1
174
+ promise==2.3
175
+ prompt-toolkit==3.0.26
176
+ protobuf==3.19.4
177
+ psutil==5.9.0
178
+ ptyprocess==0.7.0
179
+ pure-eval==0.2.2
180
+ py-cpuinfo==8.0.0
181
+ py==1.11.0
182
+ pyarrow==6.0.1
183
+ pyasn1-modules==0.2.8
184
+ pyasn1==0.4.8
185
+ pycodestyle==2.8.0
186
+ pycparser==2.21
187
+ pyctcdecode==0.3.0
188
+ pyflakes==2.4.0
189
+ pygments==2.11.2
190
+ pygtrie==2.4.2
191
+ pynvml==11.4.1
192
+ pyopenssl==22.0.0
193
+ pyparsing==3.0.7
194
+ pyperclip==1.8.2
195
+ pypng==0.0.21
196
+ pyrsistent==0.18.1
197
+ pytest-forked==1.4.0
198
+ pytest-timeout==2.1.0
199
+ pytest-xdist==2.5.0
200
+ pytest==7.1.1
201
+ python-dateutil==2.8.2
202
+ python-levenshtein==0.12.2
203
+ python-slugify==6.1.1
204
+ pytz-deprecation-shim==0.1.0.post0
205
+ pytz==2021.3
206
+ pyyaml==5.4.1
207
+ pyzmq==22.3.0
208
+ qtconsole==5.2.2
209
+ qtpy==2.0.1
210
+ ray==1.11.0
211
+ redis==4.2.2
212
+ regex==2022.1.18
213
+ requests-oauthlib==1.3.1
214
+ requests==2.27.1
215
+ resampy==0.2.2
216
+ responses==0.18.0
217
+ rfc3986==2.0.0
218
+ rouge-score==0.0.4
219
+ rsa==4.8
220
+ s3transfer==0.3.7
221
+ sacrebleu==1.5.1
222
+ sacremoses==0.0.47
223
+ scikit-learn==1.0.2
224
+ scipy==1.7.3
225
+ segments==2.2.0
226
+ send2trash==1.8.0
227
+ sentencepiece==0.1.96
228
+ sentry-sdk==1.5.6
229
+ setuptools==44.1.1
230
+ shortuuid==1.0.8
231
+ sigopt==8.3.0
232
+ six==1.16.0
233
+ smmap==5.0.0
234
+ sniffio==1.2.0
235
+ sortedcontainers==2.4.0
236
+ soundfile==0.10.3.post1
237
+ sqlalchemy==1.4.34
238
+ stack-data==0.1.4
239
+ stevedore==3.5.0
240
+ tabulate==0.8.9
241
+ tenacity==8.0.1
242
+ tensorboard-data-server==0.6.1
243
+ tensorboard-plugin-wit==1.8.1
244
+ tensorboard==2.8.0
245
+ tensorboardx==2.5
246
+ tensorflow-io-gcs-filesystem==0.24.0
247
+ tensorflow==2.8.0
248
+ termcolor==1.1.0
249
+ terminado==0.13.1
250
+ testpath==0.5.0
251
+ text-unidecode==1.3
252
+ tf-estimator-nightly==2.8.0.dev2021122109
253
+ tf2onnx==1.9.3
254
+ threadpoolctl==3.1.0
255
+ timeout-decorator==0.5.0
256
+ timm==0.5.4
257
+ tokenizers==0.11.4
258
+ toml==0.10.2
259
+ tomli==2.0.0
260
+ toolz==0.11.2
261
+ torch==1.10.2+cu113
262
+ torchaudio==0.10.2+cu113
263
+ torchvision==0.11.3
264
+ tornado==6.1
265
+ tqdm==4.62.3
266
+ traitlets==5.1.1
267
+ transformers==4.18.0.dev0
268
+ typing-extensions==3.10.0.2
269
+ tzdata==2022.1
270
+ tzlocal==4.2
271
+ unidic-lite==1.0.8
272
+ unidic==1.1.0
273
+ uritemplate==4.1.1
274
+ urllib3==1.26.8
275
+ wandb==0.12.10
276
+ wasabi==0.9.1
277
+ wcwidth==0.2.5
278
+ webencodings==0.5.1
279
+ websocket-client==1.2.3
280
+ werkzeug==2.0.2
281
+ wheel==0.37.1
282
+ widgetsnbextension==3.5.2
283
+ wrapt==1.14.0
284
+ xxhash==2.0.2
285
+ yarl==1.7.2
286
+ yaspin==2.1.0
287
+ zipp==3.7.0
wandb/run-20220505_173213-ykkm1chu/files/wandb-metadata.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-05-05T17:32:17.002298",
5
+ "startedAt": "2022-05-05T17:32:13.328589",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--overwrite_output_dir",
13
+ "--freeze_feature_encoder",
14
+ "--gradient_checkpointing",
15
+ "--predict_with_generate",
16
+ "--fp16",
17
+ "--group_by_length",
18
+ "--do_train",
19
+ "--do_eval",
20
+ "--load_best_model_at_end",
21
+ "--push_to_hub",
22
+ "--use_auth_token",
23
+ "--eval_split_name=test",
24
+ "--eval_steps=500",
25
+ "--evaluation_strategy=steps",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--gradient_accumulation_steps=16",
29
+ "--greater_is_better=True",
30
+ "--hidden_dropout=0.2",
31
+ "--language=fr.en",
32
+ "--learning_rate=0.0003287457929573604",
33
+ "--logging_steps=1",
34
+ "--max_duration_in_seconds=20",
35
+ "--metric_for_best_model=bleu",
36
+ "--model_name_or_path=./",
37
+ "--num_train_epochs=3",
38
+ "--output_dir=./",
39
+ "--per_device_eval_batch_size=8",
40
+ "--per_device_train_batch_size=8",
41
+ "--save_steps=500",
42
+ "--task=covost2",
43
+ "--warmup_steps=500"
44
+ ],
45
+ "state": "running",
46
+ "program": "/home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/run_xtreme_s.py",
47
+ "codePath": "run_xtreme_s.py",
48
+ "git": {
49
+ "remote": "https://huggingface.co/sanchit-gandhi/xtreme_s_xlsr_2_bart_covost2_fr_en",
50
+ "commit": "96386197e130d55a2be359d90ef99c414e1860b3"
51
+ },
52
+ "email": "sanchit@huggingface.co",
53
+ "root": "/home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en",
54
+ "host": "sanchit--v100",
55
+ "username": "sanchit_huggingface_co",
56
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python3"
57
+ }
wandb/run-20220505_173213-ykkm1chu/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
1
+ {"train/train_runtime": 8.4243, "train/train_samples_per_second": 73848.5, "train/train_steps_per_second": 576.902, "train/total_flos": 0.0, "train/train_loss": 0.0, "train/epoch": 3.0, "train/global_step": 9720, "_runtime": 8, "_timestamp": 1651771941, "_step": 0}
wandb/run-20220505_173213-ykkm1chu/logs/debug-internal.log ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-05-05 17:32:14,392 INFO MainThread:73923 [internal.py:wandb_internal():89] W&B internal server running at pid: 73923, started at: 2022-05-05 17:32:14.392410
2
+ 2022-05-05 17:32:14,396 DEBUG HandlerThread:73923 [handler.py:handle_request():131] handle_request: check_version
3
+ 2022-05-05 17:32:14,396 INFO WriterThread:73923 [datastore.py:open_for_write():77] open: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/run-ykkm1chu.wandb
4
+ 2022-05-05 17:32:14,396 DEBUG SenderThread:73923 [sender.py:send():235] send: header
5
+ 2022-05-05 17:32:14,396 DEBUG SenderThread:73923 [sender.py:send_request():249] send_request: check_version
6
+ 2022-05-05 17:32:14,467 DEBUG SenderThread:73923 [sender.py:send():235] send: run
7
+ 2022-05-05 17:32:14,567 INFO SenderThread:73923 [dir_watcher.py:__init__():169] watching files in: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files
8
+ 2022-05-05 17:32:14,568 INFO SenderThread:73923 [sender.py:_start_run_threads():809] run started: ykkm1chu with start time 1651771933
9
+ 2022-05-05 17:32:14,568 DEBUG SenderThread:73923 [sender.py:send():235] send: summary
10
+ 2022-05-05 17:32:14,568 INFO SenderThread:73923 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
11
+ 2022-05-05 17:32:14,569 DEBUG HandlerThread:73923 [handler.py:handle_request():131] handle_request: run_start
12
+ 2022-05-05 17:32:15,569 INFO Thread-8 :73923 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/wandb-summary.json
13
+ 2022-05-05 17:32:17,002 DEBUG HandlerThread:73923 [meta.py:__init__():36] meta init
14
+ 2022-05-05 17:32:17,002 DEBUG HandlerThread:73923 [meta.py:__init__():50] meta init done
15
+ 2022-05-05 17:32:17,002 DEBUG HandlerThread:73923 [meta.py:probe():210] probe
16
+ 2022-05-05 17:32:17,008 DEBUG HandlerThread:73923 [meta.py:_setup_git():200] setup git
17
+ 2022-05-05 17:32:17,043 DEBUG HandlerThread:73923 [meta.py:_setup_git():207] setup git done
18
+ 2022-05-05 17:32:17,043 DEBUG HandlerThread:73923 [meta.py:_save_pip():54] save pip
19
+ 2022-05-05 17:32:17,044 DEBUG HandlerThread:73923 [meta.py:_save_pip():68] save pip done
20
+ 2022-05-05 17:32:17,044 DEBUG HandlerThread:73923 [meta.py:probe():248] probe done
21
+ 2022-05-05 17:32:17,048 DEBUG SenderThread:73923 [sender.py:send():235] send: files
22
+ 2022-05-05 17:32:17,048 INFO SenderThread:73923 [sender.py:_save_file():944] saving file wandb-metadata.json with policy now
23
+ 2022-05-05 17:32:17,055 DEBUG HandlerThread:73923 [handler.py:handle_request():131] handle_request: stop_status
24
+ 2022-05-05 17:32:17,055 DEBUG SenderThread:73923 [sender.py:send_request():249] send_request: stop_status
25
+ 2022-05-05 17:32:17,099 DEBUG SenderThread:73923 [sender.py:send():235] send: config
26
+ 2022-05-05 17:32:17,100 DEBUG SenderThread:73923 [sender.py:send():235] send: metric
27
+ 2022-05-05 17:32:17,100 DEBUG SenderThread:73923 [sender.py:send():235] send: metric
28
+ 2022-05-05 17:32:17,100 WARNING SenderThread:73923 [sender.py:send_metric():902] Seen metric with glob (shouldnt happen)
29
+ 2022-05-05 17:32:17,571 INFO Thread-8 :73923 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/requirements.txt
30
+ 2022-05-05 17:32:17,572 INFO Thread-8 :73923 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/wandb-metadata.json
31
+ 2022-05-05 17:32:17,572 INFO Thread-8 :73923 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/output.log
32
+ 2022-05-05 17:32:17,892 INFO Thread-11 :73923 [upload_job.py:push():137] Uploaded file /tmp/tmpqkr07o8awandb/2ridycox-wandb-metadata.json
33
+ 2022-05-05 17:32:19,572 INFO Thread-8 :73923 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/output.log
34
+ 2022-05-05 17:32:21,452 DEBUG SenderThread:73923 [sender.py:send():235] send: metric
35
+ 2022-05-05 17:32:21,453 DEBUG SenderThread:73923 [sender.py:send():235] send: metric
36
+ 2022-05-05 17:32:21,453 DEBUG SenderThread:73923 [sender.py:send():235] send: metric
37
+ 2022-05-05 17:32:21,453 DEBUG SenderThread:73923 [sender.py:send():235] send: metric
38
+ 2022-05-05 17:32:21,453 DEBUG SenderThread:73923 [sender.py:send():235] send: metric
39
+ 2022-05-05 17:32:21,453 DEBUG SenderThread:73923 [sender.py:send():235] send: metric
40
+ 2022-05-05 17:32:21,453 DEBUG SenderThread:73923 [sender.py:send():235] send: history
41
+ 2022-05-05 17:32:21,453 DEBUG SenderThread:73923 [sender.py:send():235] send: summary
42
+ 2022-05-05 17:32:21,455 INFO SenderThread:73923 [sender.py:_save_file():944] saving file wandb-summary.json with policy end
43
+ 2022-05-05 17:32:21,573 INFO Thread-8 :73923 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/wandb-summary.json
44
+ 2022-05-05 17:32:23,574 INFO Thread-8 :73923 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/output.log
45
+ 2022-05-05 17:32:32,106 DEBUG HandlerThread:73923 [handler.py:handle_request():131] handle_request: stop_status
46
+ 2022-05-05 17:32:32,106 DEBUG SenderThread:73923 [sender.py:send_request():249] send_request: stop_status
47
+ 2022-05-05 17:32:35,578 INFO Thread-8 :73923 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/output.log
48
+ 2022-05-05 17:32:45,443 DEBUG SenderThread:73923 [sender.py:send():235] send: stats
49
+ 2022-05-05 17:32:45,582 INFO Thread-8 :73923 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/config.yaml
50
+ 2022-05-05 17:32:47,138 DEBUG HandlerThread:73923 [handler.py:handle_request():131] handle_request: stop_status
51
+ 2022-05-05 17:32:47,139 DEBUG SenderThread:73923 [sender.py:send_request():249] send_request: stop_status
52
+ 2022-05-05 17:32:55,585 INFO Thread-8 :73923 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/output.log
53
+ 2022-05-05 17:33:02,172 DEBUG HandlerThread:73923 [handler.py:handle_request():131] handle_request: stop_status
54
+ 2022-05-05 17:33:02,172 DEBUG SenderThread:73923 [sender.py:send_request():249] send_request: stop_status
55
+ 2022-05-05 17:33:15,850 DEBUG SenderThread:73923 [sender.py:send():235] send: stats
56
+ 2022-05-05 17:33:17,208 DEBUG HandlerThread:73923 [handler.py:handle_request():131] handle_request: stop_status
57
+ 2022-05-05 17:33:17,209 DEBUG SenderThread:73923 [sender.py:send_request():249] send_request: stop_status
58
+ 2022-05-05 17:33:27,473 WARNING MainThread:73923 [internal.py:is_dead():380] Internal process exiting, parent pid 73744 disappeared
59
+ 2022-05-05 17:33:27,473 ERROR MainThread:73923 [internal.py:wandb_internal():148] Internal process shutdown.
60
+ 2022-05-05 17:33:27,853 INFO WriterThread:73923 [datastore.py:close():281] close: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/run-ykkm1chu.wandb
61
+ 2022-05-05 17:33:28,248 INFO SenderThread:73923 [sender.py:finish():1075] shutting down sender
62
+ 2022-05-05 17:33:28,248 INFO SenderThread:73923 [dir_watcher.py:finish():283] shutting down directory watcher
63
+ 2022-05-05 17:33:28,253 INFO HandlerThread:73923 [handler.py:finish():739] shutting down handler
64
+ 2022-05-05 17:33:28,592 INFO SenderThread:73923 [dir_watcher.py:finish():313] scan: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files
65
+ 2022-05-05 17:33:28,592 INFO SenderThread:73923 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/wandb-metadata.json wandb-metadata.json
66
+ 2022-05-05 17:33:28,592 INFO SenderThread:73923 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/output.log output.log
67
+ 2022-05-05 17:33:28,593 INFO SenderThread:73923 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/wandb-summary.json wandb-summary.json
68
+ 2022-05-05 17:33:28,593 INFO SenderThread:73923 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/requirements.txt requirements.txt
69
+ 2022-05-05 17:33:28,597 INFO SenderThread:73923 [dir_watcher.py:finish():327] scan save: /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/config.yaml config.yaml
70
+ 2022-05-05 17:33:28,603 INFO SenderThread:73923 [file_pusher.py:finish():177] shutting down file pusher
71
+ 2022-05-05 17:33:28,603 INFO SenderThread:73923 [file_pusher.py:join():182] waiting for file pusher
72
+ 2022-05-05 17:33:28,861 INFO Thread-12 :73923 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/output.log
73
+ 2022-05-05 17:33:28,870 INFO Thread-14 :73923 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/requirements.txt
74
+ 2022-05-05 17:33:28,882 INFO Thread-13 :73923 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/wandb-summary.json
75
+ 2022-05-05 17:33:28,884 INFO Thread-15 :73923 [upload_job.py:push():137] Uploaded file /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/files/config.yaml
76
+ 2022-05-05 17:33:29,163 INFO MainThread:73923 [internal.py:handle_exit():79] Internal process exited
wandb/run-20220505_173213-ykkm1chu/logs/debug.log ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-05-05 17:32:13,330 INFO MainThread:73744 [wandb_setup.py:_flush():75] Loading settings from /home/sanchit_huggingface_co/.config/wandb/settings
2
+ 2022-05-05 17:32:13,330 INFO MainThread:73744 [wandb_setup.py:_flush():75] Loading settings from wandb/settings
3
+ 2022-05-05 17:32:13,330 INFO MainThread:73744 [wandb_setup.py:_flush():75] Loading settings from environment variables: {'entity': 'sanchit-gandhi', 'project': 'xtreme_s_xlsr_2_bart_covost2_fr_en', 'sweep_id': 'gtn0dbwf', 'root_dir': '/home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en', 'run_id': 'ykkm1chu', 'sweep_param_path': '/home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/sweep-gtn0dbwf/config-ykkm1chu.yaml'}
4
+ 2022-05-05 17:32:13,330 INFO MainThread:73744 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_xtreme_s.py', 'program': '/home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/run_xtreme_s.py'}
5
+ 2022-05-05 17:32:13,330 INFO MainThread:73744 [wandb_init.py:_log_setup():386] Logging user logs to /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/logs/debug.log
6
+ 2022-05-05 17:32:13,330 INFO MainThread:73744 [wandb_init.py:_log_setup():387] Logging internal logs to /home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/wandb/run-20220505_173213-ykkm1chu/logs/debug-internal.log
7
+ 2022-05-05 17:32:13,330 INFO MainThread:73744 [wandb_init.py:init():420] calling init triggers
8
+ 2022-05-05 17:32:13,331 INFO MainThread:73744 [wandb_init.py:init():425] wandb.init called with sweep_config: {'eval_split_name': 'test', 'eval_steps': 500, 'evaluation_strategy': 'steps', 'generation_max_length': 40, 'generation_num_beams': 1, 'gradient_accumulation_steps': 16, 'greater_is_better': True, 'hidden_dropout': 0.2, 'language': 'fr.en', 'learning_rate': 0.0003287457929573604, 'logging_steps': 1, 'max_duration_in_seconds': 20, 'metric_for_best_model': 'bleu', 'model_name_or_path': './', 'num_train_epochs': 3, 'output_dir': './', 'per_device_eval_batch_size': 8, 'per_device_train_batch_size': 8, 'save_steps': 500, 'task': 'covost2', 'warmup_steps': 500}
9
+ config: {}
10
+ 2022-05-05 17:32:13,331 INFO MainThread:73744 [wandb_init.py:init():471] starting backend
11
+ 2022-05-05 17:32:13,331 INFO MainThread:73744 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-05-05 17:32:13,410 INFO MainThread:73744 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-05-05 17:32:13,489 INFO MainThread:73744 [backend.py:ensure_launched():224] started backend process with pid: 73923
14
+ 2022-05-05 17:32:13,492 INFO MainThread:73744 [wandb_init.py:init():480] backend started and connected
15
+ 2022-05-05 17:32:13,495 INFO MainThread:73744 [wandb_run.py:_config_callback():966] config_cb None None {'eval_split_name': 'test', 'eval_steps': 500, 'evaluation_strategy': 'steps', 'generation_max_length': 40, 'generation_num_beams': 1, 'gradient_accumulation_steps': 16, 'greater_is_better': True, 'hidden_dropout': 0.2, 'language': 'fr.en', 'learning_rate': 0.0003287457929573604, 'logging_steps': 1, 'max_duration_in_seconds': 20, 'metric_for_best_model': 'bleu', 'model_name_or_path': './', 'num_train_epochs': 3, 'output_dir': './', 'per_device_eval_batch_size': 8, 'per_device_train_batch_size': 8, 'save_steps': 500, 'task': 'covost2', 'warmup_steps': 500}
16
+ 2022-05-05 17:32:13,509 INFO MainThread:73744 [wandb_init.py:init():550] updated telemetry
17
+ 2022-05-05 17:32:13,700 INFO MainThread:73744 [wandb_init.py:init():581] communicating current version
18
+ 2022-05-05 17:32:14,465 INFO MainThread:73744 [wandb_init.py:init():586] got version response upgrade_message: "wandb version 0.12.16 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
19
+
20
+ 2022-05-05 17:32:14,466 INFO MainThread:73744 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
21
+ 2022-05-05 17:32:14,568 INFO MainThread:73744 [wandb_init.py:init():624] starting run threads in backend
22
+ 2022-05-05 17:32:17,055 INFO MainThread:73744 [wandb_run.py:_console_start():1827] atexit reg
23
+ 2022-05-05 17:32:17,055 INFO MainThread:73744 [wandb_run.py:_redirect():1701] redirect: SettingsConsole.REDIRECT
24
+ 2022-05-05 17:32:17,056 INFO MainThread:73744 [wandb_run.py:_redirect():1706] Redirecting console.
25
+ 2022-05-05 17:32:17,057 INFO MainThread:73744 [wandb_run.py:_redirect():1762] Redirects installed.
26
+ 2022-05-05 17:32:17,057 INFO MainThread:73744 [wandb_init.py:init():651] run started, returning control to user process
27
+ 2022-05-05 17:32:17,060 INFO MainThread:73744 [wandb_run.py:_config_callback():966] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'torch.float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 40, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['SpeechEncoderDecoderModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': './', 'transformers_version': None, 'decoder': {'vocab_size': 50265, 'max_position_embeddings': 1024, 'd_model': 1024, 'encoder_ffn_dim': 4096, 'encoder_layers': 12, 'encoder_attention_heads': 16, 'decoder_ffn_dim': 4096, 'decoder_layers': 12, 'decoder_attention_heads': 16, 'dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.1, 'activation_function': 'gelu', 'init_std': 0.02, 'encoder_layerdrop': 0.0, 'decoder_layerdrop': 0.0, 'classifier_dropout': 0.0, 'use_cache': True, 'num_hidden_layers': 12, 'scale_embedding': False, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': True, 'cross_attention_hidden_size': None, 'add_cross_attention': True, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': True, 'num_beams': 4, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 3, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': 0, 'forced_eos_token_id': 2, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['BartModel'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1', 2: 'LABEL_2'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': 2, 'task_specific_params': {'summarization': {'length_penalty': 1.0, 'max_length': 128, 'min_length': 12, 'num_beams': 4}, 'summarization_cnn': {'length_penalty': 2.0, 'max_length': 142, 'min_length': 56, 'num_beams': 4}, 'summarization_xsum': {'length_penalty': 1.0, 'max_length': 62, 'min_length': 11, 'num_beams': 6}}, 'problem_type': None, '_name_or_path': 'facebook/bart-large', 'transformers_version': '4.19.0.dev0', 'add_bias_logits': False, 'add_final_layer_norm': False, 'classif_dropout': 0.1, 'gradient_checkpointing': False, 'normalize_before': False, 'model_type': 'bart'}, 'encoder': {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.19.0.dev0', 'feat_extract_dropout': 0.0, 'gradient_checkpointing': False, 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.2, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.1, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': True, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'model_type': 'wav2vec2'}, 'model_type': 'speech-encoder-decoder', 'processor_class': 'Wav2Vec2Processor', 'use_cache': False, 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'prediction_loss_only': False, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'eval_accumulation_steps': 'None', 'eval_delay': 0, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/May05_17-31-19_sanchit--v100', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_total_limit': 'None', 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'data_seed': 'None', 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': './', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': "['tensorboard', 'wandb', 'codecarbon']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'gradient_checkpointing': True, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'sortish_sampler': False, 'predict_with_generate': True, 'train_batch_size': 8, 'eval_batch_size': 8}
28
+ 2022-05-05 17:32:17,063 INFO MainThread:73744 [wandb_watch.py:watch():43] Watching
wandb/run-20220505_173213-ykkm1chu/run-ykkm1chu.wandb ADDED
Binary file (14.9 kB). View file
wandb/run-20220505_173748-b097rk18/files/config.yaml ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220505_173748-b097rk18/files/output.log ADDED
The diff for this file is too large to render. See raw diff
wandb/run-20220505_173748-b097rk18/files/requirements.txt ADDED
@@ -0,0 +1,287 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ alembic==1.7.7
5
+ anyio==3.5.0
6
+ appdirs==1.4.4
7
+ apscheduler==3.9.1
8
+ argon2-cffi-bindings==21.2.0
9
+ argon2-cffi==21.3.0
10
+ arrow==1.2.2
11
+ asttokens==2.0.5
12
+ astunparse==1.6.3
13
+ async-timeout==4.0.2
14
+ attrs==21.4.0
15
+ audioread==2.1.9
16
+ autopage==0.5.0
17
+ babel==2.9.1
18
+ backcall==0.2.0
19
+ backoff==1.11.1
20
+ binaryornot==0.4.4
21
+ bitsandbytes-cuda113==0.26.0
22
+ black==22.1.0
23
+ bleach==4.1.0
24
+ boto3==1.16.34
25
+ botocore==1.19.63
26
+ brotli==1.0.9
27
+ cachetools==5.0.0
28
+ certifi==2021.10.8
29
+ cffi==1.15.0
30
+ chardet==4.0.0
31
+ charset-normalizer==2.0.11
32
+ chex==0.1.0
33
+ click==8.0.3
34
+ cliff==3.10.1
35
+ clldutils==3.10.1
36
+ cmaes==0.8.2
37
+ cmd2==2.4.0
38
+ codecarbon==1.2.0
39
+ colorlog==6.6.0
40
+ cookiecutter==1.7.3
41
+ cryptography==36.0.2
42
+ csvw==1.11.0
43
+ cycler==0.11.0
44
+ dash-bootstrap-components==1.1.0
45
+ dash-core-components==2.0.0
46
+ dash-html-components==2.0.0
47
+ dash-table==5.0.0
48
+ dash==2.3.1
49
+ datasets==2.1.1.dev0
50
+ debugpy==1.5.1
51
+ decorator==5.1.1
52
+ defusedxml==0.7.1
53
+ deprecated==1.2.13
54
+ dill==0.3.4
55
+ dlinfo==1.2.1
56
+ dm-tree==0.1.6
57
+ docker-pycreds==0.4.0
58
+ docker==4.4.4
59
+ entrypoints==0.4
60
+ execnet==1.9.0
61
+ executing==0.8.2
62
+ faiss-cpu==1.7.2
63
+ filelock==3.4.2
64
+ fire==0.4.0
65
+ flake8==4.0.1
66
+ flask-compress==1.11
67
+ flask==2.1.1
68
+ flatbuffers==1.12
69
+ flax==0.4.0
70
+ fonttools==4.29.1
71
+ frozenlist==1.3.0
72
+ fsspec==2022.1.0
73
+ fugashi==1.1.2
74
+ gast==0.5.3
75
+ gitdb==4.0.9
76
+ gitpython==3.1.18
77
+ google-auth-oauthlib==0.4.6
78
+ google-auth==2.6.0
79
+ google-pasta==0.2.0
80
+ greenlet==1.1.2
81
+ grpcio==1.43.0
82
+ h5py==3.6.0
83
+ hf-doc-builder==0.2.0
84
+ huggingface-hub==0.4.0
85
+ hypothesis==6.36.1
86
+ idna==3.3
87
+ importlib-metadata==4.10.1
88
+ iniconfig==1.1.1
89
+ ipadic==1.0.0
90
+ ipdb==0.13.9
91
+ ipykernel==6.8.0
92
+ ipython-genutils==0.2.0
93
+ ipython==8.0.1
94
+ ipywidgets==7.6.5
95
+ isodate==0.6.1
96
+ isort==5.10.1
97
+ itsdangerous==2.1.2
98
+ jax==0.2.28
99
+ jaxlib==0.1.76+cuda11.cudnn82
100
+ jedi==0.18.1
101
+ jinja2-time==0.2.0
102
+ jinja2==3.0.3
103
+ jiwer==2.3.0
104
+ jmespath==0.10.0
105
+ joblib==1.1.0
106
+ json5==0.9.6
107
+ jsonschema==4.4.0
108
+ jupyter-client==7.1.2
109
+ jupyter-console==6.4.0
110
+ jupyter-core==4.9.1
111
+ jupyter-server==1.13.5
112
+ jupyter==1.0.0
113
+ jupyterlab-pygments==0.1.2
114
+ jupyterlab-server==2.10.3
115
+ jupyterlab-widgets==1.0.2
116
+ jupyterlab==3.2.9
117
+ keras-preprocessing==1.1.2
118
+ keras==2.8.0
119
+ kiwisolver==1.3.2
120
+ kubernetes==12.0.1
121
+ libclang==13.0.0
122
+ librosa==0.8.1
123
+ llvmlite==0.38.0
124
+ mako==1.2.0
125
+ markdown==3.3.6
126
+ markupsafe==2.0.1
127
+ matplotlib-inline==0.1.3
128
+ matplotlib==3.5.1
129
+ mccabe==0.6.1
130
+ mistune==0.8.4
131
+ msgpack==1.0.3
132
+ multidict==6.0.2
133
+ multiprocess==0.70.12.2
134
+ mypy-extensions==0.4.3
135
+ nbclassic==0.3.5
136
+ nbclient==0.5.10
137
+ nbconvert==6.4.1
138
+ nbformat==5.1.3
139
+ nest-asyncio==1.5.4
140
+ nltk==3.7
141
+ notebook==6.4.8
142
+ numba==0.55.1
143
+ numpy==1.21.5
144
+ oauthlib==3.2.0
145
+ onnx==1.11.0
146
+ onnxconverter-common==1.9.0
147
+ opt-einsum==3.3.0
148
+ optax==0.1.0
149
+ optuna==2.10.0
150
+ packaging==21.3
151
+ pandas==1.4.0
152
+ pandocfilters==1.5.0
153
+ parameterized==0.8.1
154
+ parso==0.8.3
155
+ pathspec==0.9.0
156
+ pathtools==0.1.2
157
+ pbr==5.8.1
158
+ pexpect==4.8.0
159
+ phonemizer==3.0.1
160
+ pickleshare==0.7.5
161
+ pillow==9.0.0
162
+ pint==0.16.1
163
+ pip==22.0.2
164
+ pkg-resources==0.0.0
165
+ plac==1.3.5
166
+ platformdirs==2.4.1
167
+ plotly==5.6.0
168
+ pluggy==1.0.0
169
+ pooch==1.6.0
170
+ portalocker==2.0.0
171
+ poyo==0.5.0
172
+ prettytable==3.2.0
173
+ prometheus-client==0.13.1
174
+ promise==2.3
175
+ prompt-toolkit==3.0.26
176
+ protobuf==3.19.4
177
+ psutil==5.9.0
178
+ ptyprocess==0.7.0
179
+ pure-eval==0.2.2
180
+ py-cpuinfo==8.0.0
181
+ py==1.11.0
182
+ pyarrow==6.0.1
183
+ pyasn1-modules==0.2.8
184
+ pyasn1==0.4.8
185
+ pycodestyle==2.8.0
186
+ pycparser==2.21
187
+ pyctcdecode==0.3.0
188
+ pyflakes==2.4.0
189
+ pygments==2.11.2
190
+ pygtrie==2.4.2
191
+ pynvml==11.4.1
192
+ pyopenssl==22.0.0
193
+ pyparsing==3.0.7
194
+ pyperclip==1.8.2
195
+ pypng==0.0.21
196
+ pyrsistent==0.18.1
197
+ pytest-forked==1.4.0
198
+ pytest-timeout==2.1.0
199
+ pytest-xdist==2.5.0
200
+ pytest==7.1.1
201
+ python-dateutil==2.8.2
202
+ python-levenshtein==0.12.2
203
+ python-slugify==6.1.1
204
+ pytz-deprecation-shim==0.1.0.post0
205
+ pytz==2021.3
206
+ pyyaml==5.4.1
207
+ pyzmq==22.3.0
208
+ qtconsole==5.2.2
209
+ qtpy==2.0.1
210
+ ray==1.11.0
211
+ redis==4.2.2
212
+ regex==2022.1.18
213
+ requests-oauthlib==1.3.1
214
+ requests==2.27.1
215
+ resampy==0.2.2
216
+ responses==0.18.0
217
+ rfc3986==2.0.0
218
+ rouge-score==0.0.4
219
+ rsa==4.8
220
+ s3transfer==0.3.7
221
+ sacrebleu==1.5.1
222
+ sacremoses==0.0.47
223
+ scikit-learn==1.0.2
224
+ scipy==1.7.3
225
+ segments==2.2.0
226
+ send2trash==1.8.0
227
+ sentencepiece==0.1.96
228
+ sentry-sdk==1.5.6
229
+ setuptools==44.1.1
230
+ shortuuid==1.0.8
231
+ sigopt==8.3.0
232
+ six==1.16.0
233
+ smmap==5.0.0
234
+ sniffio==1.2.0
235
+ sortedcontainers==2.4.0
236
+ soundfile==0.10.3.post1
237
+ sqlalchemy==1.4.34
238
+ stack-data==0.1.4
239
+ stevedore==3.5.0
240
+ tabulate==0.8.9
241
+ tenacity==8.0.1
242
+ tensorboard-data-server==0.6.1
243
+ tensorboard-plugin-wit==1.8.1
244
+ tensorboard==2.8.0
245
+ tensorboardx==2.5
246
+ tensorflow-io-gcs-filesystem==0.24.0
247
+ tensorflow==2.8.0
248
+ termcolor==1.1.0
249
+ terminado==0.13.1
250
+ testpath==0.5.0
251
+ text-unidecode==1.3
252
+ tf-estimator-nightly==2.8.0.dev2021122109
253
+ tf2onnx==1.9.3
254
+ threadpoolctl==3.1.0
255
+ timeout-decorator==0.5.0
256
+ timm==0.5.4
257
+ tokenizers==0.11.4
258
+ toml==0.10.2
259
+ tomli==2.0.0
260
+ toolz==0.11.2
261
+ torch==1.10.2+cu113
262
+ torchaudio==0.10.2+cu113
263
+ torchvision==0.11.3
264
+ tornado==6.1
265
+ tqdm==4.62.3
266
+ traitlets==5.1.1
267
+ transformers==4.18.0.dev0
268
+ typing-extensions==3.10.0.2
269
+ tzdata==2022.1
270
+ tzlocal==4.2
271
+ unidic-lite==1.0.8
272
+ unidic==1.1.0
273
+ uritemplate==4.1.1
274
+ urllib3==1.26.8
275
+ wandb==0.12.10
276
+ wasabi==0.9.1
277
+ wcwidth==0.2.5
278
+ webencodings==0.5.1
279
+ websocket-client==1.2.3
280
+ werkzeug==2.0.2
281
+ wheel==0.37.1
282
+ widgetsnbextension==3.5.2
283
+ wrapt==1.14.0
284
+ xxhash==2.0.2
285
+ yarl==1.7.2
286
+ yaspin==2.1.0
287
+ zipp==3.7.0
wandb/run-20220505_173748-b097rk18/files/wandb-metadata.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.11.0-1028-gcp-x86_64-with-glibc2.33",
3
+ "python": "3.9.5",
4
+ "heartbeatAt": "2022-05-05T17:37:52.275326",
5
+ "startedAt": "2022-05-05T17:37:48.741791",
6
+ "docker": null,
7
+ "gpu": "Tesla V100-SXM2-16GB",
8
+ "gpu_count": 2,
9
+ "cpu_count": 16,
10
+ "cuda": null,
11
+ "args": [
12
+ "--overwrite_output_dir",
13
+ "--freeze_feature_encoder",
14
+ "--gradient_checkpointing",
15
+ "--predict_with_generate",
16
+ "--fp16",
17
+ "--group_by_length",
18
+ "--do_train",
19
+ "--do_eval",
20
+ "--load_best_model_at_end",
21
+ "--push_to_hub",
22
+ "--use_auth_token",
23
+ "--eval_split_name=test",
24
+ "--eval_steps=500",
25
+ "--evaluation_strategy=steps",
26
+ "--generation_max_length=40",
27
+ "--generation_num_beams=1",
28
+ "--gradient_accumulation_steps=16",
29
+ "--greater_is_better=True",
30
+ "--hidden_dropout=0.2",
31
+ "--language=fr.en",
32
+ "--learning_rate=3e-05",
33
+ "--logging_steps=1",
34
+ "--max_duration_in_seconds=20",
35
+ "--metric_for_best_model=bleu",
36
+ "--model_name_or_path=./",
37
+ "--num_train_epochs=3",
38
+ "--output_dir=./",
39
+ "--per_device_eval_batch_size=8",
40
+ "--per_device_train_batch_size=8",
41
+ "--save_steps=500",
42
+ "--task=covost2",
43
+ "--warmup_steps=500"
44
+ ],
45
+ "state": "running",
46
+ "program": "/home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en/run_xtreme_s.py",
47
+ "codePath": "run_xtreme_s.py",
48
+ "git": {
49
+ "remote": "https://huggingface.co/sanchit-gandhi/xtreme_s_xlsr_2_bart_covost2_fr_en",
50
+ "commit": "96386197e130d55a2be359d90ef99c414e1860b3"
51
+ },
52
+ "email": "sanchit@huggingface.co",
53
+ "root": "/home/sanchit_huggingface_co/xtreme_s_xlsr_2_bart_covost2_fr_en",
54
+ "host": "sanchit--v100",
55
+ "username": "sanchit_huggingface_co",
56
+ "executable": "/home/sanchit_huggingface_co/gcp/bin/python3"
57
+ }
wandb/run-20220505_173748-b097rk18/files/wandb-summary.json ADDED
The diff for this file is too large to render. See raw diff