sanchit-gandhi HF staff commited on
Commit
cac475f
1 Parent(s): c09faa6

Saving weights and logs of epoch 0

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. .ipynb_checkpoints/create_model-checkpoint.py +29 -0
  3. config.json +277 -0
  4. events.out.tfevents.1649095813.t1v-n-4eb331dd-w-0.1239999.0.v2 +3 -0
  5. events.out.tfevents.1649095849.t1v-n-4eb331dd-w-0.1241191.0.v2 +3 -0
  6. events.out.tfevents.1649095884.t1v-n-4eb331dd-w-0.1242310.0.v2 +3 -0
  7. events.out.tfevents.1649095921.t1v-n-4eb331dd-w-0.1243418.0.v2 +3 -0
  8. events.out.tfevents.1649095961.t1v-n-4eb331dd-w-0.1244535.0.v2 +3 -0
  9. events.out.tfevents.1649192648.t1v-n-4eb331dd-w-0.1449582.0.v2 +3 -0
  10. flax_model.msgpack +3 -0
  11. merges.txt +0 -0
  12. preprocessor_config.json +9 -0
  13. run_librispeech.sh +4 -4
  14. special_tokens_map.json +1 -0
  15. sweep.yaml +65 -0
  16. tokenizer.json +0 -0
  17. tokenizer_config.json +1 -0
  18. vocab.json +0 -0
  19. wandb/debug-cli.log +147 -0
  20. wandb/debug-internal.log +1 -0
  21. wandb/debug.log +1 -0
  22. wandb/latest-run +1 -0
  23. wandb/run-20220404_180947-p4wqexfj/files/config.yaml +108 -0
  24. wandb/run-20220404_180947-p4wqexfj/files/output.log +788 -0
  25. wandb/run-20220404_180947-p4wqexfj/files/requirements.txt +177 -0
  26. wandb/run-20220404_180947-p4wqexfj/files/wandb-metadata.json +53 -0
  27. wandb/run-20220404_180947-p4wqexfj/files/wandb-summary.json +1 -0
  28. wandb/run-20220404_180947-p4wqexfj/logs/debug-internal.log +148 -0
  29. wandb/run-20220404_180947-p4wqexfj/logs/debug.log +115 -0
  30. wandb/run-20220404_180947-p4wqexfj/run-p4wqexfj.wandb +0 -0
  31. wandb/run-20220404_181027-mgg9caus/files/config.yaml +108 -0
  32. wandb/run-20220404_181027-mgg9caus/files/output.log +793 -0
  33. wandb/run-20220404_181027-mgg9caus/files/requirements.txt +177 -0
  34. wandb/run-20220404_181027-mgg9caus/files/wandb-metadata.json +53 -0
  35. wandb/run-20220404_181027-mgg9caus/files/wandb-summary.json +1 -0
  36. wandb/run-20220404_181027-mgg9caus/logs/debug-internal.log +144 -0
  37. wandb/run-20220404_181027-mgg9caus/logs/debug.log +107 -0
  38. wandb/run-20220404_181027-mgg9caus/run-mgg9caus.wandb +0 -0
  39. wandb/run-20220404_181103-88xgr1fg/files/config.yaml +108 -0
  40. wandb/run-20220404_181103-88xgr1fg/files/output.log +793 -0
  41. wandb/run-20220404_181103-88xgr1fg/files/requirements.txt +177 -0
  42. wandb/run-20220404_181103-88xgr1fg/files/wandb-metadata.json +53 -0
  43. wandb/run-20220404_181103-88xgr1fg/files/wandb-summary.json +1 -0
  44. wandb/run-20220404_181103-88xgr1fg/logs/debug-internal.log +144 -0
  45. wandb/run-20220404_181103-88xgr1fg/logs/debug.log +107 -0
  46. wandb/run-20220404_181103-88xgr1fg/run-88xgr1fg.wandb +0 -0
  47. wandb/run-20220404_181139-xmgtui21/files/config.yaml +108 -0
  48. wandb/run-20220404_181139-xmgtui21/files/output.log +793 -0
  49. wandb/run-20220404_181139-xmgtui21/files/requirements.txt +177 -0
  50. wandb/run-20220404_181139-xmgtui21/files/wandb-metadata.json +53 -0
.gitattributes CHANGED
@@ -26,3 +26,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
26
  *.zip filter=lfs diff=lfs merge=lfs -text
27
  *.zstandard filter=lfs diff=lfs merge=lfs -text
28
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
26
  *.zip filter=lfs diff=lfs merge=lfs -text
27
  *.zstandard filter=lfs diff=lfs merge=lfs -text
28
  *tfevents* filter=lfs diff=lfs merge=lfs -text
29
+ ~/cache/huggingface/datasets/downloads/tmpqvtsgw_8 filter=lfs diff=lfs merge=lfs -text
.ipynb_checkpoints/create_model-checkpoint.py ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import jax.numpy as jnp
2
+ from transformers import AutoFeatureExtractor, AutoTokenizer, FlaxSpeechEncoderDecoderModel
3
+
4
+ encoder_id = "facebook/wav2vec2-large-lv60"
5
+ decoder_id = "facebook/bart-large-cnn"
6
+
7
+ model = FlaxSpeechEncoderDecoderModel.from_encoder_decoder_pretrained(encoder_id, decoder_id, encoder_add_adapter=True, decoder_from_pt=True)
8
+
9
+ model.config.encoder.feat_proj_dropout = 0.0
10
+ model.config.encoder.final_dropout = 0.0
11
+ model.config.encoder.mask_time_prob = 0.1
12
+ model.config.decoder_start_token_id = model.config.decoder.bos_token_id
13
+ model.config.pad_token_id = model.config.decoder.pad_token_id
14
+ model.config.eos_token_id = model.config.decoder.eos_token_id
15
+ model.config.max_length = 40
16
+ model.config.num_beams = 1
17
+ model.config.encoder.layerdrop = 0.0
18
+ model.config.use_cache = False
19
+ model.config.processor_class = "Wav2Vec2Processor"
20
+
21
+ # check if generation works
22
+ out = model.generate(jnp.ones((1, 2000)))
23
+
24
+ model.save_pretrained("./")
25
+
26
+ feature_extractor = AutoFeatureExtractor.from_pretrained(encoder_id)
27
+ feature_extractor.save_pretrained("./")
28
+ tokenizer = AutoTokenizer.from_pretrained(decoder_id)
29
+ tokenizer.save_pretrained("./")
config.json ADDED
@@ -0,0 +1,277 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./",
3
+ "architectures": [
4
+ "SpeechEncoderDecoderModel"
5
+ ],
6
+ "decoder": {
7
+ "_name_or_path": "facebook/bart-large-cnn",
8
+ "_num_labels": 3,
9
+ "activation_dropout": 0.0,
10
+ "activation_function": "gelu",
11
+ "add_cross_attention": true,
12
+ "add_final_layer_norm": false,
13
+ "architectures": [
14
+ "BartForConditionalGeneration"
15
+ ],
16
+ "attention_dropout": 0.0,
17
+ "bad_words_ids": null,
18
+ "bos_token_id": 0,
19
+ "chunk_size_feed_forward": 0,
20
+ "classif_dropout": 0.0,
21
+ "classifier_dropout": 0.0,
22
+ "cross_attention_hidden_size": null,
23
+ "d_model": 1024,
24
+ "decoder_attention_heads": 16,
25
+ "decoder_ffn_dim": 4096,
26
+ "decoder_layerdrop": 0.0,
27
+ "decoder_layers": 12,
28
+ "decoder_start_token_id": 2,
29
+ "diversity_penalty": 0.0,
30
+ "do_sample": false,
31
+ "dropout": 0.1,
32
+ "early_stopping": true,
33
+ "encoder_attention_heads": 16,
34
+ "encoder_ffn_dim": 4096,
35
+ "encoder_layerdrop": 0.0,
36
+ "encoder_layers": 12,
37
+ "encoder_no_repeat_ngram_size": 0,
38
+ "eos_token_id": 2,
39
+ "exponential_decay_length_penalty": null,
40
+ "finetuning_task": null,
41
+ "force_bos_token_to_be_generated": true,
42
+ "forced_bos_token_id": 0,
43
+ "forced_eos_token_id": 2,
44
+ "gradient_checkpointing": false,
45
+ "id2label": {
46
+ "0": "LABEL_0",
47
+ "1": "LABEL_1",
48
+ "2": "LABEL_2"
49
+ },
50
+ "init_std": 0.02,
51
+ "is_decoder": true,
52
+ "is_encoder_decoder": false,
53
+ "label2id": {
54
+ "LABEL_0": 0,
55
+ "LABEL_1": 1,
56
+ "LABEL_2": 2
57
+ },
58
+ "length_penalty": 2.0,
59
+ "max_length": 142,
60
+ "max_position_embeddings": 1024,
61
+ "min_length": 56,
62
+ "model_type": "bart",
63
+ "no_repeat_ngram_size": 3,
64
+ "normalize_before": false,
65
+ "num_beam_groups": 1,
66
+ "num_beams": 4,
67
+ "num_hidden_layers": 12,
68
+ "num_return_sequences": 1,
69
+ "output_attentions": false,
70
+ "output_hidden_states": false,
71
+ "output_past": true,
72
+ "output_scores": false,
73
+ "pad_token_id": 1,
74
+ "prefix": " ",
75
+ "problem_type": null,
76
+ "pruned_heads": {},
77
+ "remove_invalid_values": false,
78
+ "repetition_penalty": 1.0,
79
+ "return_dict": true,
80
+ "return_dict_in_generate": false,
81
+ "scale_embedding": false,
82
+ "sep_token_id": null,
83
+ "task_specific_params": {
84
+ "summarization": {
85
+ "early_stopping": true,
86
+ "length_penalty": 2.0,
87
+ "max_length": 142,
88
+ "min_length": 56,
89
+ "no_repeat_ngram_size": 3,
90
+ "num_beams": 4
91
+ }
92
+ },
93
+ "temperature": 1.0,
94
+ "tie_encoder_decoder": false,
95
+ "tie_word_embeddings": true,
96
+ "tokenizer_class": null,
97
+ "top_k": 50,
98
+ "top_p": 1.0,
99
+ "torch_dtype": null,
100
+ "torchscript": false,
101
+ "transformers_version": "4.18.0.dev0",
102
+ "typical_p": 1.0,
103
+ "use_bfloat16": false,
104
+ "use_cache": true,
105
+ "vocab_size": 50264
106
+ },
107
+ "decoder_start_token_id": 0,
108
+ "encoder": {
109
+ "_name_or_path": "facebook/wav2vec2-large-lv60",
110
+ "activation_dropout": 0.1,
111
+ "adapter_kernel_size": 3,
112
+ "adapter_stride": 2,
113
+ "add_adapter": true,
114
+ "add_cross_attention": false,
115
+ "apply_spec_augment": true,
116
+ "architectures": [
117
+ "Wav2Vec2ForPreTraining"
118
+ ],
119
+ "attention_dropout": 0.1,
120
+ "bad_words_ids": null,
121
+ "bos_token_id": 1,
122
+ "chunk_size_feed_forward": 0,
123
+ "classifier_proj_size": 256,
124
+ "codevector_dim": 768,
125
+ "contrastive_logits_temperature": 0.1,
126
+ "conv_bias": true,
127
+ "conv_dim": [
128
+ 512,
129
+ 512,
130
+ 512,
131
+ 512,
132
+ 512,
133
+ 512,
134
+ 512
135
+ ],
136
+ "conv_kernel": [
137
+ 10,
138
+ 3,
139
+ 3,
140
+ 3,
141
+ 3,
142
+ 2,
143
+ 2
144
+ ],
145
+ "conv_stride": [
146
+ 5,
147
+ 2,
148
+ 2,
149
+ 2,
150
+ 2,
151
+ 2,
152
+ 2
153
+ ],
154
+ "cross_attention_hidden_size": null,
155
+ "ctc_loss_reduction": "sum",
156
+ "ctc_zero_infinity": false,
157
+ "decoder_start_token_id": null,
158
+ "diversity_loss_weight": 0.1,
159
+ "diversity_penalty": 0.0,
160
+ "do_sample": false,
161
+ "do_stable_layer_norm": true,
162
+ "early_stopping": false,
163
+ "encoder_no_repeat_ngram_size": 0,
164
+ "eos_token_id": 2,
165
+ "exponential_decay_length_penalty": null,
166
+ "feat_extract_activation": "gelu",
167
+ "feat_extract_dropout": 0.0,
168
+ "feat_extract_norm": "layer",
169
+ "feat_proj_dropout": 0.0,
170
+ "feat_quantizer_dropout": 0.0,
171
+ "final_dropout": 0.0,
172
+ "finetuning_task": null,
173
+ "forced_bos_token_id": null,
174
+ "forced_eos_token_id": null,
175
+ "gradient_checkpointing": false,
176
+ "hidden_act": "gelu",
177
+ "hidden_dropout": 0.1,
178
+ "hidden_dropout_prob": 0.1,
179
+ "hidden_size": 1024,
180
+ "id2label": {
181
+ "0": "LABEL_0",
182
+ "1": "LABEL_1"
183
+ },
184
+ "initializer_range": 0.02,
185
+ "intermediate_size": 4096,
186
+ "is_decoder": false,
187
+ "is_encoder_decoder": false,
188
+ "label2id": {
189
+ "LABEL_0": 0,
190
+ "LABEL_1": 1
191
+ },
192
+ "layer_norm_eps": 1e-05,
193
+ "layerdrop": 0.0,
194
+ "length_penalty": 1.0,
195
+ "mask_feature_length": 10,
196
+ "mask_feature_min_masks": 0,
197
+ "mask_feature_prob": 0.0,
198
+ "mask_time_length": 10,
199
+ "mask_time_min_masks": 2,
200
+ "mask_time_prob": 0.1,
201
+ "max_length": 20,
202
+ "min_length": 0,
203
+ "model_type": "wav2vec2",
204
+ "no_repeat_ngram_size": 0,
205
+ "num_adapter_layers": 3,
206
+ "num_attention_heads": 16,
207
+ "num_beam_groups": 1,
208
+ "num_beams": 1,
209
+ "num_codevector_groups": 2,
210
+ "num_codevectors_per_group": 320,
211
+ "num_conv_pos_embedding_groups": 16,
212
+ "num_conv_pos_embeddings": 128,
213
+ "num_feat_extract_layers": 7,
214
+ "num_hidden_layers": 24,
215
+ "num_negatives": 100,
216
+ "num_return_sequences": 1,
217
+ "output_attentions": false,
218
+ "output_hidden_size": 1024,
219
+ "output_hidden_states": false,
220
+ "output_scores": false,
221
+ "pad_token_id": 0,
222
+ "prefix": null,
223
+ "problem_type": null,
224
+ "proj_codevector_dim": 768,
225
+ "pruned_heads": {},
226
+ "remove_invalid_values": false,
227
+ "repetition_penalty": 1.0,
228
+ "return_dict": true,
229
+ "return_dict_in_generate": false,
230
+ "sep_token_id": null,
231
+ "task_specific_params": null,
232
+ "tdnn_dilation": [
233
+ 1,
234
+ 2,
235
+ 3,
236
+ 1,
237
+ 1
238
+ ],
239
+ "tdnn_dim": [
240
+ 512,
241
+ 512,
242
+ 512,
243
+ 512,
244
+ 1500
245
+ ],
246
+ "tdnn_kernel": [
247
+ 5,
248
+ 3,
249
+ 3,
250
+ 1,
251
+ 1
252
+ ],
253
+ "temperature": 1.0,
254
+ "tie_encoder_decoder": false,
255
+ "tie_word_embeddings": true,
256
+ "tokenizer_class": null,
257
+ "top_k": 50,
258
+ "top_p": 1.0,
259
+ "torch_dtype": null,
260
+ "torchscript": false,
261
+ "transformers_version": "4.18.0.dev0",
262
+ "typical_p": 1.0,
263
+ "use_bfloat16": false,
264
+ "use_weighted_layer_sum": false,
265
+ "vocab_size": 32,
266
+ "xvector_output_dim": 512
267
+ },
268
+ "eos_token_id": 2,
269
+ "is_encoder_decoder": true,
270
+ "max_length": 40,
271
+ "model_type": "speech-encoder-decoder",
272
+ "pad_token_id": 1,
273
+ "processor_class": "Wav2Vec2Processor",
274
+ "tie_word_embeddings": false,
275
+ "transformers_version": null,
276
+ "use_cache": false
277
+ }
events.out.tfevents.1649095813.t1v-n-4eb331dd-w-0.1239999.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cb555e32fca1e9475a0b241e3bb2cba48e3ce342bbacf6f8ffd3d81c1c3f3fe
3
+ size 40
events.out.tfevents.1649095849.t1v-n-4eb331dd-w-0.1241191.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1df2832c39a07dd1829a8818a21b3333c3c594097c21e9ed8ee9b23ad84aebe7
3
+ size 40
events.out.tfevents.1649095884.t1v-n-4eb331dd-w-0.1242310.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c601c7f5d3f941c32a4b6039cbc59b4443019c46912052b5de5410ae93ad30f7
3
+ size 40
events.out.tfevents.1649095921.t1v-n-4eb331dd-w-0.1243418.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fad3447b5ef3b5828834607b23370c874d292819b47deb7ae58f66c640a36cff
3
+ size 40
events.out.tfevents.1649095961.t1v-n-4eb331dd-w-0.1244535.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:419fe92149e756478572afcd546d07e25adaf99d129c6da7ef64cc394c96cb6a
3
+ size 40
events.out.tfevents.1649192648.t1v-n-4eb331dd-w-0.1449582.0.v2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:549a27283e51c8b5d32b4dfb3b95c7c6dfb6024b614ae828b7939926cb31bdf2
3
+ size 40
flax_model.msgpack ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f83fca6aac012f0d8d1d7998627499ab247ce39eb6e24ea266ae1c63b2f00522
3
+ size 2353635949
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
+ "feature_size": 1,
5
+ "padding_side": "right",
6
+ "padding_value": 0.0,
7
+ "return_attention_mask": true,
8
+ "sampling_rate": 16000
9
+ }
run_librispeech.sh CHANGED
@@ -10,16 +10,16 @@ python run_flax_speech_recognition_seq2seq.py \
10
  --preprocessing_num_workers="16" \
11
  --length_column_name="input_length" \
12
  --overwrite_output_dir \
13
- --num_train_epochs="15" \
14
- --per_device_train_batch_size="2" \
15
  --per_device_eval_batch_size="2" \
16
  --gradient_accumulation_steps="1" \
17
  --logging_steps="25" \
18
- --max_duration_in_seconds="15" \
19
  --max_target_length="64" \
20
  --generation_max_length="40" \
21
  --generation_num_beams="1" \
22
- --learning_rate="3e-4" \
23
  --warmup_steps="500" \
24
  --text_column_name="text" \
25
  --save_total_limit="1" \
 
10
  --preprocessing_num_workers="16" \
11
  --length_column_name="input_length" \
12
  --overwrite_output_dir \
13
+ --num_train_epochs="10" \
14
+ --per_device_train_batch_size="4" \
15
  --per_device_eval_batch_size="2" \
16
  --gradient_accumulation_steps="1" \
17
  --logging_steps="25" \
18
+ --max_duration_in_seconds="10" \
19
  --max_target_length="64" \
20
  --generation_max_length="40" \
21
  --generation_num_beams="1" \
22
+ --learning_rate="1e-4" \
23
  --warmup_steps="500" \
24
  --text_column_name="text" \
25
  --save_total_limit="1" \
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "sep_token": "</s>", "pad_token": "<pad>", "cls_token": "<s>", "mask_token": {"content": "<mask>", "single_word": false, "lstrip": true, "rstrip": false, "normalized": false}}
sweep.yaml ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ command:
2
+ - python3
3
+ - ${program}
4
+ - --overwrite_output_dir
5
+ - --freeze_feature_encoder
6
+ - --predict_with_generate
7
+ - --do_lower_case
8
+ - --do_train
9
+ - --do_eval
10
+ - ${args}
11
+ method: random
12
+ metric:
13
+ goal: minimize
14
+ name: eval/wer
15
+ parameters:
16
+ dataset_cache_dir:
17
+ value: /home/sanchitgandhi/cache/huggingface/datasets
18
+ dataset_config_name:
19
+ value: clean
20
+ dataset_name:
21
+ value: librispeech_asr
22
+ eval_split_name:
23
+ value: validation
24
+ generation_max_length:
25
+ value: 40
26
+ generation_num_beams:
27
+ value: 1
28
+ gradient_accumulation_steps:
29
+ value: 1
30
+ learning_rate:
31
+ distribution: log_uniform
32
+ max: -6.9
33
+ min: -11.5
34
+ length_column_name:
35
+ value: input_length
36
+ logging_steps:
37
+ value: 1
38
+ matmul_precision:
39
+ value: highest
40
+ max_duration_in_seconds:
41
+ value: 15
42
+ max_target_length:
43
+ value: 64
44
+ model_name_or_path:
45
+ value: ./
46
+ num_train_epochs:
47
+ value: 5
48
+ output_dir:
49
+ value: ./
50
+ per_device_eval_batch_size:
51
+ value: 2
52
+ per_device_train_batch_size:
53
+ value: 2
54
+ preprocessing_num_workers:
55
+ value: 16
56
+ text_column_name:
57
+ value: text
58
+ train_split_name:
59
+ value: train.100
60
+ wandb_project:
61
+ value: flax-wav2vec2-2-bart-large-cnn
62
+ warmup_steps:
63
+ value: 500
64
+ program: run_flax_speech_recognition_seq2seq.py
65
+ project: flax-wav2vec2-2-bart-large-cnn
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "bos_token": "<s>", "eos_token": "</s>", "sep_token": "</s>", "cls_token": "<s>", "unk_token": "<unk>", "pad_token": "<pad>", "mask_token": "<mask>", "add_prefix_space": false, "trim_offsets": true, "model_max_length": 1024, "special_tokens_map_file": null, "name_or_path": "./", "tokenizer_class": "BartTokenizer"}
vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug-cli.log ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-04-04 18:09:43 INFO Running runs: []
2
+ 2022-04-04 18:09:43 INFO Agent received command: run
3
+ 2022-04-04 18:09:43 INFO Agent starting run with config:
4
+ dataset_cache_dir: /home/sanchitgandhi/cache/huggingface/datasets
5
+ dataset_config_name: clean
6
+ dataset_name: librispeech_asr
7
+ eval_split_name: validation
8
+ generation_max_length: 40
9
+ generation_num_beams: 1
10
+ gradient_accumulation_steps: 1
11
+ learning_rate: 2.565346074198426e-05
12
+ length_column_name: input_length
13
+ logging_steps: 1
14
+ matmul_precision: highest
15
+ max_duration_in_seconds: 15
16
+ max_target_length: 64
17
+ min_duration_in_seconds: 15
18
+ model_name_or_path: ./
19
+ num_train_epochs: 5
20
+ output_dir: ./
21
+ per_device_eval_batch_size: 2
22
+ per_device_train_batch_size: 2
23
+ preprocessing_num_workers: 16
24
+ text_column_name: text
25
+ train_split_name: train.100
26
+ wandb_project: flax-wav2vec2-2-bart-large-cnn
27
+ warmup_steps: 500
28
+ 2022-04-04 18:09:43 INFO About to run command: python3 run_flax_speech_recognition_seq2seq.py --overwrite_output_dir --freeze_feature_encoder --predict_with_generate --do_lower_case --do_train --do_eval --dataset_cache_dir=/home/sanchitgandhi/cache/huggingface/datasets --dataset_config_name=clean --dataset_name=librispeech_asr --eval_split_name=validation --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=1 --learning_rate=2.565346074198426e-05 --length_column_name=input_length --logging_steps=1 --matmul_precision=highest --max_duration_in_seconds=15 --max_target_length=64 --min_duration_in_seconds=15 --model_name_or_path=./ --num_train_epochs=5 --output_dir=./ --per_device_eval_batch_size=2 --per_device_train_batch_size=2 --preprocessing_num_workers=16 --text_column_name=text --train_split_name=train.100 --wandb_project=flax-wav2vec2-2-bart-large-cnn --warmup_steps=500
29
+ 2022-04-04 18:09:48 INFO Running runs: ['p4wqexfj']
30
+ 2022-04-04 18:10:23 INFO Cleaning up finished run: p4wqexfj
31
+ 2022-04-04 18:10:24 INFO Agent received command: run
32
+ 2022-04-04 18:10:24 INFO Agent starting run with config:
33
+ dataset_cache_dir: /home/sanchitgandhi/cache/huggingface/datasets
34
+ dataset_config_name: clean
35
+ dataset_name: librispeech_asr
36
+ eval_split_name: validation
37
+ generation_max_length: 40
38
+ generation_num_beams: 1
39
+ gradient_accumulation_steps: 1
40
+ learning_rate: 0.0006871268347239357
41
+ length_column_name: input_length
42
+ logging_steps: 1
43
+ matmul_precision: highest
44
+ max_duration_in_seconds: 15
45
+ max_target_length: 64
46
+ min_duration_in_seconds: 15
47
+ model_name_or_path: ./
48
+ num_train_epochs: 5
49
+ output_dir: ./
50
+ per_device_eval_batch_size: 2
51
+ per_device_train_batch_size: 2
52
+ preprocessing_num_workers: 16
53
+ text_column_name: text
54
+ train_split_name: train.100
55
+ wandb_project: flax-wav2vec2-2-bart-large-cnn
56
+ warmup_steps: 500
57
+ 2022-04-04 18:10:24 INFO About to run command: python3 run_flax_speech_recognition_seq2seq.py --overwrite_output_dir --freeze_feature_encoder --predict_with_generate --do_lower_case --do_train --do_eval --dataset_cache_dir=/home/sanchitgandhi/cache/huggingface/datasets --dataset_config_name=clean --dataset_name=librispeech_asr --eval_split_name=validation --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=1 --learning_rate=0.0006871268347239357 --length_column_name=input_length --logging_steps=1 --matmul_precision=highest --max_duration_in_seconds=15 --max_target_length=64 --min_duration_in_seconds=15 --model_name_or_path=./ --num_train_epochs=5 --output_dir=./ --per_device_eval_batch_size=2 --per_device_train_batch_size=2 --preprocessing_num_workers=16 --text_column_name=text --train_split_name=train.100 --wandb_project=flax-wav2vec2-2-bart-large-cnn --warmup_steps=500
58
+ 2022-04-04 18:10:29 INFO Running runs: ['mgg9caus']
59
+ 2022-04-04 18:10:59 INFO Cleaning up finished run: mgg9caus
60
+ 2022-04-04 18:10:59 INFO Agent received command: run
61
+ 2022-04-04 18:10:59 INFO Agent starting run with config:
62
+ dataset_cache_dir: /home/sanchitgandhi/cache/huggingface/datasets
63
+ dataset_config_name: clean
64
+ dataset_name: librispeech_asr
65
+ eval_split_name: validation
66
+ generation_max_length: 40
67
+ generation_num_beams: 1
68
+ gradient_accumulation_steps: 1
69
+ learning_rate: 9.383495031304748e-05
70
+ length_column_name: input_length
71
+ logging_steps: 1
72
+ matmul_precision: highest
73
+ max_duration_in_seconds: 15
74
+ max_target_length: 64
75
+ min_duration_in_seconds: 15
76
+ model_name_or_path: ./
77
+ num_train_epochs: 5
78
+ output_dir: ./
79
+ per_device_eval_batch_size: 2
80
+ per_device_train_batch_size: 2
81
+ preprocessing_num_workers: 16
82
+ text_column_name: text
83
+ train_split_name: train.100
84
+ wandb_project: flax-wav2vec2-2-bart-large-cnn
85
+ warmup_steps: 500
86
+ 2022-04-04 18:10:59 INFO About to run command: python3 run_flax_speech_recognition_seq2seq.py --overwrite_output_dir --freeze_feature_encoder --predict_with_generate --do_lower_case --do_train --do_eval --dataset_cache_dir=/home/sanchitgandhi/cache/huggingface/datasets --dataset_config_name=clean --dataset_name=librispeech_asr --eval_split_name=validation --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=1 --learning_rate=9.383495031304748e-05 --length_column_name=input_length --logging_steps=1 --matmul_precision=highest --max_duration_in_seconds=15 --max_target_length=64 --min_duration_in_seconds=15 --model_name_or_path=./ --num_train_epochs=5 --output_dir=./ --per_device_eval_batch_size=2 --per_device_train_batch_size=2 --preprocessing_num_workers=16 --text_column_name=text --train_split_name=train.100 --wandb_project=flax-wav2vec2-2-bart-large-cnn --warmup_steps=500
87
+ 2022-04-04 18:11:04 INFO Running runs: ['88xgr1fg']
88
+ 2022-04-04 18:11:35 INFO Cleaning up finished run: 88xgr1fg
89
+ 2022-04-04 18:11:35 INFO Agent received command: run
90
+ 2022-04-04 18:11:35 INFO Agent starting run with config:
91
+ dataset_cache_dir: /home/sanchitgandhi/cache/huggingface/datasets
92
+ dataset_config_name: clean
93
+ dataset_name: librispeech_asr
94
+ eval_split_name: validation
95
+ generation_max_length: 40
96
+ generation_num_beams: 1
97
+ gradient_accumulation_steps: 1
98
+ learning_rate: 7.331199736432637e-05
99
+ length_column_name: input_length
100
+ logging_steps: 1
101
+ matmul_precision: highest
102
+ max_duration_in_seconds: 15
103
+ max_target_length: 64
104
+ min_duration_in_seconds: 15
105
+ model_name_or_path: ./
106
+ num_train_epochs: 5
107
+ output_dir: ./
108
+ per_device_eval_batch_size: 2
109
+ per_device_train_batch_size: 2
110
+ preprocessing_num_workers: 16
111
+ text_column_name: text
112
+ train_split_name: train.100
113
+ wandb_project: flax-wav2vec2-2-bart-large-cnn
114
+ warmup_steps: 500
115
+ 2022-04-04 18:11:35 INFO About to run command: python3 run_flax_speech_recognition_seq2seq.py --overwrite_output_dir --freeze_feature_encoder --predict_with_generate --do_lower_case --do_train --do_eval --dataset_cache_dir=/home/sanchitgandhi/cache/huggingface/datasets --dataset_config_name=clean --dataset_name=librispeech_asr --eval_split_name=validation --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=1 --learning_rate=7.331199736432637e-05 --length_column_name=input_length --logging_steps=1 --matmul_precision=highest --max_duration_in_seconds=15 --max_target_length=64 --min_duration_in_seconds=15 --model_name_or_path=./ --num_train_epochs=5 --output_dir=./ --per_device_eval_batch_size=2 --per_device_train_batch_size=2 --preprocessing_num_workers=16 --text_column_name=text --train_split_name=train.100 --wandb_project=flax-wav2vec2-2-bart-large-cnn --warmup_steps=500
116
+ 2022-04-04 18:11:40 INFO Running runs: ['xmgtui21']
117
+ 2022-04-04 18:12:15 INFO Cleaning up finished run: xmgtui21
118
+ 2022-04-04 18:12:15 INFO Agent received command: run
119
+ 2022-04-04 18:12:15 INFO Agent starting run with config:
120
+ dataset_cache_dir: /home/sanchitgandhi/cache/huggingface/datasets
121
+ dataset_config_name: clean
122
+ dataset_name: librispeech_asr
123
+ eval_split_name: validation
124
+ generation_max_length: 40
125
+ generation_num_beams: 1
126
+ gradient_accumulation_steps: 1
127
+ learning_rate: 0.0007642424770238645
128
+ length_column_name: input_length
129
+ logging_steps: 1
130
+ matmul_precision: highest
131
+ max_duration_in_seconds: 15
132
+ max_target_length: 64
133
+ min_duration_in_seconds: 15
134
+ model_name_or_path: ./
135
+ num_train_epochs: 5
136
+ output_dir: ./
137
+ per_device_eval_batch_size: 2
138
+ per_device_train_batch_size: 2
139
+ preprocessing_num_workers: 16
140
+ text_column_name: text
141
+ train_split_name: train.100
142
+ wandb_project: flax-wav2vec2-2-bart-large-cnn
143
+ warmup_steps: 500
144
+ 2022-04-04 18:12:15 INFO About to run command: python3 run_flax_speech_recognition_seq2seq.py --overwrite_output_dir --freeze_feature_encoder --predict_with_generate --do_lower_case --do_train --do_eval --dataset_cache_dir=/home/sanchitgandhi/cache/huggingface/datasets --dataset_config_name=clean --dataset_name=librispeech_asr --eval_split_name=validation --generation_max_length=40 --generation_num_beams=1 --gradient_accumulation_steps=1 --learning_rate=0.0007642424770238645 --length_column_name=input_length --logging_steps=1 --matmul_precision=highest --max_duration_in_seconds=15 --max_target_length=64 --min_duration_in_seconds=15 --model_name_or_path=./ --num_train_epochs=5 --output_dir=./ --per_device_eval_batch_size=2 --per_device_train_batch_size=2 --preprocessing_num_workers=16 --text_column_name=text --train_split_name=train.100 --wandb_project=flax-wav2vec2-2-bart-large-cnn --warmup_steps=500
145
+ 2022-04-04 18:12:20 INFO Running runs: ['4s004g1k']
146
+ 2022-04-04 18:12:51 ERROR Detected 5 failed runs in a row, shutting down.
147
+ 2022-04-04 18:12:51 INFO To change this value set WANDB_AGENT_MAX_INITIAL_FAILURES=val
wandb/debug-internal.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20220405_210343-1w1tow7f/logs/debug-internal.log
wandb/debug.log ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20220405_210343-1w1tow7f/logs/debug.log
wandb/latest-run ADDED
@@ -0,0 +1 @@
 
 
1
+ run-20220405_210343-1w1tow7f
wandb/run-20220404_180947-p4wqexfj/files/config.yaml ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ cli_version: 0.12.11
7
+ framework: huggingface
8
+ huggingface_version: 4.18.0.dev0
9
+ is_jupyter_run: false
10
+ is_kaggle_kernel: false
11
+ python_version: 3.8.10
12
+ start_time: 1649095787
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 11
19
+ - 12
20
+ - 45
21
+ - 49
22
+ - 51
23
+ 2:
24
+ - 1
25
+ - 2
26
+ - 3
27
+ - 11
28
+ - 12
29
+ - 45
30
+ - 49
31
+ - 51
32
+ 4: 3.8.10
33
+ 5: 0.12.11
34
+ 6: 4.18.0.dev0
35
+ 8:
36
+ - 5
37
+ dataset_cache_dir:
38
+ desc: null
39
+ value: /home/sanchitgandhi/cache/huggingface/datasets
40
+ dataset_config_name:
41
+ desc: null
42
+ value: clean
43
+ dataset_name:
44
+ desc: null
45
+ value: librispeech_asr
46
+ eval_split_name:
47
+ desc: null
48
+ value: validation
49
+ generation_max_length:
50
+ desc: null
51
+ value: 40
52
+ generation_num_beams:
53
+ desc: null
54
+ value: 1
55
+ gradient_accumulation_steps:
56
+ desc: null
57
+ value: 1
58
+ learning_rate:
59
+ desc: null
60
+ value: 2.565346074198426e-05
61
+ length_column_name:
62
+ desc: null
63
+ value: input_length
64
+ logging_steps:
65
+ desc: null
66
+ value: 1
67
+ matmul_precision:
68
+ desc: null
69
+ value: highest
70
+ max_duration_in_seconds:
71
+ desc: null
72
+ value: 15
73
+ max_target_length:
74
+ desc: null
75
+ value: 64
76
+ min_duration_in_seconds:
77
+ desc: null
78
+ value: 15
79
+ model_name_or_path:
80
+ desc: null
81
+ value: ./
82
+ num_train_epochs:
83
+ desc: null
84
+ value: 5
85
+ output_dir:
86
+ desc: null
87
+ value: ./
88
+ per_device_eval_batch_size:
89
+ desc: null
90
+ value: 2
91
+ per_device_train_batch_size:
92
+ desc: null
93
+ value: 2
94
+ preprocessing_num_workers:
95
+ desc: null
96
+ value: 16
97
+ text_column_name:
98
+ desc: null
99
+ value: text
100
+ train_split_name:
101
+ desc: null
102
+ value: train.100
103
+ wandb_project:
104
+ desc: null
105
+ value: flax-wav2vec2-2-bart-large-cnn
106
+ warmup_steps:
107
+ desc: null
108
+ value: 500
wandb/run-20220404_180947-p4wqexfj/files/output.log ADDED
@@ -0,0 +1,788 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 04/04/2022 18:09:50 - INFO - __main__ - Training/evaluation parameters FlaxSeq2SeqTrainingArguments(
2
+ _n_gpu=0,
3
+ adafactor=False,
4
+ adam_beta1=0.9,
5
+ adam_beta2=0.999,
6
+ adam_epsilon=1e-08,
7
+ bf16=False,
8
+ bf16_full_eval=False,
9
+ data_seed=None,
10
+ dataloader_drop_last=False,
11
+ dataloader_num_workers=0,
12
+ dataloader_pin_memory=True,
13
+ ddp_bucket_cap_mb=None,
14
+ ddp_find_unused_parameters=None,
15
+ debug=[],
16
+ deepspeed=None,
17
+ disable_tqdm=False,
18
+ do_eval=True,
19
+ do_predict=False,
20
+ do_train=True,
21
+ eval_accumulation_steps=None,
22
+ eval_delay=0,
23
+ eval_steps=None,
24
+ evaluation_strategy=IntervalStrategy.NO,
25
+ fp16=False,
26
+ fp16_backend=auto,
27
+ fp16_full_eval=False,
28
+ fp16_opt_level=O1,
29
+ generation_max_length=40,
30
+ generation_num_beams=1,
31
+ gradient_accumulation_steps=1,
32
+ gradient_checkpointing=False,
33
+ greater_is_better=None,
34
+ group_by_length=False,
35
+ half_precision_backend=auto,
36
+ hub_model_id=None,
37
+ hub_strategy=HubStrategy.EVERY_SAVE,
38
+ hub_token=<HUB_TOKEN>,
39
+ ignore_data_skip=False,
40
+ label_names=None,
41
+ label_smoothing_factor=0.0,
42
+ learning_rate=2.565346074198426e-05,
43
+ length_column_name=input_length,
44
+ load_best_model_at_end=False,
45
+ local_rank=-1,
46
+ log_level=-1,
47
+ log_level_replica=-1,
48
+ log_on_each_node=True,
49
+ logging_dir=./runs/Apr04_18-09-47_t1v-n-4eb331dd-w-0,
50
+ logging_first_step=False,
51
+ logging_nan_inf_filter=True,
52
+ logging_steps=1,
53
+ logging_strategy=IntervalStrategy.STEPS,
54
+ lr_scheduler_type=SchedulerType.LINEAR,
55
+ matmul_precision=highest,
56
+ max_grad_norm=1.0,
57
+ max_steps=-1,
58
+ metric_for_best_model=None,
59
+ mixed_precision=False,
60
+ mp_parameters=,
61
+ no_cuda=False,
62
+ num_train_epochs=5.0,
63
+ optim=OptimizerNames.ADAMW_HF,
64
+ output_dir=./,
65
+ overwrite_output_dir=True,
66
+ past_index=-1,
67
+ per_device_eval_batch_size=2,
68
+ per_device_train_batch_size=2,
69
+ predict_with_generate=True,
70
+ prediction_loss_only=False,
71
+ push_to_hub=False,
72
+ push_to_hub_model_id=None,
73
+ push_to_hub_organization=None,
74
+ push_to_hub_token=<PUSH_TO_HUB_TOKEN>,
75
+ remove_unused_columns=True,
76
+ report_to=['tensorboard', 'wandb'],
77
+ resume_from_checkpoint=None,
78
+ run_name=./,
79
+ save_on_each_node=False,
80
+ save_steps=500,
81
+ save_strategy=IntervalStrategy.STEPS,
82
+ save_total_limit=None,
83
+ seed=42,
84
+ sharded_ddp=[],
85
+ skip_memory_metrics=True,
86
+ sortish_sampler=False,
87
+ tf32=None,
88
+ tpu_metrics_debug=False,
89
+ tpu_num_cores=None,
90
+ use_legacy_prediction_loop=False,
91
+ warmup_ratio=0.0,
92
+ warmup_steps=500,
93
+ weight_decay=0.0,
94
+ xpu_backend=None,
95
+ )
96
+ 04/04/2022 18:09:50 - INFO - __main__ - JAX devices: 1, matmul precision: highest
97
+ 04/04/2022 18:09:50 - WARNING - datasets.builder - Reusing dataset librispeech_asr (/home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c)
98
+ 04/04/2022 18:09:51 - WARNING - datasets.builder - Reusing dataset librispeech_asr (/home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c)
99
+ loading configuration file ./config.json
100
+ Model config SpeechEncoderDecoderConfig {
101
+ "_name_or_path": "./",
102
+ "architectures": [
103
+ "SpeechEncoderDecoderModel"
104
+ ],
105
+ "decoder": {
106
+ "_name_or_path": "facebook/bart-large-cnn",
107
+ "_num_labels": 3,
108
+ "activation_dropout": 0.0,
109
+ "activation_function": "gelu",
110
+ "add_cross_attention": true,
111
+ "add_final_layer_norm": false,
112
+ "architectures": [
113
+ "BartForConditionalGeneration"
114
+ ],
115
+ "attention_dropout": 0.0,
116
+ "bad_words_ids": null,
117
+ "bos_token_id": 0,
118
+ "chunk_size_feed_forward": 0,
119
+ "classif_dropout": 0.0,
120
+ "classifier_dropout": 0.0,
121
+ "cross_attention_hidden_size": null,
122
+ "d_model": 1024,
123
+ "decoder_attention_heads": 16,
124
+ "decoder_ffn_dim": 4096,
125
+ "decoder_layerdrop": 0.0,
126
+ "decoder_layers": 12,
127
+ "decoder_start_token_id": 2,
128
+ "diversity_penalty": 0.0,
129
+ "do_sample": false,
130
+ "dropout": 0.1,
131
+ "early_stopping": true,
132
+ "encoder_attention_heads": 16,
133
+ "encoder_ffn_dim": 4096,
134
+ "encoder_layerdrop": 0.0,
135
+ "encoder_layers": 12,
136
+ "encoder_no_repeat_ngram_size": 0,
137
+ "eos_token_id": 2,
138
+ "exponential_decay_length_penalty": null,
139
+ "finetuning_task": null,
140
+ "force_bos_token_to_be_generated": true,
141
+ "forced_bos_token_id": 0,
142
+ "forced_eos_token_id": 2,
143
+ "gradient_checkpointing": false,
144
+ "id2label": {
145
+ "0": "LABEL_0",
146
+ "1": "LABEL_1",
147
+ "2": "LABEL_2"
148
+ },
149
+ "init_std": 0.02,
150
+ "is_decoder": true,
151
+ "is_encoder_decoder": false,
152
+ "label2id": {
153
+ "LABEL_0": 0,
154
+ "LABEL_1": 1,
155
+ "LABEL_2": 2
156
+ },
157
+ "length_penalty": 2.0,
158
+ "max_length": 142,
159
+ "max_position_embeddings": 1024,
160
+ "min_length": 56,
161
+ "model_type": "bart",
162
+ "no_repeat_ngram_size": 3,
163
+ "normalize_before": false,
164
+ "num_beam_groups": 1,
165
+ "num_beams": 4,
166
+ "num_hidden_layers": 12,
167
+ "num_return_sequences": 1,
168
+ "output_attentions": false,
169
+ "output_hidden_states": false,
170
+ "output_past": true,
171
+ "output_scores": false,
172
+ "pad_token_id": 1,
173
+ "prefix": " ",
174
+ "problem_type": null,
175
+ "pruned_heads": {},
176
+ "remove_invalid_values": false,
177
+ "repetition_penalty": 1.0,
178
+ "return_dict": true,
179
+ "return_dict_in_generate": false,
180
+ "scale_embedding": false,
181
+ "sep_token_id": null,
182
+ "task_specific_params": {
183
+ "summarization": {
184
+ "early_stopping": true,
185
+ "length_penalty": 2.0,
186
+ "max_length": 142,
187
+ "min_length": 56,
188
+ "no_repeat_ngram_size": 3,
189
+ "num_beams": 4
190
+ }
191
+ },
192
+ "temperature": 1.0,
193
+ "tie_encoder_decoder": false,
194
+ "tie_word_embeddings": true,
195
+ "tokenizer_class": null,
196
+ "top_k": 50,
197
+ "top_p": 1.0,
198
+ "torch_dtype": null,
199
+ "torchscript": false,
200
+ "transformers_version": "4.18.0.dev0",
201
+ "typical_p": 1.0,
202
+ "use_bfloat16": false,
203
+ "use_cache": true,
204
+ "vocab_size": 50264
205
+ },
206
+ "decoder_start_token_id": 0,
207
+ "encoder": {
208
+ "_name_or_path": "facebook/wav2vec2-large-lv60",
209
+ "activation_dropout": 0.1,
210
+ "adapter_kernel_size": 3,
211
+ "adapter_stride": 2,
212
+ "add_adapter": true,
213
+ "add_cross_attention": false,
214
+ "apply_spec_augment": true,
215
+ "architectures": [
216
+ "Wav2Vec2ForPreTraining"
217
+ ],
218
+ "attention_dropout": 0.1,
219
+ "bad_words_ids": null,
220
+ "bos_token_id": 1,
221
+ "chunk_size_feed_forward": 0,
222
+ "classifier_proj_size": 256,
223
+ "codevector_dim": 768,
224
+ "contrastive_logits_temperature": 0.1,
225
+ "conv_bias": true,
226
+ "conv_dim": [
227
+ 512,
228
+ 512,
229
+ 512,
230
+ 512,
231
+ 512,
232
+ 512,
233
+ 512
234
+ ],
235
+ "conv_kernel": [
236
+ 10,
237
+ 3,
238
+ 3,
239
+ 3,
240
+ 3,
241
+ 2,
242
+ 2
243
+ ],
244
+ "conv_stride": [
245
+ 5,
246
+ 2,
247
+ 2,
248
+ 2,
249
+ 2,
250
+ 2,
251
+ 2
252
+ ],
253
+ "cross_attention_hidden_size": null,
254
+ "ctc_loss_reduction": "sum",
255
+ "ctc_zero_infinity": false,
256
+ "decoder_start_token_id": null,
257
+ "diversity_loss_weight": 0.1,
258
+ "diversity_penalty": 0.0,
259
+ "do_sample": false,
260
+ "do_stable_layer_norm": true,
261
+ "early_stopping": false,
262
+ "encoder_no_repeat_ngram_size": 0,
263
+ "eos_token_id": 2,
264
+ "exponential_decay_length_penalty": null,
265
+ "feat_extract_activation": "gelu",
266
+ "feat_extract_dropout": 0.0,
267
+ "feat_extract_norm": "layer",
268
+ "feat_proj_dropout": 0.0,
269
+ "feat_quantizer_dropout": 0.0,
270
+ "final_dropout": 0.0,
271
+ "finetuning_task": null,
272
+ "forced_bos_token_id": null,
273
+ "forced_eos_token_id": null,
274
+ "gradient_checkpointing": false,
275
+ "hidden_act": "gelu",
276
+ "hidden_dropout": 0.1,
277
+ "hidden_dropout_prob": 0.1,
278
+ "hidden_size": 1024,
279
+ "id2label": {
280
+ "0": "LABEL_0",
281
+ "1": "LABEL_1"
282
+ },
283
+ "initializer_range": 0.02,
284
+ "intermediate_size": 4096,
285
+ "is_decoder": false,
286
+ "is_encoder_decoder": false,
287
+ "label2id": {
288
+ "LABEL_0": 0,
289
+ "LABEL_1": 1
290
+ },
291
+ "layer_norm_eps": 1e-05,
292
+ "layerdrop": 0.0,
293
+ "length_penalty": 1.0,
294
+ "mask_feature_length": 10,
295
+ "mask_feature_min_masks": 0,
296
+ "mask_feature_prob": 0.0,
297
+ "mask_time_length": 10,
298
+ "mask_time_min_masks": 2,
299
+ "mask_time_prob": 0.1,
300
+ "max_length": 20,
301
+ "min_length": 0,
302
+ "model_type": "wav2vec2",
303
+ "no_repeat_ngram_size": 0,
304
+ "num_adapter_layers": 3,
305
+ "num_attention_heads": 16,
306
+ "num_beam_groups": 1,
307
+ "num_beams": 1,
308
+ "num_codevector_groups": 2,
309
+ "num_codevectors_per_group": 320,
310
+ "num_conv_pos_embedding_groups": 16,
311
+ "num_conv_pos_embeddings": 128,
312
+ "num_feat_extract_layers": 7,
313
+ "num_hidden_layers": 24,
314
+ "num_negatives": 100,
315
+ "num_return_sequences": 1,
316
+ "output_attentions": false,
317
+ "output_hidden_size": 1024,
318
+ "output_hidden_states": false,
319
+ "output_scores": false,
320
+ "pad_token_id": 0,
321
+ "prefix": null,
322
+ "problem_type": null,
323
+ "proj_codevector_dim": 768,
324
+ "pruned_heads": {},
325
+ "remove_invalid_values": false,
326
+ "repetition_penalty": 1.0,
327
+ "return_dict": true,
328
+ "return_dict_in_generate": false,
329
+ "sep_token_id": null,
330
+ "task_specific_params": null,
331
+ "tdnn_dilation": [
332
+ 1,
333
+ 2,
334
+ 3,
335
+ 1,
336
+ 1
337
+ ],
338
+ "tdnn_dim": [
339
+ 512,
340
+ 512,
341
+ 512,
342
+ 512,
343
+ 1500
344
+ ],
345
+ "tdnn_kernel": [
346
+ 5,
347
+ 3,
348
+ 3,
349
+ 1,
350
+ 1
351
+ ],
352
+ "temperature": 1.0,
353
+ "tie_encoder_decoder": false,
354
+ "tie_word_embeddings": true,
355
+ "tokenizer_class": null,
356
+ "top_k": 50,
357
+ "top_p": 1.0,
358
+ "torch_dtype": null,
359
+ "torchscript": false,
360
+ "transformers_version": "4.18.0.dev0",
361
+ "typical_p": 1.0,
362
+ "use_bfloat16": false,
363
+ "use_weighted_layer_sum": false,
364
+ "vocab_size": 32,
365
+ "xvector_output_dim": 512
366
+ },
367
+ "eos_token_id": 2,
368
+ "is_encoder_decoder": true,
369
+ "max_length": 40,
370
+ "model_type": "speech-encoder-decoder",
371
+ "pad_token_id": 1,
372
+ "processor_class": "Wav2Vec2Processor",
373
+ "tie_word_embeddings": false,
374
+ "transformers_version": null,
375
+ "use_cache": false
376
+ }
377
+ loading feature extractor configuration file ./preprocessor_config.json
378
+ Feature extractor Wav2Vec2FeatureExtractor {
379
+ "do_normalize": true,
380
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
381
+ "feature_size": 1,
382
+ "padding_side": "right",
383
+ "padding_value": 0.0,
384
+ "return_attention_mask": true,
385
+ "sampling_rate": 16000
386
+ }
387
+ Didn't find file ./added_tokens.json. We won't load it.
388
+ loading file ./vocab.json
389
+ loading file ./merges.txt
390
+ loading file ./tokenizer.json
391
+ loading file None
392
+ loading file ./special_tokens_map.json
393
+ loading file ./tokenizer_config.json
394
+ loading weights file ./flax_model.msgpack
395
+ tcmalloc: large alloc 2353643520 bytes == 0x9c48e000 @ 0x7ffb21ca1680 0x7ffb21cc2824 0x5f8a01 0x648cf1 0x5c4676 0x4f290e 0x64f718 0x5048b3 0x56b1da 0x56939a 0x5f6a13 0x50af49 0x5f3547 0x56c8cd 0x56939a 0x50aaa0 0x56c28c 0x56939a 0x5f6a13 0x56b0ae 0x56939a 0x68d047 0x67e351 0x67e3cf 0x67e471 0x67e817 0x6b6fe2 0x6b736d 0x7ffb21ab30b3 0x5fa5ce
396
+ All model checkpoint weights were used when initializing FlaxSpeechEncoderDecoderModel.
397
+ All the weights of FlaxSpeechEncoderDecoderModel were initialized from the model checkpoint at ./.
398
+ If your task is similar to the task the model of the checkpoint was trained on, you can already use FlaxSpeechEncoderDecoderModel for predictions without further training.
399
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-eb8e77ddeb6c34dc.arrow
400
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e19eb46cb05f7b14.arrow
401
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-92685008e5c09846.arrow
402
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-0d478bd8b093039a.arrow
403
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-13aa6d940e98be09.arrow
404
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-2d24e62de045a0c8.arrow
405
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-c6e972d87654db26.arrow
406
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-4771f1765b989007.arrow
407
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ab9a2fefe02761db.arrow
408
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-6ce32d0e344f4c06.arrow
409
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-1e0f4ae48a3c8b80.arrow
410
+ 04/04/2022 18:10:09 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e6f5ab8655734a9f.arrow
411
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-be6a3a97adc5697a.arrow
412
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-239fcccc5b67f7a4.arrow
413
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ff31cd13241c8bf3.arrow
414
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e8bf216b19718031.arrow
415
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-03ee7ff6632282a6.arrow
416
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-bb1b7de42a389f66.arrow
417
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ed92b3340af62815.arrow
418
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-12ce4b2b8350532d.arrow
419
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-75f20b5cb7b89a9d.arrow
420
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-69941cf3a70dde04.arrow
421
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-f4d0102728859e5f.arrow
422
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-1c0eaf46e759e04d.arrow
423
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-9f8c4dccc4fa8f17.arrow
424
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-14df38b848e61b52.arrow
425
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-19c662ecc5b20d66.arrow
426
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-3276de10dacb4963.arrow
427
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-111057e4500c4cc5.arrow
428
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-d1bf30028d638290.arrow
429
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cc50c08263e878fc.arrow
430
+ 04/04/2022 18:10:10 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-f1813a8ae49a67d6.arrow
431
+ #1: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 161.70ba/s]
432
+ #0: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 116.51ba/s]
433
+ #2: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 163.38ba/s]
434
+ #3: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 156.34ba/s]
435
+ #4: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 150.68ba/s]
436
+ #6: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 124.88ba/s]
437
+ #5: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 118.73ba/s]
438
+ #7: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 129.85ba/s]
439
+ #8: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 132.48ba/s]
440
+ #9: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 139.58ba/s]
441
+ #11: 100%|███████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 137.78ba/s]
442
+ #10: 100%|████████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 90.63ba/s]
443
+ #13: 100%|███████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 142.51ba/s]
444
+ #12: 100%|███████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 106.71ba/s]
445
+ #14: 100%|███████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 142.20ba/s]
446
+ #15: 100%|███████████████████████████████████████████████████████████████████████████████| 2/2 [00:00<00:00, 143.37ba/s]
447
+ #10: 0%| | 0/2 [00:00<?, ?ba/s]
448
+ #11: 0%| | 0/2 [00:00<?, ?ba/s]
449
+ #12: 0%| | 0/2 [00:00<?, ?ba/s]
450
+ #13: 0%| | 0/2 [00:00<?, ?ba/s]
451
+ #14: 0%| | 0/2 [00:00<?, ?ba/s]
452
+ #10: 100%|███████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 336.24ba/s]
453
+ #11: 100%|███████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 278.97ba/s]
454
+ #12: 100%|█████████████████████��█████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 301.44ba/s]
455
+ #13: 100%|███████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 277.62ba/s]
456
+ #14: 100%|███████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 352.40ba/s]
457
+ #15: 100%|███████████████████████████████████████████████████████████████████████████████| 1/1 [00:00<00:00, 369.80ba/s]
458
+ Feature extractor saved in ./preprocessor_config.json | 0/1 [00:00<?, ?ba/s]
459
+ tokenizer config file saved in ./tokenizer_config.json
460
+ Special tokens file saved in ./special_tokens_map.json | 0/1 [00:00<?, ?ba/s]
461
+ Configuration saved in ./config.json
462
+ loading feature extractor configuration file ./preprocessor_config.json | 0/1 [00:00<?, ?ba/s]
463
+ loading configuration file ./config.json
464
+ Model config SpeechEncoderDecoderConfig { | 0/1 [00:00<?, ?ba/s]
465
+ "_name_or_path": "./",
466
+ "architectures": [ | 0/1 [00:00<?, ?ba/s]
467
+ "SpeechEncoderDecoderModel"
468
+ ], 0%| | 0/1 [00:00<?, ?ba/s]
469
+ "decoder": {
470
+ "_name_or_path": "facebook/bart-large-cnn", | 0/1 [00:00<?, ?ba/s]
471
+ "_num_labels": 3, | 0/1 [00:00<?, ?ba/s]
472
+ "activation_dropout": 0.0,
473
+ "activation_function": "gelu",
474
+ "add_cross_attention": true,
475
+ "add_final_layer_norm": false,
476
+ "architectures": [
477
+ "BartForConditionalGeneration"
478
+ ],
479
+ "attention_dropout": 0.0,
480
+ "bad_words_ids": null,
481
+ "bos_token_id": 0,
482
+ "chunk_size_feed_forward": 0,
483
+ "classif_dropout": 0.0,
484
+ "classifier_dropout": 0.0,
485
+ "cross_attention_hidden_size": null,
486
+ "d_model": 1024,
487
+ "decoder_attention_heads": 16,
488
+ "decoder_ffn_dim": 4096,
489
+ "decoder_layerdrop": 0.0,
490
+ "decoder_layers": 12,
491
+ "decoder_start_token_id": 2,
492
+ "diversity_penalty": 0.0,
493
+ "do_sample": false,
494
+ "dropout": 0.1,
495
+ "early_stopping": true,
496
+ "encoder_attention_heads": 16,
497
+ "encoder_ffn_dim": 4096,
498
+ "encoder_layerdrop": 0.0,
499
+ "encoder_layers": 12,
500
+ "encoder_no_repeat_ngram_size": 0,
501
+ "eos_token_id": 2,
502
+ "exponential_decay_length_penalty": null,
503
+ "finetuning_task": null,
504
+ "force_bos_token_to_be_generated": true,
505
+ "forced_bos_token_id": 0,
506
+ "forced_eos_token_id": 2,
507
+ "gradient_checkpointing": false,
508
+ "id2label": {
509
+ "0": "LABEL_0",
510
+ "1": "LABEL_1",
511
+ "2": "LABEL_2"
512
+ },
513
+ "init_std": 0.02,
514
+ "is_decoder": true,
515
+ "is_encoder_decoder": false,
516
+ "label2id": {
517
+ "LABEL_0": 0,
518
+ "LABEL_1": 1,
519
+ "LABEL_2": 2
520
+ },
521
+ "length_penalty": 2.0,
522
+ "max_length": 142,
523
+ "max_position_embeddings": 1024,
524
+ "min_length": 56,
525
+ "model_type": "bart",
526
+ "no_repeat_ngram_size": 3,
527
+ "normalize_before": false,
528
+ "num_beam_groups": 1,
529
+ "num_beams": 4,
530
+ "num_hidden_layers": 12,
531
+ "num_return_sequences": 1,
532
+ "output_attentions": false,
533
+ "output_hidden_states": false,
534
+ "output_past": true,
535
+ "output_scores": false,
536
+ "pad_token_id": 1,
537
+ "prefix": " ",
538
+ "problem_type": null,
539
+ "pruned_heads": {},
540
+ "remove_invalid_values": false,
541
+ "repetition_penalty": 1.0,
542
+ "return_dict": true,
543
+ "return_dict_in_generate": false,
544
+ "scale_embedding": false,
545
+ "sep_token_id": null,
546
+ "task_specific_params": {
547
+ "summarization": {
548
+ "early_stopping": true,
549
+ "length_penalty": 2.0,
550
+ "max_length": 142,
551
+ "min_length": 56,
552
+ "no_repeat_ngram_size": 3,
553
+ "num_beams": 4
554
+ }
555
+ },
556
+ "temperature": 1.0,
557
+ "tie_encoder_decoder": false,
558
+ "tie_word_embeddings": true,
559
+ "tokenizer_class": null,
560
+ "top_k": 50,
561
+ "top_p": 1.0,
562
+ "torch_dtype": null,
563
+ "torchscript": false,
564
+ "transformers_version": "4.18.0.dev0",
565
+ "typical_p": 1.0,
566
+ "use_bfloat16": false,
567
+ "use_cache": true,
568
+ "vocab_size": 50264
569
+ },
570
+ "decoder_start_token_id": 0,
571
+ "encoder": {
572
+ "_name_or_path": "facebook/wav2vec2-large-lv60",
573
+ "activation_dropout": 0.1,
574
+ "adapter_kernel_size": 3,
575
+ "adapter_stride": 2,
576
+ "add_adapter": true,
577
+ "add_cross_attention": false,
578
+ "apply_spec_augment": true,
579
+ "architectures": [
580
+ "Wav2Vec2ForPreTraining"
581
+ ],
582
+ "attention_dropout": 0.1,
583
+ "bad_words_ids": null,
584
+ "bos_token_id": 1,
585
+ "chunk_size_feed_forward": 0,
586
+ "classifier_proj_size": 256,
587
+ "codevector_dim": 768,
588
+ "contrastive_logits_temperature": 0.1,
589
+ "conv_bias": true,
590
+ "conv_dim": [
591
+ 512,
592
+ 512,
593
+ 512,
594
+ 512,
595
+ 512,
596
+ 512,
597
+ 512
598
+ ],
599
+ "conv_kernel": [
600
+ 10,
601
+ 3,
602
+ 3,
603
+ 3,
604
+ 3,
605
+ 2,
606
+ 2
607
+ ],
608
+ "conv_stride": [
609
+ 5,
610
+ 2,
611
+ 2,
612
+ 2,
613
+ 2,
614
+ 2,
615
+ 2
616
+ ],
617
+ "cross_attention_hidden_size": null,
618
+ "ctc_loss_reduction": "sum",
619
+ "ctc_zero_infinity": false,
620
+ "decoder_start_token_id": null,
621
+ "diversity_loss_weight": 0.1,
622
+ "diversity_penalty": 0.0,
623
+ "do_sample": false,
624
+ "do_stable_layer_norm": true,
625
+ "early_stopping": false,
626
+ "encoder_no_repeat_ngram_size": 0,
627
+ "eos_token_id": 2,
628
+ "exponential_decay_length_penalty": null,
629
+ "feat_extract_activation": "gelu",
630
+ "feat_extract_dropout": 0.0,
631
+ "feat_extract_norm": "layer",
632
+ "feat_proj_dropout": 0.0,
633
+ "feat_quantizer_dropout": 0.0,
634
+ "final_dropout": 0.0,
635
+ "finetuning_task": null,
636
+ "forced_bos_token_id": null,
637
+ "forced_eos_token_id": null,
638
+ "gradient_checkpointing": false,
639
+ "hidden_act": "gelu",
640
+ "hidden_dropout": 0.1,
641
+ "hidden_dropout_prob": 0.1,
642
+ "hidden_size": 1024,
643
+ "id2label": {
644
+ "0": "LABEL_0",
645
+ "1": "LABEL_1"
646
+ },
647
+ "initializer_range": 0.02,
648
+ "intermediate_size": 4096,
649
+ "is_decoder": false,
650
+ "is_encoder_decoder": false,
651
+ "label2id": {
652
+ "LABEL_0": 0,
653
+ "LABEL_1": 1
654
+ },
655
+ "layer_norm_eps": 1e-05,
656
+ "layerdrop": 0.0,
657
+ "length_penalty": 1.0,
658
+ "mask_feature_length": 10,
659
+ "mask_feature_min_masks": 0,
660
+ "mask_feature_prob": 0.0,
661
+ "mask_time_length": 10,
662
+ "mask_time_min_masks": 2,
663
+ "mask_time_prob": 0.1,
664
+ "max_length": 20,
665
+ "min_length": 0,
666
+ "model_type": "wav2vec2",
667
+ "no_repeat_ngram_size": 0,
668
+ "num_adapter_layers": 3,
669
+ "num_attention_heads": 16,
670
+ "num_beam_groups": 1,
671
+ "num_beams": 1,
672
+ "num_codevector_groups": 2,
673
+ "num_codevectors_per_group": 320,
674
+ "num_conv_pos_embedding_groups": 16,
675
+ "num_conv_pos_embeddings": 128,
676
+ "num_feat_extract_layers": 7,
677
+ "num_hidden_layers": 24,
678
+ "num_negatives": 100,
679
+ "num_return_sequences": 1,
680
+ "output_attentions": false,
681
+ "output_hidden_size": 1024,
682
+ "output_hidden_states": false,
683
+ "output_scores": false,
684
+ "pad_token_id": 0,
685
+ "prefix": null,
686
+ "problem_type": null,
687
+ "proj_codevector_dim": 768,
688
+ "pruned_heads": {},
689
+ "remove_invalid_values": false,
690
+ "repetition_penalty": 1.0,
691
+ "return_dict": true,
692
+ "return_dict_in_generate": false,
693
+ "sep_token_id": null,
694
+ "task_specific_params": null,
695
+ "tdnn_dilation": [
696
+ 1,
697
+ 2,
698
+ 3,
699
+ 1,
700
+ 1
701
+ ],
702
+ "tdnn_dim": [
703
+ 512,
704
+ 512,
705
+ 512,
706
+ 512,
707
+ 1500
708
+ ],
709
+ "tdnn_kernel": [
710
+ 5,
711
+ 3,
712
+ 3,
713
+ 1,
714
+ 1
715
+ ],
716
+ "temperature": 1.0,
717
+ "tie_encoder_decoder": false,
718
+ "tie_word_embeddings": true,
719
+ "tokenizer_class": null,
720
+ "top_k": 50,
721
+ "top_p": 1.0,
722
+ "torch_dtype": null,
723
+ "torchscript": false,
724
+ "transformers_version": "4.18.0.dev0",
725
+ "typical_p": 1.0,
726
+ "use_bfloat16": false,
727
+ "use_weighted_layer_sum": false,
728
+ "vocab_size": 32,
729
+ "xvector_output_dim": 512
730
+ },
731
+ "eos_token_id": 2,
732
+ "is_encoder_decoder": true,
733
+ "max_length": 40,
734
+ "model_type": "speech-encoder-decoder",
735
+ "pad_token_id": 1,
736
+ "processor_class": "Wav2Vec2Processor",
737
+ "tie_word_embeddings": false,
738
+ "transformers_version": null,
739
+ "use_cache": false
740
+ }
741
+ loading feature extractor configuration file ./preprocessor_config.json
742
+ Feature extractor Wav2Vec2FeatureExtractor {
743
+ "do_normalize": true,
744
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
745
+ "feature_size": 1,
746
+ "padding_side": "right",
747
+ "padding_value": 0.0,
748
+ "return_attention_mask": true,
749
+ "sampling_rate": 16000
750
+ }
751
+ Didn't find file ./added_tokens.json. We won't load it.
752
+ loading file ./vocab.json
753
+ loading file ./merges.txt
754
+ loading file ./tokenizer.json
755
+ loading file None
756
+ loading file ./special_tokens_map.json
757
+ loading file ./tokenizer_config.json
758
+ 2022-04-04 18:10:13.379554: W tensorflow/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libcuda.so.1'; dlerror: libcuda.so.1: cannot open shared object file: No such file or directory
759
+ 2022-04-04 18:10:13.379593: W tensorflow/stream_executor/cuda/cuda_driver.cc:269] failed call to cuInit: UNKNOWN ERROR (303)
760
+ 04/04/2022 18:10:14 - INFO - __main__ - ***** Running training *****
761
+ 04/04/2022 18:10:14 - INFO - __main__ - Num examples = 0
762
+ 04/04/2022 18:10:14 - INFO - __main__ - Num Epochs = 5
763
+ 04/04/2022 18:10:14 - INFO - __main__ - Instantaneous batch size per device = 2
764
+ 04/04/2022 18:10:14 - INFO - __main__ - Num gradient accumulation steps = 1
765
+ 04/04/2022 18:10:14 - INFO - __main__ - Total train batch size (w. parallel & distributed) = 2
766
+ 04/04/2022 18:10:14 - INFO - __main__ - Total optimization steps = 0
767
+ Epoch ... (1/5): 0%| | 0/5 [00:00<?, ?it/s]
768
+ Traceback (most recent call last):
769
+ File "run_flax_speech_recognition_seq2seq.py", line 1133, in <module>
770
+ main()
771
+ File "run_flax_speech_recognition_seq2seq.py", line 1050, in main
772
+ train_samples_idx = get_grouped_indices(vectorized_datasets["train"], batch_size_per_update, input_rng)
773
+ File "run_flax_speech_recognition_seq2seq.py", line 453, in get_grouped_indices
774
+ lengths = dataset["input_length"]
775
+ File "/home/sanchitgandhi/datasets/src/datasets/arrow_dataset.py", line 1765, in __getitem__
776
+ return self._getitem(
777
+ File "/home/sanchitgandhi/datasets/src/datasets/arrow_dataset.py", line 1749, in _getitem
778
+ pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None)
779
+ File "/home/sanchitgandhi/datasets/src/datasets/formatting/formatting.py", line 491, in query_table
780
+ pa_subtable = _query_table_with_indices_mapping(table, key, indices=indices)
781
+ File "/home/sanchitgandhi/datasets/src/datasets/formatting/formatting.py", line 69, in _query_table_with_indices_mapping
782
+ return _query_table(table, indices.column(0).to_pylist())
783
+ File "/home/sanchitgandhi/datasets/src/datasets/table.py", line 343, in column
784
+ return self.table.column(*args, **kwargs)
785
+ File "pyarrow/table.pxi", line 2115, in pyarrow.lib.Table.column
786
+ File "pyarrow/table.pxi", line 2130, in pyarrow.lib.Table._column
787
+ File "pyarrow/array.pxi", line 561, in pyarrow.lib._normalize_index
788
+ IndexError: index out of bounds
wandb/run-20220404_180947-p4wqexfj/files/requirements.txt ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ astunparse==1.6.3
10
+ async-timeout==4.0.2
11
+ attrs==21.4.0
12
+ audioread==2.1.9
13
+ babel==2.9.1
14
+ backcall==0.2.0
15
+ beautifulsoup4==4.10.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.12
21
+ chex==0.1.1
22
+ click==8.1.0
23
+ cycler==0.11.0
24
+ datasets==2.0.1.dev0
25
+ debugpy==1.6.0
26
+ decorator==5.1.1
27
+ defusedxml==0.7.1
28
+ dill==0.3.4
29
+ dm-tree==0.1.6
30
+ docker-pycreds==0.4.0
31
+ entrypoints==0.4
32
+ executing==0.8.3
33
+ filelock==3.6.0
34
+ flatbuffers==2.0
35
+ flax==0.4.1
36
+ fonttools==4.31.2
37
+ frozenlist==1.3.0
38
+ fsspec==2022.2.0
39
+ gast==0.5.3
40
+ gitdb==4.0.9
41
+ gitpython==3.1.27
42
+ google-auth-oauthlib==0.4.6
43
+ google-auth==2.6.2
44
+ google-pasta==0.2.0
45
+ grpcio==1.44.0
46
+ h5py==3.6.0
47
+ huggingface-hub==0.4.0
48
+ idna==3.3
49
+ importlib-metadata==4.11.3
50
+ importlib-resources==5.6.0
51
+ ipdb==0.13.9
52
+ ipykernel==6.10.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.2.0
55
+ jax==0.3.4
56
+ jaxlib==0.3.2
57
+ jedi==0.18.1
58
+ jinja2==3.1.1
59
+ jiwer==2.3.0
60
+ joblib==1.1.0
61
+ json5==0.9.6
62
+ jsonschema==4.4.0
63
+ jupyter-client==7.2.1
64
+ jupyter-core==4.9.2
65
+ jupyter-server==1.16.0
66
+ jupyterlab-pygments==0.1.2
67
+ jupyterlab-server==2.12.0
68
+ jupyterlab==3.3.2
69
+ keras-preprocessing==1.1.2
70
+ keras==2.8.0
71
+ kiwisolver==1.4.2
72
+ libclang==13.0.0
73
+ librosa==0.9.1
74
+ libtpu-nightly==0.1.dev20220315
75
+ llvmlite==0.38.0
76
+ markdown==3.3.6
77
+ markupsafe==2.1.1
78
+ matplotlib-inline==0.1.3
79
+ matplotlib==3.5.1
80
+ mistune==0.8.4
81
+ msgpack==1.0.3
82
+ multidict==6.0.2
83
+ multiprocess==0.70.12.2
84
+ nbclassic==0.3.7
85
+ nbclient==0.5.13
86
+ nbconvert==6.4.5
87
+ nbformat==5.2.0
88
+ nest-asyncio==1.5.4
89
+ notebook-shim==0.1.0
90
+ notebook==6.4.10
91
+ numba==0.55.1
92
+ numpy==1.21.0
93
+ oauthlib==3.2.0
94
+ opt-einsum==3.3.0
95
+ optax==0.1.1
96
+ packaging==21.3
97
+ pandas==1.4.1
98
+ pandocfilters==1.5.0
99
+ parso==0.8.3
100
+ pathtools==0.1.2
101
+ pexpect==4.8.0
102
+ pickleshare==0.7.5
103
+ pillow==9.0.1
104
+ pip==20.0.2
105
+ pkg-resources==0.0.0
106
+ pooch==1.6.0
107
+ prometheus-client==0.13.1
108
+ promise==2.3
109
+ prompt-toolkit==3.0.28
110
+ protobuf==3.19.4
111
+ psutil==5.9.0
112
+ ptyprocess==0.7.0
113
+ pure-eval==0.2.2
114
+ pyarrow==7.0.0
115
+ pyasn1-modules==0.2.8
116
+ pyasn1==0.4.8
117
+ pycparser==2.21
118
+ pygments==2.11.2
119
+ pyparsing==3.0.7
120
+ pyrsistent==0.18.1
121
+ python-dateutil==2.8.2
122
+ python-levenshtein==0.12.2
123
+ pytz==2022.1
124
+ pyyaml==6.0
125
+ pyzmq==22.3.0
126
+ regex==2022.3.15
127
+ requests-oauthlib==1.3.1
128
+ requests==2.27.1
129
+ resampy==0.2.2
130
+ responses==0.18.0
131
+ rsa==4.8
132
+ sacremoses==0.0.49
133
+ scikit-learn==1.0.2
134
+ scipy==1.8.0
135
+ send2trash==1.8.0
136
+ sentry-sdk==1.5.8
137
+ setproctitle==1.2.2
138
+ setuptools==44.0.0
139
+ shortuuid==1.0.8
140
+ six==1.16.0
141
+ smmap==5.0.0
142
+ sniffio==1.2.0
143
+ soundfile==0.10.3.post1
144
+ soupsieve==2.3.1
145
+ stack-data==0.2.0
146
+ tensorboard-data-server==0.6.1
147
+ tensorboard-plugin-wit==1.8.1
148
+ tensorboard==2.8.0
149
+ tensorflow-io-gcs-filesystem==0.24.0
150
+ tensorflow==2.8.0
151
+ termcolor==1.1.0
152
+ terminado==0.13.3
153
+ testpath==0.6.0
154
+ tf-estimator-nightly==2.8.0.dev2021122109
155
+ threadpoolctl==3.1.0
156
+ tokenizers==0.11.6
157
+ toml==0.10.2
158
+ toolz==0.11.2
159
+ torch==1.11.0+cpu
160
+ torchaudio==0.11.0+cpu
161
+ tornado==6.1
162
+ tqdm==4.63.1
163
+ traitlets==5.1.1
164
+ transformers==4.18.0.dev0
165
+ typing-extensions==4.1.1
166
+ urllib3==1.26.9
167
+ wandb==0.12.11
168
+ wcwidth==0.2.5
169
+ webencodings==0.5.1
170
+ websocket-client==1.3.2
171
+ werkzeug==2.1.0
172
+ wheel==0.37.1
173
+ wrapt==1.14.0
174
+ xxhash==3.0.0
175
+ yarl==1.7.2
176
+ yaspin==2.1.0
177
+ zipp==3.7.0
wandb/run-20220404_180947-p4wqexfj/files/wandb-metadata.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-1043-gcp-x86_64-with-glibc2.29",
3
+ "python": "3.8.10",
4
+ "heartbeatAt": "2022-04-04T18:09:50.590240",
5
+ "startedAt": "2022-04-04T18:09:47.395313",
6
+ "docker": null,
7
+ "cpu_count": 96,
8
+ "cuda": null,
9
+ "args": [
10
+ "--overwrite_output_dir",
11
+ "--freeze_feature_encoder",
12
+ "--predict_with_generate",
13
+ "--do_lower_case",
14
+ "--do_train",
15
+ "--do_eval",
16
+ "--dataset_cache_dir=/home/sanchitgandhi/cache/huggingface/datasets",
17
+ "--dataset_config_name=clean",
18
+ "--dataset_name=librispeech_asr",
19
+ "--eval_split_name=validation",
20
+ "--generation_max_length=40",
21
+ "--generation_num_beams=1",
22
+ "--gradient_accumulation_steps=1",
23
+ "--learning_rate=2.565346074198426e-05",
24
+ "--length_column_name=input_length",
25
+ "--logging_steps=1",
26
+ "--matmul_precision=highest",
27
+ "--max_duration_in_seconds=15",
28
+ "--max_target_length=64",
29
+ "--min_duration_in_seconds=15",
30
+ "--model_name_or_path=./",
31
+ "--num_train_epochs=5",
32
+ "--output_dir=./",
33
+ "--per_device_eval_batch_size=2",
34
+ "--per_device_train_batch_size=2",
35
+ "--preprocessing_num_workers=16",
36
+ "--text_column_name=text",
37
+ "--train_split_name=train.100",
38
+ "--wandb_project=flax-wav2vec2-2-bart-large-cnn",
39
+ "--warmup_steps=500"
40
+ ],
41
+ "state": "running",
42
+ "program": "run_flax_speech_recognition_seq2seq.py",
43
+ "codePath": "run_flax_speech_recognition_seq2seq.py",
44
+ "git": {
45
+ "remote": "https://huggingface.co/sanchit-gandhi/flax-wav2vec2-2-bart-large-cnn",
46
+ "commit": "c09faa6b18ac3f19251c414100acf1bad11c1b16"
47
+ },
48
+ "email": "sanchit@huggingface.co",
49
+ "root": "/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn",
50
+ "host": "t1v-n-4eb331dd-w-0",
51
+ "username": "sanchitgandhi",
52
+ "executable": "/home/sanchitgandhi/venv/bin/python3"
53
+ }
wandb/run-20220404_180947-p4wqexfj/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 28}}
wandb/run-20220404_180947-p4wqexfj/logs/debug-internal.log ADDED
@@ -0,0 +1,148 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-04-04 18:09:48,182 INFO MainThread:1240464 [internal.py:wandb_internal():92] W&B internal server running at pid: 1240464, started at: 2022-04-04 18:09:48.182168
2
+ 2022-04-04 18:09:48,184 INFO WriterThread:1240464 [datastore.py:open_for_write():77] open: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/run-p4wqexfj.wandb
3
+ 2022-04-04 18:09:48,185 DEBUG SenderThread:1240464 [sender.py:send():235] send: header
4
+ 2022-04-04 18:09:48,186 DEBUG SenderThread:1240464 [sender.py:send():235] send: run
5
+ 2022-04-04 18:09:48,316 INFO SenderThread:1240464 [dir_watcher.py:__init__():169] watching files in: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files
6
+ 2022-04-04 18:09:48,316 INFO SenderThread:1240464 [sender.py:_start_run_threads():812] run started: p4wqexfj with start time 1649095787
7
+ 2022-04-04 18:09:48,316 DEBUG SenderThread:1240464 [sender.py:send():235] send: summary
8
+ 2022-04-04 18:09:48,317 INFO SenderThread:1240464 [sender.py:_save_file():947] saving file wandb-summary.json with policy end
9
+ 2022-04-04 18:09:48,320 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: check_version
10
+ 2022-04-04 18:09:48,320 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: check_version
11
+ 2022-04-04 18:09:48,381 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: run_start
12
+ 2022-04-04 18:09:49,318 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/wandb-summary.json
13
+ 2022-04-04 18:09:50,590 DEBUG HandlerThread:1240464 [meta.py:__init__():37] meta init
14
+ 2022-04-04 18:09:50,590 DEBUG HandlerThread:1240464 [meta.py:__init__():51] meta init done
15
+ 2022-04-04 18:09:50,590 DEBUG HandlerThread:1240464 [meta.py:probe():211] probe
16
+ 2022-04-04 18:09:50,591 DEBUG HandlerThread:1240464 [meta.py:_setup_git():201] setup git
17
+ 2022-04-04 18:09:50,622 DEBUG HandlerThread:1240464 [meta.py:_setup_git():208] setup git done
18
+ 2022-04-04 18:09:50,623 DEBUG HandlerThread:1240464 [meta.py:_save_pip():55] save pip
19
+ 2022-04-04 18:09:50,623 DEBUG HandlerThread:1240464 [meta.py:_save_pip():69] save pip done
20
+ 2022-04-04 18:09:50,623 DEBUG HandlerThread:1240464 [meta.py:probe():249] probe done
21
+ 2022-04-04 18:09:50,626 DEBUG SenderThread:1240464 [sender.py:send():235] send: files
22
+ 2022-04-04 18:09:50,627 INFO SenderThread:1240464 [sender.py:_save_file():947] saving file wandb-metadata.json with policy now
23
+ 2022-04-04 18:09:50,633 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: stop_status
24
+ 2022-04-04 18:09:50,633 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: stop_status
25
+ 2022-04-04 18:09:50,664 DEBUG SenderThread:1240464 [sender.py:send():235] send: telemetry
26
+ 2022-04-04 18:09:50,908 INFO Thread-11 :1240464 [upload_job.py:push():137] Uploaded file /tmp/tmp1c0a214pwandb/2czfbyut-wandb-metadata.json
27
+ 2022-04-04 18:09:51,331 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/wandb-metadata.json
28
+ 2022-04-04 18:09:51,331 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log
29
+ 2022-04-04 18:09:51,332 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/requirements.txt
30
+ 2022-04-04 18:09:53,332 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log
31
+ 2022-04-04 18:10:05,666 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: stop_status
32
+ 2022-04-04 18:10:05,667 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: stop_status
33
+ 2022-04-04 18:10:09,338 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log
34
+ 2022-04-04 18:10:11,339 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log
35
+ 2022-04-04 18:10:13,340 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log
36
+ 2022-04-04 18:10:14,824 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: partial_history
37
+ 2022-04-04 18:10:14,826 DEBUG SenderThread:1240464 [sender.py:send():235] send: telemetry
38
+ 2022-04-04 18:10:15,342 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log
39
+ 2022-04-04 18:10:16,342 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log
40
+ 2022-04-04 18:10:17,066 DEBUG SenderThread:1240464 [sender.py:send():235] send: exit
41
+ 2022-04-04 18:10:17,066 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
42
+ 2022-04-04 18:10:17,066 INFO SenderThread:1240464 [sender.py:send_exit():371] handling exit code: 1
43
+ 2022-04-04 18:10:17,067 INFO SenderThread:1240464 [sender.py:send_exit():373] handling runtime: 28
44
+ 2022-04-04 18:10:17,067 INFO SenderThread:1240464 [sender.py:_save_file():947] saving file wandb-summary.json with policy end
45
+ 2022-04-04 18:10:17,067 INFO SenderThread:1240464 [sender.py:send_exit():379] send defer
46
+ 2022-04-04 18:10:17,067 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
47
+ 2022-04-04 18:10:17,068 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
48
+ 2022-04-04 18:10:17,068 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 0
49
+ 2022-04-04 18:10:17,068 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
50
+ 2022-04-04 18:10:17,068 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 0
51
+ 2022-04-04 18:10:17,068 INFO SenderThread:1240464 [sender.py:transition_state():392] send defer: 1
52
+ 2022-04-04 18:10:17,069 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
53
+ 2022-04-04 18:10:17,069 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 1
54
+ 2022-04-04 18:10:17,169 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
55
+ 2022-04-04 18:10:17,170 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
56
+ 2022-04-04 18:10:17,170 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 1
57
+ 2022-04-04 18:10:17,170 INFO SenderThread:1240464 [sender.py:transition_state():392] send defer: 2
58
+ 2022-04-04 18:10:17,170 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
59
+ 2022-04-04 18:10:17,170 DEBUG SenderThread:1240464 [sender.py:send():235] send: stats
60
+ 2022-04-04 18:10:17,171 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
61
+ 2022-04-04 18:10:17,171 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 2
62
+ 2022-04-04 18:10:17,171 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
63
+ 2022-04-04 18:10:17,171 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 2
64
+ 2022-04-04 18:10:17,171 INFO SenderThread:1240464 [sender.py:transition_state():392] send defer: 3
65
+ 2022-04-04 18:10:17,171 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
66
+ 2022-04-04 18:10:17,171 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 3
67
+ 2022-04-04 18:10:17,172 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
68
+ 2022-04-04 18:10:17,172 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 3
69
+ 2022-04-04 18:10:17,172 INFO SenderThread:1240464 [sender.py:transition_state():392] send defer: 4
70
+ 2022-04-04 18:10:17,172 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
71
+ 2022-04-04 18:10:17,172 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 4
72
+ 2022-04-04 18:10:17,172 DEBUG SenderThread:1240464 [sender.py:send():235] send: summary
73
+ 2022-04-04 18:10:17,172 INFO SenderThread:1240464 [sender.py:_save_file():947] saving file wandb-summary.json with policy end
74
+ 2022-04-04 18:10:17,172 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
75
+ 2022-04-04 18:10:17,172 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 4
76
+ 2022-04-04 18:10:17,172 INFO SenderThread:1240464 [sender.py:transition_state():392] send defer: 5
77
+ 2022-04-04 18:10:17,173 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
78
+ 2022-04-04 18:10:17,173 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 5
79
+ 2022-04-04 18:10:17,173 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
80
+ 2022-04-04 18:10:17,173 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 5
81
+ 2022-04-04 18:10:17,242 INFO SenderThread:1240464 [sender.py:transition_state():392] send defer: 6
82
+ 2022-04-04 18:10:17,243 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
83
+ 2022-04-04 18:10:17,243 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 6
84
+ 2022-04-04 18:10:17,243 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
85
+ 2022-04-04 18:10:17,243 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 6
86
+ 2022-04-04 18:10:17,243 INFO SenderThread:1240464 [dir_watcher.py:finish():283] shutting down directory watcher
87
+ 2022-04-04 18:10:17,272 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
88
+ 2022-04-04 18:10:17,342 INFO Thread-7 :1240464 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log
89
+ 2022-04-04 18:10:17,343 INFO SenderThread:1240464 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/config.yaml
90
+ 2022-04-04 18:10:17,343 INFO SenderThread:1240464 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/wandb-summary.json
91
+ 2022-04-04 18:10:17,343 INFO SenderThread:1240464 [dir_watcher.py:finish():313] scan: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files
92
+ 2022-04-04 18:10:17,343 INFO SenderThread:1240464 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/requirements.txt requirements.txt
93
+ 2022-04-04 18:10:17,343 INFO SenderThread:1240464 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log output.log
94
+ 2022-04-04 18:10:17,344 INFO SenderThread:1240464 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/wandb-metadata.json wandb-metadata.json
95
+ 2022-04-04 18:10:17,344 INFO SenderThread:1240464 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/config.yaml config.yaml
96
+ 2022-04-04 18:10:17,344 INFO SenderThread:1240464 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/wandb-summary.json wandb-summary.json
97
+ 2022-04-04 18:10:17,347 INFO SenderThread:1240464 [sender.py:transition_state():392] send defer: 7
98
+ 2022-04-04 18:10:17,347 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
99
+ 2022-04-04 18:10:17,348 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
100
+ 2022-04-04 18:10:17,348 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 7
101
+ 2022-04-04 18:10:17,348 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
102
+ 2022-04-04 18:10:17,348 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 7
103
+ 2022-04-04 18:10:17,348 INFO SenderThread:1240464 [file_pusher.py:finish():145] shutting down file pusher
104
+ 2022-04-04 18:10:17,449 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
105
+ 2022-04-04 18:10:17,449 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
106
+ 2022-04-04 18:10:17,551 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
107
+ 2022-04-04 18:10:17,551 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
108
+ 2022-04-04 18:10:17,621 INFO Thread-12 :1240464 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/requirements.txt
109
+ 2022-04-04 18:10:17,644 INFO Thread-14 :1240464 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/config.yaml
110
+ 2022-04-04 18:10:17,653 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
111
+ 2022-04-04 18:10:17,653 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
112
+ 2022-04-04 18:10:17,662 INFO Thread-13 :1240464 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/output.log
113
+ 2022-04-04 18:10:17,694 INFO Thread-15 :1240464 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/files/wandb-summary.json
114
+ 2022-04-04 18:10:17,754 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
115
+ 2022-04-04 18:10:17,755 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
116
+ 2022-04-04 18:10:17,856 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
117
+ 2022-04-04 18:10:17,856 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
118
+ 2022-04-04 18:10:17,894 INFO Thread-6 :1240464 [sender.py:transition_state():392] send defer: 8
119
+ 2022-04-04 18:10:17,894 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
120
+ 2022-04-04 18:10:17,895 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 8
121
+ 2022-04-04 18:10:17,895 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
122
+ 2022-04-04 18:10:17,895 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 8
123
+ 2022-04-04 18:10:17,958 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
124
+ 2022-04-04 18:10:18,009 INFO SenderThread:1240464 [sender.py:transition_state():392] send defer: 9
125
+ 2022-04-04 18:10:18,009 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
126
+ 2022-04-04 18:10:18,009 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
127
+ 2022-04-04 18:10:18,009 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 9
128
+ 2022-04-04 18:10:18,010 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
129
+ 2022-04-04 18:10:18,010 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 9
130
+ 2022-04-04 18:10:18,010 INFO SenderThread:1240464 [sender.py:transition_state():392] send defer: 10
131
+ 2022-04-04 18:10:18,010 DEBUG SenderThread:1240464 [sender.py:send():235] send: final
132
+ 2022-04-04 18:10:18,010 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: defer
133
+ 2022-04-04 18:10:18,010 DEBUG SenderThread:1240464 [sender.py:send():235] send: footer
134
+ 2022-04-04 18:10:18,010 INFO HandlerThread:1240464 [handler.py:handle_request_defer():164] handle defer: 10
135
+ 2022-04-04 18:10:18,011 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: defer
136
+ 2022-04-04 18:10:18,011 INFO SenderThread:1240464 [sender.py:send_request_defer():388] handle sender defer: 10
137
+ 2022-04-04 18:10:18,110 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: poll_exit
138
+ 2022-04-04 18:10:18,111 DEBUG SenderThread:1240464 [sender.py:send_request():249] send_request: poll_exit
139
+ 2022-04-04 18:10:18,111 INFO SenderThread:1240464 [file_pusher.py:join():150] waiting for file pusher
140
+ 2022-04-04 18:10:18,262 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: sampled_history
141
+ 2022-04-04 18:10:18,263 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: get_summary
142
+ 2022-04-04 18:10:18,264 DEBUG HandlerThread:1240464 [handler.py:handle_request():141] handle_request: shutdown
143
+ 2022-04-04 18:10:18,264 INFO HandlerThread:1240464 [handler.py:finish():778] shutting down handler
144
+ 2022-04-04 18:10:19,011 INFO WriterThread:1240464 [datastore.py:close():281] close: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/run-p4wqexfj.wandb
145
+ 2022-04-04 18:10:19,161 INFO SenderThread:1240464 [sender.py:finish():1078] shutting down sender
146
+ 2022-04-04 18:10:19,161 INFO SenderThread:1240464 [file_pusher.py:finish():145] shutting down file pusher
147
+ 2022-04-04 18:10:19,161 INFO SenderThread:1240464 [file_pusher.py:join():150] waiting for file pusher
148
+ 2022-04-04 18:10:19,249 INFO MainThread:1240464 [internal.py:handle_exit():82] Internal process exited
wandb/run-20220404_180947-p4wqexfj/logs/debug.log ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-04-04 18:09:47,396 INFO MainThread:1239999 [wandb_setup.py:_flush():75] Loading settings from /home/sanchitgandhi/.config/wandb/settings
2
+ 2022-04-04 18:09:47,396 INFO MainThread:1239999 [wandb_setup.py:_flush():75] Loading settings from wandb/settings
3
+ 2022-04-04 18:09:47,396 INFO MainThread:1239999 [wandb_setup.py:_flush():75] Loading settings from environment variables: {'entity': 'sanchit-gandhi', 'project': 'flax-wav2vec2-2-bart-large-cnn', 'sweep_id': 'a50sny4c', 'root_dir': '/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn', 'run_id': 'p4wqexfj', 'sweep_param_path': '/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/sweep-a50sny4c/config-p4wqexfj.yaml'}
4
+ 2022-04-04 18:09:47,396 INFO MainThread:1239999 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_flax_speech_recognition_seq2seq.py', 'program': 'run_flax_speech_recognition_seq2seq.py'}
5
+ 2022-04-04 18:09:47,397 INFO MainThread:1239999 [wandb_init.py:_log_setup():405] Logging user logs to /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/logs/debug.log
6
+ 2022-04-04 18:09:47,397 INFO MainThread:1239999 [wandb_init.py:_log_setup():406] Logging internal logs to /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_180947-p4wqexfj/logs/debug-internal.log
7
+ 2022-04-04 18:09:47,397 INFO MainThread:1239999 [wandb_init.py:init():439] calling init triggers
8
+ 2022-04-04 18:09:47,397 INFO MainThread:1239999 [wandb_init.py:init():442] wandb.init called with sweep_config: {'dataset_cache_dir': '/home/sanchitgandhi/cache/huggingface/datasets', 'dataset_config_name': 'clean', 'dataset_name': 'librispeech_asr', 'eval_split_name': 'validation', 'generation_max_length': 40, 'generation_num_beams': 1, 'gradient_accumulation_steps': 1, 'learning_rate': 2.565346074198426e-05, 'length_column_name': 'input_length', 'logging_steps': 1, 'matmul_precision': 'highest', 'max_duration_in_seconds': 15, 'max_target_length': 64, 'min_duration_in_seconds': 15, 'model_name_or_path': './', 'num_train_epochs': 5, 'output_dir': './', 'per_device_eval_batch_size': 2, 'per_device_train_batch_size': 2, 'preprocessing_num_workers': 16, 'text_column_name': 'text', 'train_split_name': 'train.100', 'wandb_project': 'flax-wav2vec2-2-bart-large-cnn', 'warmup_steps': 500}
9
+ config: {}
10
+ 2022-04-04 18:09:47,397 INFO MainThread:1239999 [wandb_init.py:init():492] starting backend
11
+ 2022-04-04 18:09:47,397 INFO MainThread:1239999 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-04-04 18:09:47,411 INFO MainThread:1239999 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-04-04 18:09:47,420 INFO MainThread:1239999 [backend.py:ensure_launched():224] started backend process with pid: 1240464
14
+ 2022-04-04 18:09:47,422 INFO MainThread:1239999 [wandb_init.py:init():501] backend started and connected
15
+ 2022-04-04 18:09:47,429 INFO MainThread:1239999 [wandb_run.py:_config_callback():992] config_cb None None {'dataset_cache_dir': '/home/sanchitgandhi/cache/huggingface/datasets', 'dataset_config_name': 'clean', 'dataset_name': 'librispeech_asr', 'eval_split_name': 'validation', 'generation_max_length': 40, 'generation_num_beams': 1, 'gradient_accumulation_steps': 1, 'learning_rate': 2.565346074198426e-05, 'length_column_name': 'input_length', 'logging_steps': 1, 'matmul_precision': 'highest', 'max_duration_in_seconds': 15, 'max_target_length': 64, 'min_duration_in_seconds': 15, 'model_name_or_path': './', 'num_train_epochs': 5, 'output_dir': './', 'per_device_eval_batch_size': 2, 'per_device_train_batch_size': 2, 'preprocessing_num_workers': 16, 'text_column_name': 'text', 'train_split_name': 'train.100', 'wandb_project': 'flax-wav2vec2-2-bart-large-cnn', 'warmup_steps': 500}
16
+ 2022-04-04 18:09:47,433 INFO MainThread:1239999 [wandb_init.py:init():565] updated telemetry
17
+ 2022-04-04 18:09:47,454 INFO MainThread:1239999 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
18
+ 2022-04-04 18:09:48,319 INFO MainThread:1239999 [wandb_run.py:_on_init():1759] communicating current version
19
+ 2022-04-04 18:09:48,379 INFO MainThread:1239999 [wandb_run.py:_on_init():1763] got version response
20
+ 2022-04-04 18:09:48,379 INFO MainThread:1239999 [wandb_init.py:init():625] starting run threads in backend
21
+ 2022-04-04 18:09:50,633 INFO MainThread:1239999 [wandb_run.py:_console_start():1733] atexit reg
22
+ 2022-04-04 18:09:50,633 INFO MainThread:1239999 [wandb_run.py:_redirect():1606] redirect: SettingsConsole.REDIRECT
23
+ 2022-04-04 18:09:50,634 INFO MainThread:1239999 [wandb_run.py:_redirect():1611] Redirecting console.
24
+ 2022-04-04 18:09:50,636 INFO MainThread:1239999 [wandb_run.py:_redirect():1667] Redirects installed.
25
+ 2022-04-04 18:09:50,636 INFO MainThread:1239999 [wandb_init.py:init():664] run started, returning control to user process
26
+ 2022-04-04 18:10:14,814 INFO MainThread:1239999 [wandb_run.py:_atexit_cleanup():1702] got exitcode: 1
27
+ 2022-04-04 18:10:14,824 INFO MainThread:1239999 [wandb_run.py:_restore():1674] restore
28
+ 2022-04-04 18:10:17,068 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
29
+ wandb_count: 1
30
+ }
31
+ pusher_stats {
32
+ uploaded_bytes: 1948
33
+ total_bytes: 1948
34
+ }
35
+
36
+ 2022-04-04 18:10:17,171 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
37
+ wandb_count: 1
38
+ }
39
+ pusher_stats {
40
+ uploaded_bytes: 1948
41
+ total_bytes: 1948
42
+ }
43
+
44
+ 2022-04-04 18:10:17,348 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
45
+ wandb_count: 2
46
+ }
47
+ pusher_stats {
48
+ uploaded_bytes: 1948
49
+ total_bytes: 5092
50
+ }
51
+
52
+ 2022-04-04 18:10:17,450 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
53
+ wandb_count: 5
54
+ }
55
+ pusher_stats {
56
+ uploaded_bytes: 43815
57
+ total_bytes: 43842
58
+ }
59
+
60
+ 2022-04-04 18:10:17,552 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
61
+ wandb_count: 5
62
+ }
63
+ pusher_stats {
64
+ uploaded_bytes: 43842
65
+ total_bytes: 43842
66
+ }
67
+
68
+ 2022-04-04 18:10:17,654 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
69
+ wandb_count: 5
70
+ }
71
+ pusher_stats {
72
+ uploaded_bytes: 43842
73
+ total_bytes: 43842
74
+ }
75
+
76
+ 2022-04-04 18:10:17,755 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
77
+ wandb_count: 5
78
+ }
79
+ pusher_stats {
80
+ uploaded_bytes: 43842
81
+ total_bytes: 43842
82
+ }
83
+
84
+ 2022-04-04 18:10:17,857 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
85
+ wandb_count: 5
86
+ }
87
+ pusher_stats {
88
+ uploaded_bytes: 43842
89
+ total_bytes: 43842
90
+ }
91
+
92
+ 2022-04-04 18:10:18,010 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
93
+ wandb_count: 5
94
+ }
95
+ pusher_stats {
96
+ uploaded_bytes: 43842
97
+ total_bytes: 43842
98
+ }
99
+
100
+ 2022-04-04 18:10:18,161 INFO MainThread:1239999 [wandb_run.py:_on_finish():1831] got exit ret: done: true
101
+ exit_result {
102
+ }
103
+ file_counts {
104
+ wandb_count: 5
105
+ }
106
+ pusher_stats {
107
+ uploaded_bytes: 43842
108
+ total_bytes: 43842
109
+ }
110
+ local_info {
111
+ }
112
+
113
+ 2022-04-04 18:10:19,745 INFO MainThread:1239999 [wandb_run.py:_footer_history_summary_info():2865] rendering history
114
+ 2022-04-04 18:10:19,745 INFO MainThread:1239999 [wandb_run.py:_footer_history_summary_info():2894] rendering summary
115
+ 2022-04-04 18:10:19,748 INFO MainThread:1239999 [wandb_run.py:_footer_sync_info():2822] logging synced files
wandb/run-20220404_180947-p4wqexfj/run-p4wqexfj.wandb ADDED
Binary file (39 kB). View file
 
wandb/run-20220404_181027-mgg9caus/files/config.yaml ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ cli_version: 0.12.11
7
+ framework: huggingface
8
+ huggingface_version: 4.18.0.dev0
9
+ is_jupyter_run: false
10
+ is_kaggle_kernel: false
11
+ python_version: 3.8.10
12
+ start_time: 1649095827
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 11
19
+ - 12
20
+ - 45
21
+ - 49
22
+ - 51
23
+ 2:
24
+ - 1
25
+ - 2
26
+ - 3
27
+ - 11
28
+ - 12
29
+ - 45
30
+ - 49
31
+ - 51
32
+ 4: 3.8.10
33
+ 5: 0.12.11
34
+ 6: 4.18.0.dev0
35
+ 8:
36
+ - 5
37
+ dataset_cache_dir:
38
+ desc: null
39
+ value: /home/sanchitgandhi/cache/huggingface/datasets
40
+ dataset_config_name:
41
+ desc: null
42
+ value: clean
43
+ dataset_name:
44
+ desc: null
45
+ value: librispeech_asr
46
+ eval_split_name:
47
+ desc: null
48
+ value: validation
49
+ generation_max_length:
50
+ desc: null
51
+ value: 40
52
+ generation_num_beams:
53
+ desc: null
54
+ value: 1
55
+ gradient_accumulation_steps:
56
+ desc: null
57
+ value: 1
58
+ learning_rate:
59
+ desc: null
60
+ value: 0.0006871268347239357
61
+ length_column_name:
62
+ desc: null
63
+ value: input_length
64
+ logging_steps:
65
+ desc: null
66
+ value: 1
67
+ matmul_precision:
68
+ desc: null
69
+ value: highest
70
+ max_duration_in_seconds:
71
+ desc: null
72
+ value: 15
73
+ max_target_length:
74
+ desc: null
75
+ value: 64
76
+ min_duration_in_seconds:
77
+ desc: null
78
+ value: 15
79
+ model_name_or_path:
80
+ desc: null
81
+ value: ./
82
+ num_train_epochs:
83
+ desc: null
84
+ value: 5
85
+ output_dir:
86
+ desc: null
87
+ value: ./
88
+ per_device_eval_batch_size:
89
+ desc: null
90
+ value: 2
91
+ per_device_train_batch_size:
92
+ desc: null
93
+ value: 2
94
+ preprocessing_num_workers:
95
+ desc: null
96
+ value: 16
97
+ text_column_name:
98
+ desc: null
99
+ value: text
100
+ train_split_name:
101
+ desc: null
102
+ value: train.100
103
+ wandb_project:
104
+ desc: null
105
+ value: flax-wav2vec2-2-bart-large-cnn
106
+ warmup_steps:
107
+ desc: null
108
+ value: 500
wandb/run-20220404_181027-mgg9caus/files/output.log ADDED
@@ -0,0 +1,793 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ loading configuration file ./config.json
2
+ Model config SpeechEncoderDecoderConfig {
3
+ "_name_or_path": "./",
4
+ "architectures": [
5
+ "SpeechEncoderDecoderModel"
6
+ ],
7
+ "decoder": {
8
+ "_name_or_path": "facebook/bart-large-cnn",
9
+ "_num_labels": 3,
10
+ "activation_dropout": 0.0,
11
+ "activation_function": "gelu",
12
+ "add_cross_attention": true,
13
+ "add_final_layer_norm": false,
14
+ "architectures": [
15
+ "BartForConditionalGeneration"
16
+ ],
17
+ "attention_dropout": 0.0,
18
+ "bad_words_ids": null,
19
+ "bos_token_id": 0,
20
+ "chunk_size_feed_forward": 0,
21
+ "classif_dropout": 0.0,
22
+ "classifier_dropout": 0.0,
23
+ "cross_attention_hidden_size": null,
24
+ "d_model": 1024,
25
+ "decoder_attention_heads": 16,
26
+ "decoder_ffn_dim": 4096,
27
+ "decoder_layerdrop": 0.0,
28
+ "decoder_layers": 12,
29
+ "decoder_start_token_id": 2,
30
+ "diversity_penalty": 0.0,
31
+ "do_sample": false,
32
+ "dropout": 0.1,
33
+ "early_stopping": true,
34
+ "encoder_attention_heads": 16,
35
+ "encoder_ffn_dim": 4096,
36
+ "encoder_layerdrop": 0.0,
37
+ "encoder_layers": 12,
38
+ "encoder_no_repeat_ngram_size": 0,
39
+ "eos_token_id": 2,
40
+ "exponential_decay_length_penalty": null,
41
+ "finetuning_task": null,
42
+ "force_bos_token_to_be_generated": true,
43
+ "forced_bos_token_id": 0,
44
+ "forced_eos_token_id": 2,
45
+ "gradient_checkpointing": false,
46
+ "id2label": {
47
+ "0": "LABEL_0",
48
+ "1": "LABEL_1",
49
+ "2": "LABEL_2"
50
+ },
51
+ "init_std": 0.02,
52
+ "is_decoder": true,
53
+ "is_encoder_decoder": false,
54
+ "label2id": {
55
+ "LABEL_0": 0,
56
+ "LABEL_1": 1,
57
+ "LABEL_2": 2
58
+ },
59
+ "length_penalty": 2.0,
60
+ "max_length": 142,
61
+ "max_position_embeddings": 1024,
62
+ "min_length": 56,
63
+ "model_type": "bart",
64
+ "no_repeat_ngram_size": 3,
65
+ "normalize_before": false,
66
+ "num_beam_groups": 1,
67
+ "num_beams": 4,
68
+ "num_hidden_layers": 12,
69
+ "num_return_sequences": 1,
70
+ "output_attentions": false,
71
+ "output_hidden_states": false,
72
+ "output_past": true,
73
+ "output_scores": false,
74
+ "pad_token_id": 1,
75
+ "prefix": " ",
76
+ "problem_type": null,
77
+ "pruned_heads": {},
78
+ "remove_invalid_values": false,
79
+ "repetition_penalty": 1.0,
80
+ "return_dict": true,
81
+ "return_dict_in_generate": false,
82
+ "scale_embedding": false,
83
+ "sep_token_id": null,
84
+ "task_specific_params": {
85
+ "summarization": {
86
+ "early_stopping": true,
87
+ "length_penalty": 2.0,
88
+ "max_length": 142,
89
+ "min_length": 56,
90
+ "no_repeat_ngram_size": 3,
91
+ "num_beams": 4
92
+ }
93
+ },
94
+ "temperature": 1.0,
95
+ "tie_encoder_decoder": false,
96
+ "tie_word_embeddings": true,
97
+ "tokenizer_class": null,
98
+ "top_k": 50,
99
+ "top_p": 1.0,
100
+ "torch_dtype": null,
101
+ "torchscript": false,
102
+ "transformers_version": "4.18.0.dev0",
103
+ "typical_p": 1.0,
104
+ "use_bfloat16": false,
105
+ "use_cache": true,
106
+ "vocab_size": 50264
107
+ },
108
+ "decoder_start_token_id": 0,
109
+ "encoder": {
110
+ "_name_or_path": "facebook/wav2vec2-large-lv60",
111
+ "activation_dropout": 0.1,
112
+ "adapter_kernel_size": 3,
113
+ "adapter_stride": 2,
114
+ "add_adapter": true,
115
+ "add_cross_attention": false,
116
+ "apply_spec_augment": true,
117
+ "architectures": [
118
+ "Wav2Vec2ForPreTraining"
119
+ ],
120
+ "attention_dropout": 0.1,
121
+ "bad_words_ids": null,
122
+ "bos_token_id": 1,
123
+ "chunk_size_feed_forward": 0,
124
+ "classifier_proj_size": 256,
125
+ "codevector_dim": 768,
126
+ "contrastive_logits_temperature": 0.1,
127
+ "conv_bias": true,
128
+ "conv_dim": [
129
+ 512,
130
+ 512,
131
+ 512,
132
+ 512,
133
+ 512,
134
+ 512,
135
+ 512
136
+ ],
137
+ "conv_kernel": [
138
+ 10,
139
+ 3,
140
+ 3,
141
+ 3,
142
+ 3,
143
+ 2,
144
+ 2
145
+ ],
146
+ "conv_stride": [
147
+ 5,
148
+ 2,
149
+ 2,
150
+ 2,
151
+ 2,
152
+ 2,
153
+ 2
154
+ ],
155
+ "cross_attention_hidden_size": null,
156
+ "ctc_loss_reduction": "sum",
157
+ "ctc_zero_infinity": false,
158
+ "decoder_start_token_id": null,
159
+ "diversity_loss_weight": 0.1,
160
+ "diversity_penalty": 0.0,
161
+ "do_sample": false,
162
+ "do_stable_layer_norm": true,
163
+ "early_stopping": false,
164
+ "encoder_no_repeat_ngram_size": 0,
165
+ "eos_token_id": 2,
166
+ "exponential_decay_length_penalty": null,
167
+ "feat_extract_activation": "gelu",
168
+ "feat_extract_dropout": 0.0,
169
+ "feat_extract_norm": "layer",
170
+ "feat_proj_dropout": 0.0,
171
+ "feat_quantizer_dropout": 0.0,
172
+ "final_dropout": 0.0,
173
+ "finetuning_task": null,
174
+ "forced_bos_token_id": null,
175
+ "forced_eos_token_id": null,
176
+ "gradient_checkpointing": false,
177
+ "hidden_act": "gelu",
178
+ "hidden_dropout": 0.1,
179
+ "hidden_dropout_prob": 0.1,
180
+ "hidden_size": 1024,
181
+ "id2label": {
182
+ "0": "LABEL_0",
183
+ "1": "LABEL_1"
184
+ },
185
+ "initializer_range": 0.02,
186
+ "intermediate_size": 4096,
187
+ "is_decoder": false,
188
+ "is_encoder_decoder": false,
189
+ "label2id": {
190
+ "LABEL_0": 0,
191
+ "LABEL_1": 1
192
+ },
193
+ "layer_norm_eps": 1e-05,
194
+ "layerdrop": 0.0,
195
+ "length_penalty": 1.0,
196
+ "mask_feature_length": 10,
197
+ "mask_feature_min_masks": 0,
198
+ "mask_feature_prob": 0.0,
199
+ "mask_time_length": 10,
200
+ "mask_time_min_masks": 2,
201
+ "mask_time_prob": 0.1,
202
+ "max_length": 20,
203
+ "min_length": 0,
204
+ "model_type": "wav2vec2",
205
+ "no_repeat_ngram_size": 0,
206
+ "num_adapter_layers": 3,
207
+ "num_attention_heads": 16,
208
+ "num_beam_groups": 1,
209
+ "num_beams": 1,
210
+ "num_codevector_groups": 2,
211
+ "num_codevectors_per_group": 320,
212
+ "num_conv_pos_embedding_groups": 16,
213
+ "num_conv_pos_embeddings": 128,
214
+ "num_feat_extract_layers": 7,
215
+ "num_hidden_layers": 24,
216
+ "num_negatives": 100,
217
+ "num_return_sequences": 1,
218
+ "output_attentions": false,
219
+ "output_hidden_size": 1024,
220
+ "output_hidden_states": false,
221
+ "output_scores": false,
222
+ "pad_token_id": 0,
223
+ "prefix": null,
224
+ "problem_type": null,
225
+ "proj_codevector_dim": 768,
226
+ "pruned_heads": {},
227
+ "remove_invalid_values": false,
228
+ "repetition_penalty": 1.0,
229
+ "return_dict": true,
230
+ "return_dict_in_generate": false,
231
+ "sep_token_id": null,
232
+ "task_specific_params": null,
233
+ "tdnn_dilation": [
234
+ 1,
235
+ 2,
236
+ 3,
237
+ 1,
238
+ 1
239
+ ],
240
+ "tdnn_dim": [
241
+ 512,
242
+ 512,
243
+ 512,
244
+ 512,
245
+ 1500
246
+ ],
247
+ "tdnn_kernel": [
248
+ 5,
249
+ 3,
250
+ 3,
251
+ 1,
252
+ 1
253
+ ],
254
+ "temperature": 1.0,
255
+ "tie_encoder_decoder": false,
256
+ "tie_word_embeddings": true,
257
+ "tokenizer_class": null,
258
+ "top_k": 50,
259
+ "top_p": 1.0,
260
+ "torch_dtype": null,
261
+ "torchscript": false,
262
+ "transformers_version": "4.18.0.dev0",
263
+ "typical_p": 1.0,
264
+ "use_bfloat16": false,
265
+ "use_weighted_layer_sum": false,
266
+ "vocab_size": 32,
267
+ "xvector_output_dim": 512
268
+ },
269
+ "eos_token_id": 2,
270
+ "is_encoder_decoder": true,
271
+ "max_length": 40,
272
+ "model_type": "speech-encoder-decoder",
273
+ "pad_token_id": 1,
274
+ "processor_class": "Wav2Vec2Processor",
275
+ "tie_word_embeddings": false,
276
+ "transformers_version": null,
277
+ "use_cache": false
278
+ }
279
+ loading feature extractor configuration file ./preprocessor_config.json
280
+ Feature extractor Wav2Vec2FeatureExtractor {
281
+ "do_normalize": true,
282
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
283
+ "feature_size": 1,
284
+ "padding_side": "right",
285
+ "padding_value": 0.0,
286
+ "return_attention_mask": true,
287
+ "sampling_rate": 16000
288
+ }
289
+ Didn't find file ./added_tokens.json. We won't load it.
290
+ loading file ./vocab.json
291
+ loading file ./merges.txt
292
+ loading file ./tokenizer.json
293
+ loading file None
294
+ loading file ./special_tokens_map.json
295
+ loading file ./tokenizer_config.json
296
+ loading weights file ./flax_model.msgpack
297
+ 04/04/2022 18:10:31 - INFO - __main__ - Training/evaluation parameters FlaxSeq2SeqTrainingArguments(
298
+ _n_gpu=0,
299
+ adafactor=False,
300
+ adam_beta1=0.9,
301
+ adam_beta2=0.999,
302
+ adam_epsilon=1e-08,
303
+ bf16=False,
304
+ bf16_full_eval=False,
305
+ data_seed=None,
306
+ dataloader_drop_last=False,
307
+ dataloader_num_workers=0,
308
+ dataloader_pin_memory=True,
309
+ ddp_bucket_cap_mb=None,
310
+ ddp_find_unused_parameters=None,
311
+ debug=[],
312
+ deepspeed=None,
313
+ disable_tqdm=False,
314
+ do_eval=True,
315
+ do_predict=False,
316
+ do_train=True,
317
+ eval_accumulation_steps=None,
318
+ eval_delay=0,
319
+ eval_steps=None,
320
+ evaluation_strategy=IntervalStrategy.NO,
321
+ fp16=False,
322
+ fp16_backend=auto,
323
+ fp16_full_eval=False,
324
+ fp16_opt_level=O1,
325
+ generation_max_length=40,
326
+ generation_num_beams=1,
327
+ gradient_accumulation_steps=1,
328
+ gradient_checkpointing=False,
329
+ greater_is_better=None,
330
+ group_by_length=False,
331
+ half_precision_backend=auto,
332
+ hub_model_id=None,
333
+ hub_strategy=HubStrategy.EVERY_SAVE,
334
+ hub_token=<HUB_TOKEN>,
335
+ ignore_data_skip=False,
336
+ label_names=None,
337
+ label_smoothing_factor=0.0,
338
+ learning_rate=0.0006871268347239357,
339
+ length_column_name=input_length,
340
+ load_best_model_at_end=False,
341
+ local_rank=-1,
342
+ log_level=-1,
343
+ log_level_replica=-1,
344
+ log_on_each_node=True,
345
+ logging_dir=./runs/Apr04_18-10-27_t1v-n-4eb331dd-w-0,
346
+ logging_first_step=False,
347
+ logging_nan_inf_filter=True,
348
+ logging_steps=1,
349
+ logging_strategy=IntervalStrategy.STEPS,
350
+ lr_scheduler_type=SchedulerType.LINEAR,
351
+ matmul_precision=highest,
352
+ max_grad_norm=1.0,
353
+ max_steps=-1,
354
+ metric_for_best_model=None,
355
+ mixed_precision=False,
356
+ mp_parameters=,
357
+ no_cuda=False,
358
+ num_train_epochs=5.0,
359
+ optim=OptimizerNames.ADAMW_HF,
360
+ output_dir=./,
361
+ overwrite_output_dir=True,
362
+ past_index=-1,
363
+ per_device_eval_batch_size=2,
364
+ per_device_train_batch_size=2,
365
+ predict_with_generate=True,
366
+ prediction_loss_only=False,
367
+ push_to_hub=False,
368
+ push_to_hub_model_id=None,
369
+ push_to_hub_organization=None,
370
+ push_to_hub_token=<PUSH_TO_HUB_TOKEN>,
371
+ remove_unused_columns=True,
372
+ report_to=['tensorboard', 'wandb'],
373
+ resume_from_checkpoint=None,
374
+ run_name=./,
375
+ save_on_each_node=False,
376
+ save_steps=500,
377
+ save_strategy=IntervalStrategy.STEPS,
378
+ save_total_limit=None,
379
+ seed=42,
380
+ sharded_ddp=[],
381
+ skip_memory_metrics=True,
382
+ sortish_sampler=False,
383
+ tf32=None,
384
+ tpu_metrics_debug=False,
385
+ tpu_num_cores=None,
386
+ use_legacy_prediction_loop=False,
387
+ warmup_ratio=0.0,
388
+ warmup_steps=500,
389
+ weight_decay=0.0,
390
+ xpu_backend=None,
391
+ )
392
+ 04/04/2022 18:10:31 - INFO - __main__ - JAX devices: 1, matmul precision: highest
393
+ 04/04/2022 18:10:31 - WARNING - datasets.builder - Reusing dataset librispeech_asr (/home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c)
394
+ 04/04/2022 18:10:31 - WARNING - datasets.builder - Reusing dataset librispeech_asr (/home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c)
395
+ tcmalloc: large alloc 2353643520 bytes == 0x9cd7c000 @ 0x7f28e53d2680 0x7f28e53f3824 0x5f8a01 0x648cf1 0x5c4676 0x4f290e 0x64f718 0x5048b3 0x56b1da 0x56939a 0x5f6a13 0x50af49 0x5f3547 0x56c8cd 0x56939a 0x50aaa0 0x56c28c 0x56939a 0x5f6a13 0x56b0ae 0x56939a 0x68d047 0x67e351 0x67e3cf 0x67e471 0x67e817 0x6b6fe2 0x6b736d 0x7f28e51e40b3 0x5fa5ce
396
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-eb8e77ddeb6c34dc.arrow
397
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e19eb46cb05f7b14.arrow
398
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-92685008e5c09846.arrow
399
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-0d478bd8b093039a.arrow
400
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-13aa6d940e98be09.arrow
401
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-2d24e62de045a0c8.arrow
402
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-c6e972d87654db26.arrow
403
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-4771f1765b989007.arrow
404
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ab9a2fefe02761db.arrow
405
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-6ce32d0e344f4c06.arrow
406
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-1e0f4ae48a3c8b80.arrow
407
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e6f5ab8655734a9f.arrow
408
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-be6a3a97adc5697a.arrow
409
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-239fcccc5b67f7a4.arrow
410
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ff31cd13241c8bf3.arrow
411
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e8bf216b19718031.arrow
412
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-03ee7ff6632282a6.arrow
413
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-bb1b7de42a389f66.arrow
414
+ All model checkpoint weights were used when initializing FlaxSpeechEncoderDecoderModel.
415
+ All the weights of FlaxSpeechEncoderDecoderModel were initialized from the model checkpoint at ./.
416
+ If your task is similar to the task the model of the checkpoint was trained on, you can already use FlaxSpeechEncoderDecoderModel for predictions without further training.
417
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ed92b3340af62815.arrow
418
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-12ce4b2b8350532d.arrow
419
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-75f20b5cb7b89a9d.arrow
420
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-69941cf3a70dde04.arrow
421
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-f4d0102728859e5f.arrow
422
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-1c0eaf46e759e04d.arrow
423
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-9f8c4dccc4fa8f17.arrow
424
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-14df38b848e61b52.arrow
425
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-19c662ecc5b20d66.arrow
426
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-3276de10dacb4963.arrow
427
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-111057e4500c4cc5.arrow
428
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-d1bf30028d638290.arrow
429
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cc50c08263e878fc.arrow
430
+ 04/04/2022 18:10:48 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-f1813a8ae49a67d6.arrow
431
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00000_of_00016.arrow
432
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00001_of_00016.arrow
433
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00002_of_00016.arrow
434
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00003_of_00016.arrow
435
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00004_of_00016.arrow
436
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00005_of_00016.arrow
437
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00006_of_00016.arrow
438
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00007_of_00016.arrow
439
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00008_of_00016.arrow
440
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00009_of_00016.arrow
441
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00010_of_00016.arrow
442
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00011_of_00016.arrow
443
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00012_of_00016.arrow
444
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00013_of_00016.arrow
445
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00014_of_00016.arrow
446
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00015_of_00016.arrow
447
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00000_of_00016.arrow
448
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00001_of_00016.arrow
449
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00002_of_00016.arrow
450
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00003_of_00016.arrow
451
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00004_of_00016.arrow
452
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00005_of_00016.arrow
453
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00006_of_00016.arrow
454
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00007_of_00016.arrow
455
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00008_of_00016.arrow
456
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00009_of_00016.arrow
457
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00010_of_00016.arrow
458
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00011_of_00016.arrow
459
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00012_of_00016.arrow
460
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00013_of_00016.arrow
461
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00014_of_00016.arrow
462
+ 04/04/2022 18:10:49 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00015_of_00016.arrow
463
+ 04/04/2022 18:10:50 - INFO - __main__ - ***** Running training *****
464
+ 04/04/2022 18:10:50 - INFO - __main__ - Num examples = 0
465
+ 04/04/2022 18:10:50 - INFO - __main__ - Num Epochs = 5
466
+ 04/04/2022 18:10:50 - INFO - __main__ - Instantaneous batch size per device = 2
467
+ 04/04/2022 18:10:50 - INFO - __main__ - Num gradient accumulation steps = 1
468
+ 04/04/2022 18:10:50 - INFO - __main__ - Total train batch size (w. parallel & distributed) = 2
469
+ 04/04/2022 18:10:50 - INFO - __main__ - Total optimization steps = 0
470
+ Feature extractor saved in ./preprocessor_config.json
471
+ tokenizer config file saved in ./tokenizer_config.json
472
+ Special tokens file saved in ./special_tokens_map.json
473
+ Configuration saved in ./config.json
474
+ loading feature extractor configuration file ./preprocessor_config.json
475
+ loading configuration file ./config.json
476
+ Model config SpeechEncoderDecoderConfig {
477
+ "_name_or_path": "./",
478
+ "architectures": [
479
+ "SpeechEncoderDecoderModel"
480
+ ],
481
+ "decoder": {
482
+ "_name_or_path": "facebook/bart-large-cnn",
483
+ "_num_labels": 3,
484
+ "activation_dropout": 0.0,
485
+ "activation_function": "gelu",
486
+ "add_cross_attention": true,
487
+ "add_final_layer_norm": false,
488
+ "architectures": [
489
+ "BartForConditionalGeneration"
490
+ ],
491
+ "attention_dropout": 0.0,
492
+ "bad_words_ids": null,
493
+ "bos_token_id": 0,
494
+ "chunk_size_feed_forward": 0,
495
+ "classif_dropout": 0.0,
496
+ "classifier_dropout": 0.0,
497
+ "cross_attention_hidden_size": null,
498
+ "d_model": 1024,
499
+ "decoder_attention_heads": 16,
500
+ "decoder_ffn_dim": 4096,
501
+ "decoder_layerdrop": 0.0,
502
+ "decoder_layers": 12,
503
+ "decoder_start_token_id": 2,
504
+ "diversity_penalty": 0.0,
505
+ "do_sample": false,
506
+ "dropout": 0.1,
507
+ "early_stopping": true,
508
+ "encoder_attention_heads": 16,
509
+ "encoder_ffn_dim": 4096,
510
+ "encoder_layerdrop": 0.0,
511
+ "encoder_layers": 12,
512
+ "encoder_no_repeat_ngram_size": 0,
513
+ "eos_token_id": 2,
514
+ "exponential_decay_length_penalty": null,
515
+ "finetuning_task": null,
516
+ "force_bos_token_to_be_generated": true,
517
+ "forced_bos_token_id": 0,
518
+ "forced_eos_token_id": 2,
519
+ "gradient_checkpointing": false,
520
+ "id2label": {
521
+ "0": "LABEL_0",
522
+ "1": "LABEL_1",
523
+ "2": "LABEL_2"
524
+ },
525
+ "init_std": 0.02,
526
+ "is_decoder": true,
527
+ "is_encoder_decoder": false,
528
+ "label2id": {
529
+ "LABEL_0": 0,
530
+ "LABEL_1": 1,
531
+ "LABEL_2": 2
532
+ },
533
+ "length_penalty": 2.0,
534
+ "max_length": 142,
535
+ "max_position_embeddings": 1024,
536
+ "min_length": 56,
537
+ "model_type": "bart",
538
+ "no_repeat_ngram_size": 3,
539
+ "normalize_before": false,
540
+ "num_beam_groups": 1,
541
+ "num_beams": 4,
542
+ "num_hidden_layers": 12,
543
+ "num_return_sequences": 1,
544
+ "output_attentions": false,
545
+ "output_hidden_states": false,
546
+ "output_past": true,
547
+ "output_scores": false,
548
+ "pad_token_id": 1,
549
+ "prefix": " ",
550
+ "problem_type": null,
551
+ "pruned_heads": {},
552
+ "remove_invalid_values": false,
553
+ "repetition_penalty": 1.0,
554
+ "return_dict": true,
555
+ "return_dict_in_generate": false,
556
+ "scale_embedding": false,
557
+ "sep_token_id": null,
558
+ "task_specific_params": {
559
+ "summarization": {
560
+ "early_stopping": true,
561
+ "length_penalty": 2.0,
562
+ "max_length": 142,
563
+ "min_length": 56,
564
+ "no_repeat_ngram_size": 3,
565
+ "num_beams": 4
566
+ }
567
+ },
568
+ "temperature": 1.0,
569
+ "tie_encoder_decoder": false,
570
+ "tie_word_embeddings": true,
571
+ "tokenizer_class": null,
572
+ "top_k": 50,
573
+ "top_p": 1.0,
574
+ "torch_dtype": null,
575
+ "torchscript": false,
576
+ "transformers_version": "4.18.0.dev0",
577
+ "typical_p": 1.0,
578
+ "use_bfloat16": false,
579
+ "use_cache": true,
580
+ "vocab_size": 50264
581
+ },
582
+ "decoder_start_token_id": 0,
583
+ "encoder": {
584
+ "_name_or_path": "facebook/wav2vec2-large-lv60",
585
+ "activation_dropout": 0.1,
586
+ "adapter_kernel_size": 3,
587
+ "adapter_stride": 2,
588
+ "add_adapter": true,
589
+ "add_cross_attention": false,
590
+ "apply_spec_augment": true,
591
+ "architectures": [
592
+ "Wav2Vec2ForPreTraining"
593
+ ],
594
+ "attention_dropout": 0.1,
595
+ "bad_words_ids": null,
596
+ "bos_token_id": 1,
597
+ "chunk_size_feed_forward": 0,
598
+ "classifier_proj_size": 256,
599
+ "codevector_dim": 768,
600
+ "contrastive_logits_temperature": 0.1,
601
+ "conv_bias": true,
602
+ "conv_dim": [
603
+ 512,
604
+ 512,
605
+ 512,
606
+ 512,
607
+ 512,
608
+ 512,
609
+ 512
610
+ ],
611
+ "conv_kernel": [
612
+ 10,
613
+ 3,
614
+ 3,
615
+ 3,
616
+ 3,
617
+ 2,
618
+ 2
619
+ ],
620
+ "conv_stride": [
621
+ 5,
622
+ 2,
623
+ 2,
624
+ 2,
625
+ 2,
626
+ 2,
627
+ 2
628
+ ],
629
+ "cross_attention_hidden_size": null,
630
+ "ctc_loss_reduction": "sum",
631
+ "ctc_zero_infinity": false,
632
+ "decoder_start_token_id": null,
633
+ "diversity_loss_weight": 0.1,
634
+ "diversity_penalty": 0.0,
635
+ "do_sample": false,
636
+ "do_stable_layer_norm": true,
637
+ "early_stopping": false,
638
+ "encoder_no_repeat_ngram_size": 0,
639
+ "eos_token_id": 2,
640
+ "exponential_decay_length_penalty": null,
641
+ "feat_extract_activation": "gelu",
642
+ "feat_extract_dropout": 0.0,
643
+ "feat_extract_norm": "layer",
644
+ "feat_proj_dropout": 0.0,
645
+ "feat_quantizer_dropout": 0.0,
646
+ "final_dropout": 0.0,
647
+ "finetuning_task": null,
648
+ "forced_bos_token_id": null,
649
+ "forced_eos_token_id": null,
650
+ "gradient_checkpointing": false,
651
+ "hidden_act": "gelu",
652
+ "hidden_dropout": 0.1,
653
+ "hidden_dropout_prob": 0.1,
654
+ "hidden_size": 1024,
655
+ "id2label": {
656
+ "0": "LABEL_0",
657
+ "1": "LABEL_1"
658
+ },
659
+ "initializer_range": 0.02,
660
+ "intermediate_size": 4096,
661
+ "is_decoder": false,
662
+ "is_encoder_decoder": false,
663
+ "label2id": {
664
+ "LABEL_0": 0,
665
+ "LABEL_1": 1
666
+ },
667
+ "layer_norm_eps": 1e-05,
668
+ "layerdrop": 0.0,
669
+ "length_penalty": 1.0,
670
+ "mask_feature_length": 10,
671
+ "mask_feature_min_masks": 0,
672
+ "mask_feature_prob": 0.0,
673
+ "mask_time_length": 10,
674
+ "mask_time_min_masks": 2,
675
+ "mask_time_prob": 0.1,
676
+ "max_length": 20,
677
+ "min_length": 0,
678
+ "model_type": "wav2vec2",
679
+ "no_repeat_ngram_size": 0,
680
+ "num_adapter_layers": 3,
681
+ "num_attention_heads": 16,
682
+ "num_beam_groups": 1,
683
+ "num_beams": 1,
684
+ "num_codevector_groups": 2,
685
+ "num_codevectors_per_group": 320,
686
+ "num_conv_pos_embedding_groups": 16,
687
+ "num_conv_pos_embeddings": 128,
688
+ "num_feat_extract_layers": 7,
689
+ "num_hidden_layers": 24,
690
+ "num_negatives": 100,
691
+ "num_return_sequences": 1,
692
+ "output_attentions": false,
693
+ "output_hidden_size": 1024,
694
+ "output_hidden_states": false,
695
+ "output_scores": false,
696
+ "pad_token_id": 0,
697
+ "prefix": null,
698
+ "problem_type": null,
699
+ "proj_codevector_dim": 768,
700
+ "pruned_heads": {},
701
+ "remove_invalid_values": false,
702
+ "repetition_penalty": 1.0,
703
+ "return_dict": true,
704
+ "return_dict_in_generate": false,
705
+ "sep_token_id": null,
706
+ "task_specific_params": null,
707
+ "tdnn_dilation": [
708
+ 1,
709
+ 2,
710
+ 3,
711
+ 1,
712
+ 1
713
+ ],
714
+ "tdnn_dim": [
715
+ 512,
716
+ 512,
717
+ 512,
718
+ 512,
719
+ 1500
720
+ ],
721
+ "tdnn_kernel": [
722
+ 5,
723
+ 3,
724
+ 3,
725
+ 1,
726
+ 1
727
+ ],
728
+ "temperature": 1.0,
729
+ "tie_encoder_decoder": false,
730
+ "tie_word_embeddings": true,
731
+ "tokenizer_class": null,
732
+ "top_k": 50,
733
+ "top_p": 1.0,
734
+ "torch_dtype": null,
735
+ "torchscript": false,
736
+ "transformers_version": "4.18.0.dev0",
737
+ "typical_p": 1.0,
738
+ "use_bfloat16": false,
739
+ "use_weighted_layer_sum": false,
740
+ "vocab_size": 32,
741
+ "xvector_output_dim": 512
742
+ },
743
+ "eos_token_id": 2,
744
+ "is_encoder_decoder": true,
745
+ "max_length": 40,
746
+ "model_type": "speech-encoder-decoder",
747
+ "pad_token_id": 1,
748
+ "processor_class": "Wav2Vec2Processor",
749
+ "tie_word_embeddings": false,
750
+ "transformers_version": null,
751
+ "use_cache": false
752
+ }
753
+ loading feature extractor configuration file ./preprocessor_config.json
754
+ Feature extractor Wav2Vec2FeatureExtractor {
755
+ "do_normalize": true,
756
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
757
+ "feature_size": 1,
758
+ "padding_side": "right",
759
+ "padding_value": 0.0,
760
+ "return_attention_mask": true,
761
+ "sampling_rate": 16000
762
+ }
763
+ Didn't find file ./added_tokens.json. We won't load it.
764
+ loading file ./vocab.json
765
+ loading file ./merges.txt
766
+ loading file ./tokenizer.json
767
+ loading file None
768
+ loading file ./special_tokens_map.json
769
+ loading file ./tokenizer_config.json
770
+ 2022-04-04 18:10:49.472551: W tensorflow/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libcuda.so.1'; dlerror: libcuda.so.1: cannot open shared object file: No such file or directory
771
+ 2022-04-04 18:10:49.472596: W tensorflow/stream_executor/cuda/cuda_driver.cc:269] failed call to cuInit: UNKNOWN ERROR (303)
772
+ Epoch ... (1/5): 0%| | 0/5 [00:00<?, ?it/s]
773
+ Traceback (most recent call last):
774
+ File "run_flax_speech_recognition_seq2seq.py", line 1133, in <module>
775
+ main()
776
+ File "run_flax_speech_recognition_seq2seq.py", line 1050, in main
777
+ train_samples_idx = get_grouped_indices(vectorized_datasets["train"], batch_size_per_update, input_rng)
778
+ File "run_flax_speech_recognition_seq2seq.py", line 453, in get_grouped_indices
779
+ lengths = dataset["input_length"]
780
+ File "/home/sanchitgandhi/datasets/src/datasets/arrow_dataset.py", line 1765, in __getitem__
781
+ return self._getitem(
782
+ File "/home/sanchitgandhi/datasets/src/datasets/arrow_dataset.py", line 1749, in _getitem
783
+ pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None)
784
+ File "/home/sanchitgandhi/datasets/src/datasets/formatting/formatting.py", line 491, in query_table
785
+ pa_subtable = _query_table_with_indices_mapping(table, key, indices=indices)
786
+ File "/home/sanchitgandhi/datasets/src/datasets/formatting/formatting.py", line 69, in _query_table_with_indices_mapping
787
+ return _query_table(table, indices.column(0).to_pylist())
788
+ File "/home/sanchitgandhi/datasets/src/datasets/table.py", line 343, in column
789
+ return self.table.column(*args, **kwargs)
790
+ File "pyarrow/table.pxi", line 2115, in pyarrow.lib.Table.column
791
+ File "pyarrow/table.pxi", line 2130, in pyarrow.lib.Table._column
792
+ File "pyarrow/array.pxi", line 561, in pyarrow.lib._normalize_index
793
+ IndexError: index out of bounds
wandb/run-20220404_181027-mgg9caus/files/requirements.txt ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ astunparse==1.6.3
10
+ async-timeout==4.0.2
11
+ attrs==21.4.0
12
+ audioread==2.1.9
13
+ babel==2.9.1
14
+ backcall==0.2.0
15
+ beautifulsoup4==4.10.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.12
21
+ chex==0.1.1
22
+ click==8.1.0
23
+ cycler==0.11.0
24
+ datasets==2.0.1.dev0
25
+ debugpy==1.6.0
26
+ decorator==5.1.1
27
+ defusedxml==0.7.1
28
+ dill==0.3.4
29
+ dm-tree==0.1.6
30
+ docker-pycreds==0.4.0
31
+ entrypoints==0.4
32
+ executing==0.8.3
33
+ filelock==3.6.0
34
+ flatbuffers==2.0
35
+ flax==0.4.1
36
+ fonttools==4.31.2
37
+ frozenlist==1.3.0
38
+ fsspec==2022.2.0
39
+ gast==0.5.3
40
+ gitdb==4.0.9
41
+ gitpython==3.1.27
42
+ google-auth-oauthlib==0.4.6
43
+ google-auth==2.6.2
44
+ google-pasta==0.2.0
45
+ grpcio==1.44.0
46
+ h5py==3.6.0
47
+ huggingface-hub==0.4.0
48
+ idna==3.3
49
+ importlib-metadata==4.11.3
50
+ importlib-resources==5.6.0
51
+ ipdb==0.13.9
52
+ ipykernel==6.10.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.2.0
55
+ jax==0.3.4
56
+ jaxlib==0.3.2
57
+ jedi==0.18.1
58
+ jinja2==3.1.1
59
+ jiwer==2.3.0
60
+ joblib==1.1.0
61
+ json5==0.9.6
62
+ jsonschema==4.4.0
63
+ jupyter-client==7.2.1
64
+ jupyter-core==4.9.2
65
+ jupyter-server==1.16.0
66
+ jupyterlab-pygments==0.1.2
67
+ jupyterlab-server==2.12.0
68
+ jupyterlab==3.3.2
69
+ keras-preprocessing==1.1.2
70
+ keras==2.8.0
71
+ kiwisolver==1.4.2
72
+ libclang==13.0.0
73
+ librosa==0.9.1
74
+ libtpu-nightly==0.1.dev20220315
75
+ llvmlite==0.38.0
76
+ markdown==3.3.6
77
+ markupsafe==2.1.1
78
+ matplotlib-inline==0.1.3
79
+ matplotlib==3.5.1
80
+ mistune==0.8.4
81
+ msgpack==1.0.3
82
+ multidict==6.0.2
83
+ multiprocess==0.70.12.2
84
+ nbclassic==0.3.7
85
+ nbclient==0.5.13
86
+ nbconvert==6.4.5
87
+ nbformat==5.2.0
88
+ nest-asyncio==1.5.4
89
+ notebook-shim==0.1.0
90
+ notebook==6.4.10
91
+ numba==0.55.1
92
+ numpy==1.21.0
93
+ oauthlib==3.2.0
94
+ opt-einsum==3.3.0
95
+ optax==0.1.1
96
+ packaging==21.3
97
+ pandas==1.4.1
98
+ pandocfilters==1.5.0
99
+ parso==0.8.3
100
+ pathtools==0.1.2
101
+ pexpect==4.8.0
102
+ pickleshare==0.7.5
103
+ pillow==9.0.1
104
+ pip==20.0.2
105
+ pkg-resources==0.0.0
106
+ pooch==1.6.0
107
+ prometheus-client==0.13.1
108
+ promise==2.3
109
+ prompt-toolkit==3.0.28
110
+ protobuf==3.19.4
111
+ psutil==5.9.0
112
+ ptyprocess==0.7.0
113
+ pure-eval==0.2.2
114
+ pyarrow==7.0.0
115
+ pyasn1-modules==0.2.8
116
+ pyasn1==0.4.8
117
+ pycparser==2.21
118
+ pygments==2.11.2
119
+ pyparsing==3.0.7
120
+ pyrsistent==0.18.1
121
+ python-dateutil==2.8.2
122
+ python-levenshtein==0.12.2
123
+ pytz==2022.1
124
+ pyyaml==6.0
125
+ pyzmq==22.3.0
126
+ regex==2022.3.15
127
+ requests-oauthlib==1.3.1
128
+ requests==2.27.1
129
+ resampy==0.2.2
130
+ responses==0.18.0
131
+ rsa==4.8
132
+ sacremoses==0.0.49
133
+ scikit-learn==1.0.2
134
+ scipy==1.8.0
135
+ send2trash==1.8.0
136
+ sentry-sdk==1.5.8
137
+ setproctitle==1.2.2
138
+ setuptools==44.0.0
139
+ shortuuid==1.0.8
140
+ six==1.16.0
141
+ smmap==5.0.0
142
+ sniffio==1.2.0
143
+ soundfile==0.10.3.post1
144
+ soupsieve==2.3.1
145
+ stack-data==0.2.0
146
+ tensorboard-data-server==0.6.1
147
+ tensorboard-plugin-wit==1.8.1
148
+ tensorboard==2.8.0
149
+ tensorflow-io-gcs-filesystem==0.24.0
150
+ tensorflow==2.8.0
151
+ termcolor==1.1.0
152
+ terminado==0.13.3
153
+ testpath==0.6.0
154
+ tf-estimator-nightly==2.8.0.dev2021122109
155
+ threadpoolctl==3.1.0
156
+ tokenizers==0.11.6
157
+ toml==0.10.2
158
+ toolz==0.11.2
159
+ torch==1.11.0+cpu
160
+ torchaudio==0.11.0+cpu
161
+ tornado==6.1
162
+ tqdm==4.63.1
163
+ traitlets==5.1.1
164
+ transformers==4.18.0.dev0
165
+ typing-extensions==4.1.1
166
+ urllib3==1.26.9
167
+ wandb==0.12.11
168
+ wcwidth==0.2.5
169
+ webencodings==0.5.1
170
+ websocket-client==1.3.2
171
+ werkzeug==2.1.0
172
+ wheel==0.37.1
173
+ wrapt==1.14.0
174
+ xxhash==3.0.0
175
+ yarl==1.7.2
176
+ yaspin==2.1.0
177
+ zipp==3.7.0
wandb/run-20220404_181027-mgg9caus/files/wandb-metadata.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-1043-gcp-x86_64-with-glibc2.29",
3
+ "python": "3.8.10",
4
+ "heartbeatAt": "2022-04-04T18:10:31.051842",
5
+ "startedAt": "2022-04-04T18:10:27.961456",
6
+ "docker": null,
7
+ "cpu_count": 96,
8
+ "cuda": null,
9
+ "args": [
10
+ "--overwrite_output_dir",
11
+ "--freeze_feature_encoder",
12
+ "--predict_with_generate",
13
+ "--do_lower_case",
14
+ "--do_train",
15
+ "--do_eval",
16
+ "--dataset_cache_dir=/home/sanchitgandhi/cache/huggingface/datasets",
17
+ "--dataset_config_name=clean",
18
+ "--dataset_name=librispeech_asr",
19
+ "--eval_split_name=validation",
20
+ "--generation_max_length=40",
21
+ "--generation_num_beams=1",
22
+ "--gradient_accumulation_steps=1",
23
+ "--learning_rate=0.0006871268347239357",
24
+ "--length_column_name=input_length",
25
+ "--logging_steps=1",
26
+ "--matmul_precision=highest",
27
+ "--max_duration_in_seconds=15",
28
+ "--max_target_length=64",
29
+ "--min_duration_in_seconds=15",
30
+ "--model_name_or_path=./",
31
+ "--num_train_epochs=5",
32
+ "--output_dir=./",
33
+ "--per_device_eval_batch_size=2",
34
+ "--per_device_train_batch_size=2",
35
+ "--preprocessing_num_workers=16",
36
+ "--text_column_name=text",
37
+ "--train_split_name=train.100",
38
+ "--wandb_project=flax-wav2vec2-2-bart-large-cnn",
39
+ "--warmup_steps=500"
40
+ ],
41
+ "state": "running",
42
+ "program": "run_flax_speech_recognition_seq2seq.py",
43
+ "codePath": "run_flax_speech_recognition_seq2seq.py",
44
+ "git": {
45
+ "remote": "https://huggingface.co/sanchit-gandhi/flax-wav2vec2-2-bart-large-cnn",
46
+ "commit": "c09faa6b18ac3f19251c414100acf1bad11c1b16"
47
+ },
48
+ "email": "sanchit@huggingface.co",
49
+ "root": "/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn",
50
+ "host": "t1v-n-4eb331dd-w-0",
51
+ "username": "sanchitgandhi",
52
+ "executable": "/home/sanchitgandhi/venv/bin/python3"
53
+ }
wandb/run-20220404_181027-mgg9caus/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 24}}
wandb/run-20220404_181027-mgg9caus/logs/debug-internal.log ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-04-04 18:10:28,689 INFO MainThread:1241646 [internal.py:wandb_internal():92] W&B internal server running at pid: 1241646, started at: 2022-04-04 18:10:28.689690
2
+ 2022-04-04 18:10:28,692 INFO WriterThread:1241646 [datastore.py:open_for_write():77] open: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/run-mgg9caus.wandb
3
+ 2022-04-04 18:10:28,693 DEBUG SenderThread:1241646 [sender.py:send():235] send: header
4
+ 2022-04-04 18:10:28,693 DEBUG SenderThread:1241646 [sender.py:send():235] send: run
5
+ 2022-04-04 18:10:28,789 INFO SenderThread:1241646 [dir_watcher.py:__init__():169] watching files in: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files
6
+ 2022-04-04 18:10:28,789 INFO SenderThread:1241646 [sender.py:_start_run_threads():812] run started: mgg9caus with start time 1649095827
7
+ 2022-04-04 18:10:28,789 DEBUG SenderThread:1241646 [sender.py:send():235] send: summary
8
+ 2022-04-04 18:10:28,789 INFO SenderThread:1241646 [sender.py:_save_file():947] saving file wandb-summary.json with policy end
9
+ 2022-04-04 18:10:28,793 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: check_version
10
+ 2022-04-04 18:10:28,794 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: check_version
11
+ 2022-04-04 18:10:28,854 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: run_start
12
+ 2022-04-04 18:10:29,791 INFO Thread-7 :1241646 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/wandb-summary.json
13
+ 2022-04-04 18:10:31,051 DEBUG HandlerThread:1241646 [meta.py:__init__():37] meta init
14
+ 2022-04-04 18:10:31,051 DEBUG HandlerThread:1241646 [meta.py:__init__():51] meta init done
15
+ 2022-04-04 18:10:31,051 DEBUG HandlerThread:1241646 [meta.py:probe():211] probe
16
+ 2022-04-04 18:10:31,052 DEBUG HandlerThread:1241646 [meta.py:_setup_git():201] setup git
17
+ 2022-04-04 18:10:31,084 DEBUG HandlerThread:1241646 [meta.py:_setup_git():208] setup git done
18
+ 2022-04-04 18:10:31,084 DEBUG HandlerThread:1241646 [meta.py:_save_pip():55] save pip
19
+ 2022-04-04 18:10:31,084 DEBUG HandlerThread:1241646 [meta.py:_save_pip():69] save pip done
20
+ 2022-04-04 18:10:31,084 DEBUG HandlerThread:1241646 [meta.py:probe():249] probe done
21
+ 2022-04-04 18:10:31,088 DEBUG SenderThread:1241646 [sender.py:send():235] send: files
22
+ 2022-04-04 18:10:31,088 INFO SenderThread:1241646 [sender.py:_save_file():947] saving file wandb-metadata.json with policy now
23
+ 2022-04-04 18:10:31,096 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: stop_status
24
+ 2022-04-04 18:10:31,096 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: stop_status
25
+ 2022-04-04 18:10:31,139 DEBUG SenderThread:1241646 [sender.py:send():235] send: telemetry
26
+ 2022-04-04 18:10:31,335 INFO Thread-11 :1241646 [upload_job.py:push():137] Uploaded file /tmp/tmpxnk1ubkiwandb/3n64l3r4-wandb-metadata.json
27
+ 2022-04-04 18:10:31,791 INFO Thread-7 :1241646 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/wandb-metadata.json
28
+ 2022-04-04 18:10:31,792 INFO Thread-7 :1241646 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/output.log
29
+ 2022-04-04 18:10:31,792 INFO Thread-7 :1241646 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/requirements.txt
30
+ 2022-04-04 18:10:33,792 INFO Thread-7 :1241646 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/output.log
31
+ 2022-04-04 18:10:46,142 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: stop_status
32
+ 2022-04-04 18:10:46,143 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: stop_status
33
+ 2022-04-04 18:10:47,800 INFO Thread-7 :1241646 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/output.log
34
+ 2022-04-04 18:10:49,801 INFO Thread-7 :1241646 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/output.log
35
+ 2022-04-04 18:10:50,725 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: partial_history
36
+ 2022-04-04 18:10:50,725 DEBUG SenderThread:1241646 [sender.py:send():235] send: telemetry
37
+ 2022-04-04 18:10:51,801 INFO Thread-7 :1241646 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/output.log
38
+ 2022-04-04 18:10:53,234 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: poll_exit
39
+ 2022-04-04 18:10:53,234 DEBUG SenderThread:1241646 [sender.py:send():235] send: exit
40
+ 2022-04-04 18:10:53,235 INFO SenderThread:1241646 [sender.py:send_exit():371] handling exit code: 1
41
+ 2022-04-04 18:10:53,235 INFO SenderThread:1241646 [sender.py:send_exit():373] handling runtime: 24
42
+ 2022-04-04 18:10:53,237 INFO SenderThread:1241646 [sender.py:_save_file():947] saving file wandb-summary.json with policy end
43
+ 2022-04-04 18:10:53,237 INFO SenderThread:1241646 [sender.py:send_exit():379] send defer
44
+ 2022-04-04 18:10:53,237 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: poll_exit
45
+ 2022-04-04 18:10:53,238 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
46
+ 2022-04-04 18:10:53,238 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 0
47
+ 2022-04-04 18:10:53,238 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
48
+ 2022-04-04 18:10:53,238 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 0
49
+ 2022-04-04 18:10:53,238 INFO SenderThread:1241646 [sender.py:transition_state():392] send defer: 1
50
+ 2022-04-04 18:10:53,238 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
51
+ 2022-04-04 18:10:53,238 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 1
52
+ 2022-04-04 18:10:53,321 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
53
+ 2022-04-04 18:10:53,321 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 1
54
+ 2022-04-04 18:10:53,321 INFO SenderThread:1241646 [sender.py:transition_state():392] send defer: 2
55
+ 2022-04-04 18:10:53,321 DEBUG SenderThread:1241646 [sender.py:send():235] send: stats
56
+ 2022-04-04 18:10:53,322 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
57
+ 2022-04-04 18:10:53,322 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 2
58
+ 2022-04-04 18:10:53,322 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
59
+ 2022-04-04 18:10:53,322 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 2
60
+ 2022-04-04 18:10:53,322 INFO SenderThread:1241646 [sender.py:transition_state():392] send defer: 3
61
+ 2022-04-04 18:10:53,322 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
62
+ 2022-04-04 18:10:53,322 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 3
63
+ 2022-04-04 18:10:53,322 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
64
+ 2022-04-04 18:10:53,322 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 3
65
+ 2022-04-04 18:10:53,322 INFO SenderThread:1241646 [sender.py:transition_state():392] send defer: 4
66
+ 2022-04-04 18:10:53,323 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
67
+ 2022-04-04 18:10:53,323 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 4
68
+ 2022-04-04 18:10:53,323 DEBUG SenderThread:1241646 [sender.py:send():235] send: summary
69
+ 2022-04-04 18:10:53,323 INFO SenderThread:1241646 [sender.py:_save_file():947] saving file wandb-summary.json with policy end
70
+ 2022-04-04 18:10:53,324 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
71
+ 2022-04-04 18:10:53,324 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 4
72
+ 2022-04-04 18:10:53,324 INFO SenderThread:1241646 [sender.py:transition_state():392] send defer: 5
73
+ 2022-04-04 18:10:53,324 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
74
+ 2022-04-04 18:10:53,324 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 5
75
+ 2022-04-04 18:10:53,324 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
76
+ 2022-04-04 18:10:53,324 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 5
77
+ 2022-04-04 18:10:53,339 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: poll_exit
78
+ 2022-04-04 18:10:53,392 INFO SenderThread:1241646 [sender.py:transition_state():392] send defer: 6
79
+ 2022-04-04 18:10:53,392 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: poll_exit
80
+ 2022-04-04 18:10:53,393 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
81
+ 2022-04-04 18:10:53,393 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 6
82
+ 2022-04-04 18:10:53,393 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
83
+ 2022-04-04 18:10:53,393 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 6
84
+ 2022-04-04 18:10:53,393 INFO SenderThread:1241646 [dir_watcher.py:finish():283] shutting down directory watcher
85
+ 2022-04-04 18:10:53,494 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: poll_exit
86
+ 2022-04-04 18:10:53,802 INFO Thread-7 :1241646 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/output.log
87
+ 2022-04-04 18:10:53,803 INFO SenderThread:1241646 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/config.yaml
88
+ 2022-04-04 18:10:53,803 INFO SenderThread:1241646 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/wandb-summary.json
89
+ 2022-04-04 18:10:53,804 INFO SenderThread:1241646 [dir_watcher.py:finish():313] scan: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files
90
+ 2022-04-04 18:10:53,804 INFO SenderThread:1241646 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/requirements.txt requirements.txt
91
+ 2022-04-04 18:10:53,804 INFO SenderThread:1241646 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/output.log output.log
92
+ 2022-04-04 18:10:53,804 INFO SenderThread:1241646 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/wandb-metadata.json wandb-metadata.json
93
+ 2022-04-04 18:10:53,804 INFO SenderThread:1241646 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/config.yaml config.yaml
94
+ 2022-04-04 18:10:53,804 INFO SenderThread:1241646 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/wandb-summary.json wandb-summary.json
95
+ 2022-04-04 18:10:53,805 INFO SenderThread:1241646 [sender.py:transition_state():392] send defer: 7
96
+ 2022-04-04 18:10:53,805 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: poll_exit
97
+ 2022-04-04 18:10:53,808 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
98
+ 2022-04-04 18:10:53,808 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 7
99
+ 2022-04-04 18:10:53,808 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
100
+ 2022-04-04 18:10:53,808 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 7
101
+ 2022-04-04 18:10:53,808 INFO SenderThread:1241646 [file_pusher.py:finish():145] shutting down file pusher
102
+ 2022-04-04 18:10:53,907 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: poll_exit
103
+ 2022-04-04 18:10:53,907 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: poll_exit
104
+ 2022-04-04 18:10:54,008 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: poll_exit
105
+ 2022-04-04 18:10:54,009 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: poll_exit
106
+ 2022-04-04 18:10:54,056 INFO Thread-13 :1241646 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/output.log
107
+ 2022-04-04 18:10:54,061 INFO Thread-15 :1241646 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/wandb-summary.json
108
+ 2022-04-04 18:10:54,082 INFO Thread-14 :1241646 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/config.yaml
109
+ 2022-04-04 18:10:54,102 INFO Thread-12 :1241646 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/files/requirements.txt
110
+ 2022-04-04 18:10:54,110 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: poll_exit
111
+ 2022-04-04 18:10:54,110 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: poll_exit
112
+ 2022-04-04 18:10:54,212 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: poll_exit
113
+ 2022-04-04 18:10:54,212 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: poll_exit
114
+ 2022-04-04 18:10:54,303 INFO Thread-6 :1241646 [sender.py:transition_state():392] send defer: 8
115
+ 2022-04-04 18:10:54,303 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
116
+ 2022-04-04 18:10:54,303 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 8
117
+ 2022-04-04 18:10:54,304 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
118
+ 2022-04-04 18:10:54,304 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 8
119
+ 2022-04-04 18:10:54,314 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: poll_exit
120
+ 2022-04-04 18:10:54,430 INFO SenderThread:1241646 [sender.py:transition_state():392] send defer: 9
121
+ 2022-04-04 18:10:54,431 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: poll_exit
122
+ 2022-04-04 18:10:54,431 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
123
+ 2022-04-04 18:10:54,431 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 9
124
+ 2022-04-04 18:10:54,431 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
125
+ 2022-04-04 18:10:54,431 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 9
126
+ 2022-04-04 18:10:54,431 INFO SenderThread:1241646 [sender.py:transition_state():392] send defer: 10
127
+ 2022-04-04 18:10:54,432 DEBUG SenderThread:1241646 [sender.py:send():235] send: final
128
+ 2022-04-04 18:10:54,432 DEBUG SenderThread:1241646 [sender.py:send():235] send: footer
129
+ 2022-04-04 18:10:54,432 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: defer
130
+ 2022-04-04 18:10:54,432 INFO HandlerThread:1241646 [handler.py:handle_request_defer():164] handle defer: 10
131
+ 2022-04-04 18:10:54,432 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: defer
132
+ 2022-04-04 18:10:54,432 INFO SenderThread:1241646 [sender.py:send_request_defer():388] handle sender defer: 10
133
+ 2022-04-04 18:10:54,532 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: poll_exit
134
+ 2022-04-04 18:10:54,532 DEBUG SenderThread:1241646 [sender.py:send_request():249] send_request: poll_exit
135
+ 2022-04-04 18:10:54,533 INFO SenderThread:1241646 [file_pusher.py:join():150] waiting for file pusher
136
+ 2022-04-04 18:10:54,686 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: sampled_history
137
+ 2022-04-04 18:10:54,687 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: get_summary
138
+ 2022-04-04 18:10:54,688 DEBUG HandlerThread:1241646 [handler.py:handle_request():141] handle_request: shutdown
139
+ 2022-04-04 18:10:54,688 INFO HandlerThread:1241646 [handler.py:finish():778] shutting down handler
140
+ 2022-04-04 18:10:55,432 INFO WriterThread:1241646 [datastore.py:close():281] close: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/run-mgg9caus.wandb
141
+ 2022-04-04 18:10:55,585 INFO SenderThread:1241646 [sender.py:finish():1078] shutting down sender
142
+ 2022-04-04 18:10:55,585 INFO SenderThread:1241646 [file_pusher.py:finish():145] shutting down file pusher
143
+ 2022-04-04 18:10:55,585 INFO SenderThread:1241646 [file_pusher.py:join():150] waiting for file pusher
144
+ 2022-04-04 18:10:55,587 INFO MainThread:1241646 [internal.py:handle_exit():82] Internal process exited
wandb/run-20220404_181027-mgg9caus/logs/debug.log ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-04-04 18:10:27,962 INFO MainThread:1241191 [wandb_setup.py:_flush():75] Loading settings from /home/sanchitgandhi/.config/wandb/settings
2
+ 2022-04-04 18:10:27,962 INFO MainThread:1241191 [wandb_setup.py:_flush():75] Loading settings from wandb/settings
3
+ 2022-04-04 18:10:27,962 INFO MainThread:1241191 [wandb_setup.py:_flush():75] Loading settings from environment variables: {'entity': 'sanchit-gandhi', 'project': 'flax-wav2vec2-2-bart-large-cnn', 'sweep_id': 'a50sny4c', 'root_dir': '/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn', 'run_id': 'mgg9caus', 'sweep_param_path': '/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/sweep-a50sny4c/config-mgg9caus.yaml'}
4
+ 2022-04-04 18:10:27,962 INFO MainThread:1241191 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_flax_speech_recognition_seq2seq.py', 'program': 'run_flax_speech_recognition_seq2seq.py'}
5
+ 2022-04-04 18:10:27,963 INFO MainThread:1241191 [wandb_init.py:_log_setup():405] Logging user logs to /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/logs/debug.log
6
+ 2022-04-04 18:10:27,963 INFO MainThread:1241191 [wandb_init.py:_log_setup():406] Logging internal logs to /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181027-mgg9caus/logs/debug-internal.log
7
+ 2022-04-04 18:10:27,963 INFO MainThread:1241191 [wandb_init.py:init():439] calling init triggers
8
+ 2022-04-04 18:10:27,963 INFO MainThread:1241191 [wandb_init.py:init():442] wandb.init called with sweep_config: {'dataset_cache_dir': '/home/sanchitgandhi/cache/huggingface/datasets', 'dataset_config_name': 'clean', 'dataset_name': 'librispeech_asr', 'eval_split_name': 'validation', 'generation_max_length': 40, 'generation_num_beams': 1, 'gradient_accumulation_steps': 1, 'learning_rate': 0.0006871268347239357, 'length_column_name': 'input_length', 'logging_steps': 1, 'matmul_precision': 'highest', 'max_duration_in_seconds': 15, 'max_target_length': 64, 'min_duration_in_seconds': 15, 'model_name_or_path': './', 'num_train_epochs': 5, 'output_dir': './', 'per_device_eval_batch_size': 2, 'per_device_train_batch_size': 2, 'preprocessing_num_workers': 16, 'text_column_name': 'text', 'train_split_name': 'train.100', 'wandb_project': 'flax-wav2vec2-2-bart-large-cnn', 'warmup_steps': 500}
9
+ config: {}
10
+ 2022-04-04 18:10:27,963 INFO MainThread:1241191 [wandb_init.py:init():492] starting backend
11
+ 2022-04-04 18:10:27,963 INFO MainThread:1241191 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-04-04 18:10:27,976 INFO MainThread:1241191 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-04-04 18:10:27,984 INFO MainThread:1241191 [backend.py:ensure_launched():224] started backend process with pid: 1241646
14
+ 2022-04-04 18:10:27,986 INFO MainThread:1241191 [wandb_init.py:init():501] backend started and connected
15
+ 2022-04-04 18:10:27,992 INFO MainThread:1241191 [wandb_run.py:_config_callback():992] config_cb None None {'dataset_cache_dir': '/home/sanchitgandhi/cache/huggingface/datasets', 'dataset_config_name': 'clean', 'dataset_name': 'librispeech_asr', 'eval_split_name': 'validation', 'generation_max_length': 40, 'generation_num_beams': 1, 'gradient_accumulation_steps': 1, 'learning_rate': 0.0006871268347239357, 'length_column_name': 'input_length', 'logging_steps': 1, 'matmul_precision': 'highest', 'max_duration_in_seconds': 15, 'max_target_length': 64, 'min_duration_in_seconds': 15, 'model_name_or_path': './', 'num_train_epochs': 5, 'output_dir': './', 'per_device_eval_batch_size': 2, 'per_device_train_batch_size': 2, 'preprocessing_num_workers': 16, 'text_column_name': 'text', 'train_split_name': 'train.100', 'wandb_project': 'flax-wav2vec2-2-bart-large-cnn', 'warmup_steps': 500}
16
+ 2022-04-04 18:10:27,996 INFO MainThread:1241191 [wandb_init.py:init():565] updated telemetry
17
+ 2022-04-04 18:10:28,015 INFO MainThread:1241191 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
18
+ 2022-04-04 18:10:28,793 INFO MainThread:1241191 [wandb_run.py:_on_init():1759] communicating current version
19
+ 2022-04-04 18:10:28,852 INFO MainThread:1241191 [wandb_run.py:_on_init():1763] got version response
20
+ 2022-04-04 18:10:28,852 INFO MainThread:1241191 [wandb_init.py:init():625] starting run threads in backend
21
+ 2022-04-04 18:10:31,094 INFO MainThread:1241191 [wandb_run.py:_console_start():1733] atexit reg
22
+ 2022-04-04 18:10:31,094 INFO MainThread:1241191 [wandb_run.py:_redirect():1606] redirect: SettingsConsole.REDIRECT
23
+ 2022-04-04 18:10:31,095 INFO MainThread:1241191 [wandb_run.py:_redirect():1611] Redirecting console.
24
+ 2022-04-04 18:10:31,096 INFO MainThread:1241191 [wandb_run.py:_redirect():1667] Redirects installed.
25
+ 2022-04-04 18:10:31,097 INFO MainThread:1241191 [wandb_init.py:init():664] run started, returning control to user process
26
+ 2022-04-04 18:10:50,720 INFO MainThread:1241191 [wandb_run.py:_atexit_cleanup():1702] got exitcode: 1
27
+ 2022-04-04 18:10:50,724 INFO MainThread:1241191 [wandb_run.py:_restore():1674] restore
28
+ 2022-04-04 18:10:53,238 INFO MainThread:1241191 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
29
+ wandb_count: 1
30
+ }
31
+ pusher_stats {
32
+ uploaded_bytes: 1948
33
+ total_bytes: 1948
34
+ }
35
+
36
+ 2022-04-04 18:10:53,393 INFO MainThread:1241191 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
37
+ wandb_count: 1
38
+ }
39
+ pusher_stats {
40
+ uploaded_bytes: 1948
41
+ total_bytes: 1948
42
+ }
43
+
44
+ 2022-04-04 18:10:53,806 INFO MainThread:1241191 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
45
+ wandb_count: 4
46
+ }
47
+ pusher_stats {
48
+ uploaded_bytes: 1948
49
+ total_bytes: 45265
50
+ }
51
+
52
+ 2022-04-04 18:10:53,908 INFO MainThread:1241191 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
53
+ wandb_count: 5
54
+ }
55
+ pusher_stats {
56
+ uploaded_bytes: 45292
57
+ total_bytes: 45292
58
+ }
59
+
60
+ 2022-04-04 18:10:54,009 INFO MainThread:1241191 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
61
+ wandb_count: 5
62
+ }
63
+ pusher_stats {
64
+ uploaded_bytes: 45292
65
+ total_bytes: 45292
66
+ }
67
+
68
+ 2022-04-04 18:10:54,111 INFO MainThread:1241191 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
69
+ wandb_count: 5
70
+ }
71
+ pusher_stats {
72
+ uploaded_bytes: 45292
73
+ total_bytes: 45292
74
+ }
75
+
76
+ 2022-04-04 18:10:54,213 INFO MainThread:1241191 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
77
+ wandb_count: 5
78
+ }
79
+ pusher_stats {
80
+ uploaded_bytes: 45292
81
+ total_bytes: 45292
82
+ }
83
+
84
+ 2022-04-04 18:10:54,431 INFO MainThread:1241191 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
85
+ wandb_count: 5
86
+ }
87
+ pusher_stats {
88
+ uploaded_bytes: 45292
89
+ total_bytes: 45292
90
+ }
91
+
92
+ 2022-04-04 18:10:54,585 INFO MainThread:1241191 [wandb_run.py:_on_finish():1831] got exit ret: done: true
93
+ exit_result {
94
+ }
95
+ file_counts {
96
+ wandb_count: 5
97
+ }
98
+ pusher_stats {
99
+ uploaded_bytes: 45292
100
+ total_bytes: 45292
101
+ }
102
+ local_info {
103
+ }
104
+
105
+ 2022-04-04 18:10:56,074 INFO MainThread:1241191 [wandb_run.py:_footer_history_summary_info():2865] rendering history
106
+ 2022-04-04 18:10:56,074 INFO MainThread:1241191 [wandb_run.py:_footer_history_summary_info():2894] rendering summary
107
+ 2022-04-04 18:10:56,075 INFO MainThread:1241191 [wandb_run.py:_footer_sync_info():2822] logging synced files
wandb/run-20220404_181027-mgg9caus/run-mgg9caus.wandb ADDED
Binary file (40.3 kB). View file
 
wandb/run-20220404_181103-88xgr1fg/files/config.yaml ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ cli_version: 0.12.11
7
+ framework: huggingface
8
+ huggingface_version: 4.18.0.dev0
9
+ is_jupyter_run: false
10
+ is_kaggle_kernel: false
11
+ python_version: 3.8.10
12
+ start_time: 1649095863
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 11
19
+ - 12
20
+ - 45
21
+ - 49
22
+ - 51
23
+ 2:
24
+ - 1
25
+ - 2
26
+ - 3
27
+ - 11
28
+ - 12
29
+ - 45
30
+ - 49
31
+ - 51
32
+ 4: 3.8.10
33
+ 5: 0.12.11
34
+ 6: 4.18.0.dev0
35
+ 8:
36
+ - 5
37
+ dataset_cache_dir:
38
+ desc: null
39
+ value: /home/sanchitgandhi/cache/huggingface/datasets
40
+ dataset_config_name:
41
+ desc: null
42
+ value: clean
43
+ dataset_name:
44
+ desc: null
45
+ value: librispeech_asr
46
+ eval_split_name:
47
+ desc: null
48
+ value: validation
49
+ generation_max_length:
50
+ desc: null
51
+ value: 40
52
+ generation_num_beams:
53
+ desc: null
54
+ value: 1
55
+ gradient_accumulation_steps:
56
+ desc: null
57
+ value: 1
58
+ learning_rate:
59
+ desc: null
60
+ value: 9.383495031304748e-05
61
+ length_column_name:
62
+ desc: null
63
+ value: input_length
64
+ logging_steps:
65
+ desc: null
66
+ value: 1
67
+ matmul_precision:
68
+ desc: null
69
+ value: highest
70
+ max_duration_in_seconds:
71
+ desc: null
72
+ value: 15
73
+ max_target_length:
74
+ desc: null
75
+ value: 64
76
+ min_duration_in_seconds:
77
+ desc: null
78
+ value: 15
79
+ model_name_or_path:
80
+ desc: null
81
+ value: ./
82
+ num_train_epochs:
83
+ desc: null
84
+ value: 5
85
+ output_dir:
86
+ desc: null
87
+ value: ./
88
+ per_device_eval_batch_size:
89
+ desc: null
90
+ value: 2
91
+ per_device_train_batch_size:
92
+ desc: null
93
+ value: 2
94
+ preprocessing_num_workers:
95
+ desc: null
96
+ value: 16
97
+ text_column_name:
98
+ desc: null
99
+ value: text
100
+ train_split_name:
101
+ desc: null
102
+ value: train.100
103
+ wandb_project:
104
+ desc: null
105
+ value: flax-wav2vec2-2-bart-large-cnn
106
+ warmup_steps:
107
+ desc: null
108
+ value: 500
wandb/run-20220404_181103-88xgr1fg/files/output.log ADDED
@@ -0,0 +1,793 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ loading configuration file ./config.json
2
+ Model config SpeechEncoderDecoderConfig {
3
+ "_name_or_path": "./",
4
+ "architectures": [
5
+ "SpeechEncoderDecoderModel"
6
+ ],
7
+ "decoder": {
8
+ "_name_or_path": "facebook/bart-large-cnn",
9
+ "_num_labels": 3,
10
+ "activation_dropout": 0.0,
11
+ "activation_function": "gelu",
12
+ "add_cross_attention": true,
13
+ "add_final_layer_norm": false,
14
+ "architectures": [
15
+ "BartForConditionalGeneration"
16
+ ],
17
+ "attention_dropout": 0.0,
18
+ "bad_words_ids": null,
19
+ "bos_token_id": 0,
20
+ "chunk_size_feed_forward": 0,
21
+ "classif_dropout": 0.0,
22
+ "classifier_dropout": 0.0,
23
+ "cross_attention_hidden_size": null,
24
+ "d_model": 1024,
25
+ "decoder_attention_heads": 16,
26
+ "decoder_ffn_dim": 4096,
27
+ "decoder_layerdrop": 0.0,
28
+ "decoder_layers": 12,
29
+ "decoder_start_token_id": 2,
30
+ "diversity_penalty": 0.0,
31
+ "do_sample": false,
32
+ "dropout": 0.1,
33
+ "early_stopping": true,
34
+ "encoder_attention_heads": 16,
35
+ "encoder_ffn_dim": 4096,
36
+ "encoder_layerdrop": 0.0,
37
+ "encoder_layers": 12,
38
+ "encoder_no_repeat_ngram_size": 0,
39
+ "eos_token_id": 2,
40
+ "exponential_decay_length_penalty": null,
41
+ "finetuning_task": null,
42
+ "force_bos_token_to_be_generated": true,
43
+ "forced_bos_token_id": 0,
44
+ "forced_eos_token_id": 2,
45
+ "gradient_checkpointing": false,
46
+ "id2label": {
47
+ "0": "LABEL_0",
48
+ "1": "LABEL_1",
49
+ "2": "LABEL_2"
50
+ },
51
+ "init_std": 0.02,
52
+ "is_decoder": true,
53
+ "is_encoder_decoder": false,
54
+ "label2id": {
55
+ "LABEL_0": 0,
56
+ "LABEL_1": 1,
57
+ "LABEL_2": 2
58
+ },
59
+ "length_penalty": 2.0,
60
+ "max_length": 142,
61
+ "max_position_embeddings": 1024,
62
+ "min_length": 56,
63
+ "model_type": "bart",
64
+ "no_repeat_ngram_size": 3,
65
+ "normalize_before": false,
66
+ "num_beam_groups": 1,
67
+ "num_beams": 4,
68
+ "num_hidden_layers": 12,
69
+ "num_return_sequences": 1,
70
+ "output_attentions": false,
71
+ "output_hidden_states": false,
72
+ "output_past": true,
73
+ "output_scores": false,
74
+ "pad_token_id": 1,
75
+ "prefix": " ",
76
+ "problem_type": null,
77
+ "pruned_heads": {},
78
+ "remove_invalid_values": false,
79
+ "repetition_penalty": 1.0,
80
+ "return_dict": true,
81
+ "return_dict_in_generate": false,
82
+ "scale_embedding": false,
83
+ "sep_token_id": null,
84
+ "task_specific_params": {
85
+ "summarization": {
86
+ "early_stopping": true,
87
+ "length_penalty": 2.0,
88
+ "max_length": 142,
89
+ "min_length": 56,
90
+ "no_repeat_ngram_size": 3,
91
+ "num_beams": 4
92
+ }
93
+ },
94
+ "temperature": 1.0,
95
+ "tie_encoder_decoder": false,
96
+ "tie_word_embeddings": true,
97
+ "tokenizer_class": null,
98
+ "top_k": 50,
99
+ "top_p": 1.0,
100
+ "torch_dtype": null,
101
+ "torchscript": false,
102
+ "transformers_version": "4.18.0.dev0",
103
+ "typical_p": 1.0,
104
+ "use_bfloat16": false,
105
+ "use_cache": true,
106
+ "vocab_size": 50264
107
+ },
108
+ "decoder_start_token_id": 0,
109
+ "encoder": {
110
+ "_name_or_path": "facebook/wav2vec2-large-lv60",
111
+ "activation_dropout": 0.1,
112
+ "adapter_kernel_size": 3,
113
+ "adapter_stride": 2,
114
+ "add_adapter": true,
115
+ "add_cross_attention": false,
116
+ "apply_spec_augment": true,
117
+ "architectures": [
118
+ "Wav2Vec2ForPreTraining"
119
+ ],
120
+ "attention_dropout": 0.1,
121
+ "bad_words_ids": null,
122
+ "bos_token_id": 1,
123
+ "chunk_size_feed_forward": 0,
124
+ "classifier_proj_size": 256,
125
+ "codevector_dim": 768,
126
+ "contrastive_logits_temperature": 0.1,
127
+ "conv_bias": true,
128
+ "conv_dim": [
129
+ 512,
130
+ 512,
131
+ 512,
132
+ 512,
133
+ 512,
134
+ 512,
135
+ 512
136
+ ],
137
+ "conv_kernel": [
138
+ 10,
139
+ 3,
140
+ 3,
141
+ 3,
142
+ 3,
143
+ 2,
144
+ 2
145
+ ],
146
+ "conv_stride": [
147
+ 5,
148
+ 2,
149
+ 2,
150
+ 2,
151
+ 2,
152
+ 2,
153
+ 2
154
+ ],
155
+ "cross_attention_hidden_size": null,
156
+ "ctc_loss_reduction": "sum",
157
+ "ctc_zero_infinity": false,
158
+ "decoder_start_token_id": null,
159
+ "diversity_loss_weight": 0.1,
160
+ "diversity_penalty": 0.0,
161
+ "do_sample": false,
162
+ "do_stable_layer_norm": true,
163
+ "early_stopping": false,
164
+ "encoder_no_repeat_ngram_size": 0,
165
+ "eos_token_id": 2,
166
+ "exponential_decay_length_penalty": null,
167
+ "feat_extract_activation": "gelu",
168
+ "feat_extract_dropout": 0.0,
169
+ "feat_extract_norm": "layer",
170
+ "feat_proj_dropout": 0.0,
171
+ "feat_quantizer_dropout": 0.0,
172
+ "final_dropout": 0.0,
173
+ "finetuning_task": null,
174
+ "forced_bos_token_id": null,
175
+ "forced_eos_token_id": null,
176
+ "gradient_checkpointing": false,
177
+ "hidden_act": "gelu",
178
+ "hidden_dropout": 0.1,
179
+ "hidden_dropout_prob": 0.1,
180
+ "hidden_size": 1024,
181
+ "id2label": {
182
+ "0": "LABEL_0",
183
+ "1": "LABEL_1"
184
+ },
185
+ "initializer_range": 0.02,
186
+ "intermediate_size": 4096,
187
+ "is_decoder": false,
188
+ "is_encoder_decoder": false,
189
+ "label2id": {
190
+ "LABEL_0": 0,
191
+ "LABEL_1": 1
192
+ },
193
+ "layer_norm_eps": 1e-05,
194
+ "layerdrop": 0.0,
195
+ "length_penalty": 1.0,
196
+ "mask_feature_length": 10,
197
+ "mask_feature_min_masks": 0,
198
+ "mask_feature_prob": 0.0,
199
+ "mask_time_length": 10,
200
+ "mask_time_min_masks": 2,
201
+ "mask_time_prob": 0.1,
202
+ "max_length": 20,
203
+ "min_length": 0,
204
+ "model_type": "wav2vec2",
205
+ "no_repeat_ngram_size": 0,
206
+ "num_adapter_layers": 3,
207
+ "num_attention_heads": 16,
208
+ "num_beam_groups": 1,
209
+ "num_beams": 1,
210
+ "num_codevector_groups": 2,
211
+ "num_codevectors_per_group": 320,
212
+ "num_conv_pos_embedding_groups": 16,
213
+ "num_conv_pos_embeddings": 128,
214
+ "num_feat_extract_layers": 7,
215
+ "num_hidden_layers": 24,
216
+ "num_negatives": 100,
217
+ "num_return_sequences": 1,
218
+ "output_attentions": false,
219
+ "output_hidden_size": 1024,
220
+ "output_hidden_states": false,
221
+ "output_scores": false,
222
+ "pad_token_id": 0,
223
+ "prefix": null,
224
+ "problem_type": null,
225
+ "proj_codevector_dim": 768,
226
+ "pruned_heads": {},
227
+ "remove_invalid_values": false,
228
+ "repetition_penalty": 1.0,
229
+ "return_dict": true,
230
+ "return_dict_in_generate": false,
231
+ "sep_token_id": null,
232
+ "task_specific_params": null,
233
+ "tdnn_dilation": [
234
+ 1,
235
+ 2,
236
+ 3,
237
+ 1,
238
+ 1
239
+ ],
240
+ "tdnn_dim": [
241
+ 512,
242
+ 512,
243
+ 512,
244
+ 512,
245
+ 1500
246
+ ],
247
+ "tdnn_kernel": [
248
+ 5,
249
+ 3,
250
+ 3,
251
+ 1,
252
+ 1
253
+ ],
254
+ "temperature": 1.0,
255
+ "tie_encoder_decoder": false,
256
+ "tie_word_embeddings": true,
257
+ "tokenizer_class": null,
258
+ "top_k": 50,
259
+ "top_p": 1.0,
260
+ "torch_dtype": null,
261
+ "torchscript": false,
262
+ "transformers_version": "4.18.0.dev0",
263
+ "typical_p": 1.0,
264
+ "use_bfloat16": false,
265
+ "use_weighted_layer_sum": false,
266
+ "vocab_size": 32,
267
+ "xvector_output_dim": 512
268
+ },
269
+ "eos_token_id": 2,
270
+ "is_encoder_decoder": true,
271
+ "max_length": 40,
272
+ "model_type": "speech-encoder-decoder",
273
+ "pad_token_id": 1,
274
+ "processor_class": "Wav2Vec2Processor",
275
+ "tie_word_embeddings": false,
276
+ "transformers_version": null,
277
+ "use_cache": false
278
+ }
279
+ loading feature extractor configuration file ./preprocessor_config.json
280
+ Feature extractor Wav2Vec2FeatureExtractor {
281
+ "do_normalize": true,
282
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
283
+ "feature_size": 1,
284
+ "padding_side": "right",
285
+ "padding_value": 0.0,
286
+ "return_attention_mask": true,
287
+ "sampling_rate": 16000
288
+ }
289
+ Didn't find file ./added_tokens.json. We won't load it.
290
+ loading file ./vocab.json
291
+ loading file ./merges.txt
292
+ loading file ./tokenizer.json
293
+ loading file None
294
+ loading file ./special_tokens_map.json
295
+ loading file ./tokenizer_config.json
296
+ loading weights file ./flax_model.msgpack
297
+ 04/04/2022 18:11:06 - INFO - __main__ - Training/evaluation parameters FlaxSeq2SeqTrainingArguments(
298
+ _n_gpu=0,
299
+ adafactor=False,
300
+ adam_beta1=0.9,
301
+ adam_beta2=0.999,
302
+ adam_epsilon=1e-08,
303
+ bf16=False,
304
+ bf16_full_eval=False,
305
+ data_seed=None,
306
+ dataloader_drop_last=False,
307
+ dataloader_num_workers=0,
308
+ dataloader_pin_memory=True,
309
+ ddp_bucket_cap_mb=None,
310
+ ddp_find_unused_parameters=None,
311
+ debug=[],
312
+ deepspeed=None,
313
+ disable_tqdm=False,
314
+ do_eval=True,
315
+ do_predict=False,
316
+ do_train=True,
317
+ eval_accumulation_steps=None,
318
+ eval_delay=0,
319
+ eval_steps=None,
320
+ evaluation_strategy=IntervalStrategy.NO,
321
+ fp16=False,
322
+ fp16_backend=auto,
323
+ fp16_full_eval=False,
324
+ fp16_opt_level=O1,
325
+ generation_max_length=40,
326
+ generation_num_beams=1,
327
+ gradient_accumulation_steps=1,
328
+ gradient_checkpointing=False,
329
+ greater_is_better=None,
330
+ group_by_length=False,
331
+ half_precision_backend=auto,
332
+ hub_model_id=None,
333
+ hub_strategy=HubStrategy.EVERY_SAVE,
334
+ hub_token=<HUB_TOKEN>,
335
+ ignore_data_skip=False,
336
+ label_names=None,
337
+ label_smoothing_factor=0.0,
338
+ learning_rate=9.383495031304748e-05,
339
+ length_column_name=input_length,
340
+ load_best_model_at_end=False,
341
+ local_rank=-1,
342
+ log_level=-1,
343
+ log_level_replica=-1,
344
+ log_on_each_node=True,
345
+ logging_dir=./runs/Apr04_18-11-03_t1v-n-4eb331dd-w-0,
346
+ logging_first_step=False,
347
+ logging_nan_inf_filter=True,
348
+ logging_steps=1,
349
+ logging_strategy=IntervalStrategy.STEPS,
350
+ lr_scheduler_type=SchedulerType.LINEAR,
351
+ matmul_precision=highest,
352
+ max_grad_norm=1.0,
353
+ max_steps=-1,
354
+ metric_for_best_model=None,
355
+ mixed_precision=False,
356
+ mp_parameters=,
357
+ no_cuda=False,
358
+ num_train_epochs=5.0,
359
+ optim=OptimizerNames.ADAMW_HF,
360
+ output_dir=./,
361
+ overwrite_output_dir=True,
362
+ past_index=-1,
363
+ per_device_eval_batch_size=2,
364
+ per_device_train_batch_size=2,
365
+ predict_with_generate=True,
366
+ prediction_loss_only=False,
367
+ push_to_hub=False,
368
+ push_to_hub_model_id=None,
369
+ push_to_hub_organization=None,
370
+ push_to_hub_token=<PUSH_TO_HUB_TOKEN>,
371
+ remove_unused_columns=True,
372
+ report_to=['tensorboard', 'wandb'],
373
+ resume_from_checkpoint=None,
374
+ run_name=./,
375
+ save_on_each_node=False,
376
+ save_steps=500,
377
+ save_strategy=IntervalStrategy.STEPS,
378
+ save_total_limit=None,
379
+ seed=42,
380
+ sharded_ddp=[],
381
+ skip_memory_metrics=True,
382
+ sortish_sampler=False,
383
+ tf32=None,
384
+ tpu_metrics_debug=False,
385
+ tpu_num_cores=None,
386
+ use_legacy_prediction_loop=False,
387
+ warmup_ratio=0.0,
388
+ warmup_steps=500,
389
+ weight_decay=0.0,
390
+ xpu_backend=None,
391
+ )
392
+ 04/04/2022 18:11:06 - INFO - __main__ - JAX devices: 1, matmul precision: highest
393
+ 04/04/2022 18:11:06 - WARNING - datasets.builder - Reusing dataset librispeech_asr (/home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c)
394
+ 04/04/2022 18:11:07 - WARNING - datasets.builder - Reusing dataset librispeech_asr (/home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c)
395
+ tcmalloc: large alloc 2353643520 bytes == 0x9dd4e000 @ 0x7fc0f868d680 0x7fc0f86ae824 0x5f8a01 0x648cf1 0x5c4676 0x4f290e 0x64f718 0x5048b3 0x56b1da 0x56939a 0x5f6a13 0x50af49 0x5f3547 0x56c8cd 0x56939a 0x50aaa0 0x56c28c 0x56939a 0x5f6a13 0x56b0ae 0x56939a 0x68d047 0x67e351 0x67e3cf 0x67e471 0x67e817 0x6b6fe2 0x6b736d 0x7fc0f849f0b3 0x5fa5ce
396
+ All model checkpoint weights were used when initializing FlaxSpeechEncoderDecoderModel.
397
+ All the weights of FlaxSpeechEncoderDecoderModel were initialized from the model checkpoint at ./.
398
+ If your task is similar to the task the model of the checkpoint was trained on, you can already use FlaxSpeechEncoderDecoderModel for predictions without further training.
399
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-eb8e77ddeb6c34dc.arrow
400
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e19eb46cb05f7b14.arrow
401
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-92685008e5c09846.arrow
402
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-0d478bd8b093039a.arrow
403
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-13aa6d940e98be09.arrow
404
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-2d24e62de045a0c8.arrow
405
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-c6e972d87654db26.arrow
406
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-4771f1765b989007.arrow
407
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ab9a2fefe02761db.arrow
408
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-6ce32d0e344f4c06.arrow
409
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-1e0f4ae48a3c8b80.arrow
410
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e6f5ab8655734a9f.arrow
411
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-be6a3a97adc5697a.arrow
412
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-239fcccc5b67f7a4.arrow
413
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ff31cd13241c8bf3.arrow
414
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e8bf216b19718031.arrow
415
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-03ee7ff6632282a6.arrow
416
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-bb1b7de42a389f66.arrow
417
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ed92b3340af62815.arrow
418
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-12ce4b2b8350532d.arrow
419
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-75f20b5cb7b89a9d.arrow
420
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-69941cf3a70dde04.arrow
421
+ 04/04/2022 18:11:23 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-f4d0102728859e5f.arrow
422
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-1c0eaf46e759e04d.arrow
423
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-9f8c4dccc4fa8f17.arrow
424
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-14df38b848e61b52.arrow
425
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-19c662ecc5b20d66.arrow
426
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-3276de10dacb4963.arrow
427
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-111057e4500c4cc5.arrow
428
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-d1bf30028d638290.arrow
429
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cc50c08263e878fc.arrow
430
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-f1813a8ae49a67d6.arrow
431
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00000_of_00016.arrow
432
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00001_of_00016.arrow
433
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00002_of_00016.arrow
434
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00003_of_00016.arrow
435
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00004_of_00016.arrow
436
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00005_of_00016.arrow
437
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00006_of_00016.arrow
438
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00007_of_00016.arrow
439
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00008_of_00016.arrow
440
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00009_of_00016.arrow
441
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00010_of_00016.arrow
442
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00011_of_00016.arrow
443
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00012_of_00016.arrow
444
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00013_of_00016.arrow
445
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00014_of_00016.arrow
446
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00015_of_00016.arrow
447
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00000_of_00016.arrow
448
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00001_of_00016.arrow
449
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00002_of_00016.arrow
450
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00003_of_00016.arrow
451
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00004_of_00016.arrow
452
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00005_of_00016.arrow
453
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00006_of_00016.arrow
454
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00007_of_00016.arrow
455
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00008_of_00016.arrow
456
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00009_of_00016.arrow
457
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00010_of_00016.arrow
458
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00011_of_00016.arrow
459
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00012_of_00016.arrow
460
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00013_of_00016.arrow
461
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00014_of_00016.arrow
462
+ 04/04/2022 18:11:24 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00015_of_00016.arrow
463
+ 04/04/2022 18:11:25 - INFO - __main__ - ***** Running training *****
464
+ 04/04/2022 18:11:25 - INFO - __main__ - Num examples = 0
465
+ 04/04/2022 18:11:25 - INFO - __main__ - Num Epochs = 5
466
+ 04/04/2022 18:11:25 - INFO - __main__ - Instantaneous batch size per device = 2
467
+ 04/04/2022 18:11:25 - INFO - __main__ - Num gradient accumulation steps = 1
468
+ 04/04/2022 18:11:25 - INFO - __main__ - Total train batch size (w. parallel & distributed) = 2
469
+ 04/04/2022 18:11:25 - INFO - __main__ - Total optimization steps = 0
470
+ Feature extractor saved in ./preprocessor_config.json
471
+ tokenizer config file saved in ./tokenizer_config.json
472
+ Special tokens file saved in ./special_tokens_map.json
473
+ Configuration saved in ./config.json
474
+ loading feature extractor configuration file ./preprocessor_config.json
475
+ loading configuration file ./config.json
476
+ Model config SpeechEncoderDecoderConfig {
477
+ "_name_or_path": "./",
478
+ "architectures": [
479
+ "SpeechEncoderDecoderModel"
480
+ ],
481
+ "decoder": {
482
+ "_name_or_path": "facebook/bart-large-cnn",
483
+ "_num_labels": 3,
484
+ "activation_dropout": 0.0,
485
+ "activation_function": "gelu",
486
+ "add_cross_attention": true,
487
+ "add_final_layer_norm": false,
488
+ "architectures": [
489
+ "BartForConditionalGeneration"
490
+ ],
491
+ "attention_dropout": 0.0,
492
+ "bad_words_ids": null,
493
+ "bos_token_id": 0,
494
+ "chunk_size_feed_forward": 0,
495
+ "classif_dropout": 0.0,
496
+ "classifier_dropout": 0.0,
497
+ "cross_attention_hidden_size": null,
498
+ "d_model": 1024,
499
+ "decoder_attention_heads": 16,
500
+ "decoder_ffn_dim": 4096,
501
+ "decoder_layerdrop": 0.0,
502
+ "decoder_layers": 12,
503
+ "decoder_start_token_id": 2,
504
+ "diversity_penalty": 0.0,
505
+ "do_sample": false,
506
+ "dropout": 0.1,
507
+ "early_stopping": true,
508
+ "encoder_attention_heads": 16,
509
+ "encoder_ffn_dim": 4096,
510
+ "encoder_layerdrop": 0.0,
511
+ "encoder_layers": 12,
512
+ "encoder_no_repeat_ngram_size": 0,
513
+ "eos_token_id": 2,
514
+ "exponential_decay_length_penalty": null,
515
+ "finetuning_task": null,
516
+ "force_bos_token_to_be_generated": true,
517
+ "forced_bos_token_id": 0,
518
+ "forced_eos_token_id": 2,
519
+ "gradient_checkpointing": false,
520
+ "id2label": {
521
+ "0": "LABEL_0",
522
+ "1": "LABEL_1",
523
+ "2": "LABEL_2"
524
+ },
525
+ "init_std": 0.02,
526
+ "is_decoder": true,
527
+ "is_encoder_decoder": false,
528
+ "label2id": {
529
+ "LABEL_0": 0,
530
+ "LABEL_1": 1,
531
+ "LABEL_2": 2
532
+ },
533
+ "length_penalty": 2.0,
534
+ "max_length": 142,
535
+ "max_position_embeddings": 1024,
536
+ "min_length": 56,
537
+ "model_type": "bart",
538
+ "no_repeat_ngram_size": 3,
539
+ "normalize_before": false,
540
+ "num_beam_groups": 1,
541
+ "num_beams": 4,
542
+ "num_hidden_layers": 12,
543
+ "num_return_sequences": 1,
544
+ "output_attentions": false,
545
+ "output_hidden_states": false,
546
+ "output_past": true,
547
+ "output_scores": false,
548
+ "pad_token_id": 1,
549
+ "prefix": " ",
550
+ "problem_type": null,
551
+ "pruned_heads": {},
552
+ "remove_invalid_values": false,
553
+ "repetition_penalty": 1.0,
554
+ "return_dict": true,
555
+ "return_dict_in_generate": false,
556
+ "scale_embedding": false,
557
+ "sep_token_id": null,
558
+ "task_specific_params": {
559
+ "summarization": {
560
+ "early_stopping": true,
561
+ "length_penalty": 2.0,
562
+ "max_length": 142,
563
+ "min_length": 56,
564
+ "no_repeat_ngram_size": 3,
565
+ "num_beams": 4
566
+ }
567
+ },
568
+ "temperature": 1.0,
569
+ "tie_encoder_decoder": false,
570
+ "tie_word_embeddings": true,
571
+ "tokenizer_class": null,
572
+ "top_k": 50,
573
+ "top_p": 1.0,
574
+ "torch_dtype": null,
575
+ "torchscript": false,
576
+ "transformers_version": "4.18.0.dev0",
577
+ "typical_p": 1.0,
578
+ "use_bfloat16": false,
579
+ "use_cache": true,
580
+ "vocab_size": 50264
581
+ },
582
+ "decoder_start_token_id": 0,
583
+ "encoder": {
584
+ "_name_or_path": "facebook/wav2vec2-large-lv60",
585
+ "activation_dropout": 0.1,
586
+ "adapter_kernel_size": 3,
587
+ "adapter_stride": 2,
588
+ "add_adapter": true,
589
+ "add_cross_attention": false,
590
+ "apply_spec_augment": true,
591
+ "architectures": [
592
+ "Wav2Vec2ForPreTraining"
593
+ ],
594
+ "attention_dropout": 0.1,
595
+ "bad_words_ids": null,
596
+ "bos_token_id": 1,
597
+ "chunk_size_feed_forward": 0,
598
+ "classifier_proj_size": 256,
599
+ "codevector_dim": 768,
600
+ "contrastive_logits_temperature": 0.1,
601
+ "conv_bias": true,
602
+ "conv_dim": [
603
+ 512,
604
+ 512,
605
+ 512,
606
+ 512,
607
+ 512,
608
+ 512,
609
+ 512
610
+ ],
611
+ "conv_kernel": [
612
+ 10,
613
+ 3,
614
+ 3,
615
+ 3,
616
+ 3,
617
+ 2,
618
+ 2
619
+ ],
620
+ "conv_stride": [
621
+ 5,
622
+ 2,
623
+ 2,
624
+ 2,
625
+ 2,
626
+ 2,
627
+ 2
628
+ ],
629
+ "cross_attention_hidden_size": null,
630
+ "ctc_loss_reduction": "sum",
631
+ "ctc_zero_infinity": false,
632
+ "decoder_start_token_id": null,
633
+ "diversity_loss_weight": 0.1,
634
+ "diversity_penalty": 0.0,
635
+ "do_sample": false,
636
+ "do_stable_layer_norm": true,
637
+ "early_stopping": false,
638
+ "encoder_no_repeat_ngram_size": 0,
639
+ "eos_token_id": 2,
640
+ "exponential_decay_length_penalty": null,
641
+ "feat_extract_activation": "gelu",
642
+ "feat_extract_dropout": 0.0,
643
+ "feat_extract_norm": "layer",
644
+ "feat_proj_dropout": 0.0,
645
+ "feat_quantizer_dropout": 0.0,
646
+ "final_dropout": 0.0,
647
+ "finetuning_task": null,
648
+ "forced_bos_token_id": null,
649
+ "forced_eos_token_id": null,
650
+ "gradient_checkpointing": false,
651
+ "hidden_act": "gelu",
652
+ "hidden_dropout": 0.1,
653
+ "hidden_dropout_prob": 0.1,
654
+ "hidden_size": 1024,
655
+ "id2label": {
656
+ "0": "LABEL_0",
657
+ "1": "LABEL_1"
658
+ },
659
+ "initializer_range": 0.02,
660
+ "intermediate_size": 4096,
661
+ "is_decoder": false,
662
+ "is_encoder_decoder": false,
663
+ "label2id": {
664
+ "LABEL_0": 0,
665
+ "LABEL_1": 1
666
+ },
667
+ "layer_norm_eps": 1e-05,
668
+ "layerdrop": 0.0,
669
+ "length_penalty": 1.0,
670
+ "mask_feature_length": 10,
671
+ "mask_feature_min_masks": 0,
672
+ "mask_feature_prob": 0.0,
673
+ "mask_time_length": 10,
674
+ "mask_time_min_masks": 2,
675
+ "mask_time_prob": 0.1,
676
+ "max_length": 20,
677
+ "min_length": 0,
678
+ "model_type": "wav2vec2",
679
+ "no_repeat_ngram_size": 0,
680
+ "num_adapter_layers": 3,
681
+ "num_attention_heads": 16,
682
+ "num_beam_groups": 1,
683
+ "num_beams": 1,
684
+ "num_codevector_groups": 2,
685
+ "num_codevectors_per_group": 320,
686
+ "num_conv_pos_embedding_groups": 16,
687
+ "num_conv_pos_embeddings": 128,
688
+ "num_feat_extract_layers": 7,
689
+ "num_hidden_layers": 24,
690
+ "num_negatives": 100,
691
+ "num_return_sequences": 1,
692
+ "output_attentions": false,
693
+ "output_hidden_size": 1024,
694
+ "output_hidden_states": false,
695
+ "output_scores": false,
696
+ "pad_token_id": 0,
697
+ "prefix": null,
698
+ "problem_type": null,
699
+ "proj_codevector_dim": 768,
700
+ "pruned_heads": {},
701
+ "remove_invalid_values": false,
702
+ "repetition_penalty": 1.0,
703
+ "return_dict": true,
704
+ "return_dict_in_generate": false,
705
+ "sep_token_id": null,
706
+ "task_specific_params": null,
707
+ "tdnn_dilation": [
708
+ 1,
709
+ 2,
710
+ 3,
711
+ 1,
712
+ 1
713
+ ],
714
+ "tdnn_dim": [
715
+ 512,
716
+ 512,
717
+ 512,
718
+ 512,
719
+ 1500
720
+ ],
721
+ "tdnn_kernel": [
722
+ 5,
723
+ 3,
724
+ 3,
725
+ 1,
726
+ 1
727
+ ],
728
+ "temperature": 1.0,
729
+ "tie_encoder_decoder": false,
730
+ "tie_word_embeddings": true,
731
+ "tokenizer_class": null,
732
+ "top_k": 50,
733
+ "top_p": 1.0,
734
+ "torch_dtype": null,
735
+ "torchscript": false,
736
+ "transformers_version": "4.18.0.dev0",
737
+ "typical_p": 1.0,
738
+ "use_bfloat16": false,
739
+ "use_weighted_layer_sum": false,
740
+ "vocab_size": 32,
741
+ "xvector_output_dim": 512
742
+ },
743
+ "eos_token_id": 2,
744
+ "is_encoder_decoder": true,
745
+ "max_length": 40,
746
+ "model_type": "speech-encoder-decoder",
747
+ "pad_token_id": 1,
748
+ "processor_class": "Wav2Vec2Processor",
749
+ "tie_word_embeddings": false,
750
+ "transformers_version": null,
751
+ "use_cache": false
752
+ }
753
+ loading feature extractor configuration file ./preprocessor_config.json
754
+ Feature extractor Wav2Vec2FeatureExtractor {
755
+ "do_normalize": true,
756
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
757
+ "feature_size": 1,
758
+ "padding_side": "right",
759
+ "padding_value": 0.0,
760
+ "return_attention_mask": true,
761
+ "sampling_rate": 16000
762
+ }
763
+ Didn't find file ./added_tokens.json. We won't load it.
764
+ loading file ./vocab.json
765
+ loading file ./merges.txt
766
+ loading file ./tokenizer.json
767
+ loading file None
768
+ loading file ./special_tokens_map.json
769
+ loading file ./tokenizer_config.json
770
+ 2022-04-04 18:11:24.732185: W tensorflow/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libcuda.so.1'; dlerror: libcuda.so.1: cannot open shared object file: No such file or directory
771
+ 2022-04-04 18:11:24.732232: W tensorflow/stream_executor/cuda/cuda_driver.cc:269] failed call to cuInit: UNKNOWN ERROR (303)
772
+ Epoch ... (1/5): 0%| | 0/5 [00:00<?, ?it/s]
773
+ Traceback (most recent call last):
774
+ File "run_flax_speech_recognition_seq2seq.py", line 1133, in <module>
775
+ main()
776
+ File "run_flax_speech_recognition_seq2seq.py", line 1050, in main
777
+ train_samples_idx = get_grouped_indices(vectorized_datasets["train"], batch_size_per_update, input_rng)
778
+ File "run_flax_speech_recognition_seq2seq.py", line 453, in get_grouped_indices
779
+ lengths = dataset["input_length"]
780
+ File "/home/sanchitgandhi/datasets/src/datasets/arrow_dataset.py", line 1765, in __getitem__
781
+ return self._getitem(
782
+ File "/home/sanchitgandhi/datasets/src/datasets/arrow_dataset.py", line 1749, in _getitem
783
+ pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None)
784
+ File "/home/sanchitgandhi/datasets/src/datasets/formatting/formatting.py", line 491, in query_table
785
+ pa_subtable = _query_table_with_indices_mapping(table, key, indices=indices)
786
+ File "/home/sanchitgandhi/datasets/src/datasets/formatting/formatting.py", line 69, in _query_table_with_indices_mapping
787
+ return _query_table(table, indices.column(0).to_pylist())
788
+ File "/home/sanchitgandhi/datasets/src/datasets/table.py", line 343, in column
789
+ return self.table.column(*args, **kwargs)
790
+ File "pyarrow/table.pxi", line 2115, in pyarrow.lib.Table.column
791
+ File "pyarrow/table.pxi", line 2130, in pyarrow.lib.Table._column
792
+ File "pyarrow/array.pxi", line 561, in pyarrow.lib._normalize_index
793
+ IndexError: index out of bounds
wandb/run-20220404_181103-88xgr1fg/files/requirements.txt ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ astunparse==1.6.3
10
+ async-timeout==4.0.2
11
+ attrs==21.4.0
12
+ audioread==2.1.9
13
+ babel==2.9.1
14
+ backcall==0.2.0
15
+ beautifulsoup4==4.10.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.12
21
+ chex==0.1.1
22
+ click==8.1.0
23
+ cycler==0.11.0
24
+ datasets==2.0.1.dev0
25
+ debugpy==1.6.0
26
+ decorator==5.1.1
27
+ defusedxml==0.7.1
28
+ dill==0.3.4
29
+ dm-tree==0.1.6
30
+ docker-pycreds==0.4.0
31
+ entrypoints==0.4
32
+ executing==0.8.3
33
+ filelock==3.6.0
34
+ flatbuffers==2.0
35
+ flax==0.4.1
36
+ fonttools==4.31.2
37
+ frozenlist==1.3.0
38
+ fsspec==2022.2.0
39
+ gast==0.5.3
40
+ gitdb==4.0.9
41
+ gitpython==3.1.27
42
+ google-auth-oauthlib==0.4.6
43
+ google-auth==2.6.2
44
+ google-pasta==0.2.0
45
+ grpcio==1.44.0
46
+ h5py==3.6.0
47
+ huggingface-hub==0.4.0
48
+ idna==3.3
49
+ importlib-metadata==4.11.3
50
+ importlib-resources==5.6.0
51
+ ipdb==0.13.9
52
+ ipykernel==6.10.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.2.0
55
+ jax==0.3.4
56
+ jaxlib==0.3.2
57
+ jedi==0.18.1
58
+ jinja2==3.1.1
59
+ jiwer==2.3.0
60
+ joblib==1.1.0
61
+ json5==0.9.6
62
+ jsonschema==4.4.0
63
+ jupyter-client==7.2.1
64
+ jupyter-core==4.9.2
65
+ jupyter-server==1.16.0
66
+ jupyterlab-pygments==0.1.2
67
+ jupyterlab-server==2.12.0
68
+ jupyterlab==3.3.2
69
+ keras-preprocessing==1.1.2
70
+ keras==2.8.0
71
+ kiwisolver==1.4.2
72
+ libclang==13.0.0
73
+ librosa==0.9.1
74
+ libtpu-nightly==0.1.dev20220315
75
+ llvmlite==0.38.0
76
+ markdown==3.3.6
77
+ markupsafe==2.1.1
78
+ matplotlib-inline==0.1.3
79
+ matplotlib==3.5.1
80
+ mistune==0.8.4
81
+ msgpack==1.0.3
82
+ multidict==6.0.2
83
+ multiprocess==0.70.12.2
84
+ nbclassic==0.3.7
85
+ nbclient==0.5.13
86
+ nbconvert==6.4.5
87
+ nbformat==5.2.0
88
+ nest-asyncio==1.5.4
89
+ notebook-shim==0.1.0
90
+ notebook==6.4.10
91
+ numba==0.55.1
92
+ numpy==1.21.0
93
+ oauthlib==3.2.0
94
+ opt-einsum==3.3.0
95
+ optax==0.1.1
96
+ packaging==21.3
97
+ pandas==1.4.1
98
+ pandocfilters==1.5.0
99
+ parso==0.8.3
100
+ pathtools==0.1.2
101
+ pexpect==4.8.0
102
+ pickleshare==0.7.5
103
+ pillow==9.0.1
104
+ pip==20.0.2
105
+ pkg-resources==0.0.0
106
+ pooch==1.6.0
107
+ prometheus-client==0.13.1
108
+ promise==2.3
109
+ prompt-toolkit==3.0.28
110
+ protobuf==3.19.4
111
+ psutil==5.9.0
112
+ ptyprocess==0.7.0
113
+ pure-eval==0.2.2
114
+ pyarrow==7.0.0
115
+ pyasn1-modules==0.2.8
116
+ pyasn1==0.4.8
117
+ pycparser==2.21
118
+ pygments==2.11.2
119
+ pyparsing==3.0.7
120
+ pyrsistent==0.18.1
121
+ python-dateutil==2.8.2
122
+ python-levenshtein==0.12.2
123
+ pytz==2022.1
124
+ pyyaml==6.0
125
+ pyzmq==22.3.0
126
+ regex==2022.3.15
127
+ requests-oauthlib==1.3.1
128
+ requests==2.27.1
129
+ resampy==0.2.2
130
+ responses==0.18.0
131
+ rsa==4.8
132
+ sacremoses==0.0.49
133
+ scikit-learn==1.0.2
134
+ scipy==1.8.0
135
+ send2trash==1.8.0
136
+ sentry-sdk==1.5.8
137
+ setproctitle==1.2.2
138
+ setuptools==44.0.0
139
+ shortuuid==1.0.8
140
+ six==1.16.0
141
+ smmap==5.0.0
142
+ sniffio==1.2.0
143
+ soundfile==0.10.3.post1
144
+ soupsieve==2.3.1
145
+ stack-data==0.2.0
146
+ tensorboard-data-server==0.6.1
147
+ tensorboard-plugin-wit==1.8.1
148
+ tensorboard==2.8.0
149
+ tensorflow-io-gcs-filesystem==0.24.0
150
+ tensorflow==2.8.0
151
+ termcolor==1.1.0
152
+ terminado==0.13.3
153
+ testpath==0.6.0
154
+ tf-estimator-nightly==2.8.0.dev2021122109
155
+ threadpoolctl==3.1.0
156
+ tokenizers==0.11.6
157
+ toml==0.10.2
158
+ toolz==0.11.2
159
+ torch==1.11.0+cpu
160
+ torchaudio==0.11.0+cpu
161
+ tornado==6.1
162
+ tqdm==4.63.1
163
+ traitlets==5.1.1
164
+ transformers==4.18.0.dev0
165
+ typing-extensions==4.1.1
166
+ urllib3==1.26.9
167
+ wandb==0.12.11
168
+ wcwidth==0.2.5
169
+ webencodings==0.5.1
170
+ websocket-client==1.3.2
171
+ werkzeug==2.1.0
172
+ wheel==0.37.1
173
+ wrapt==1.14.0
174
+ xxhash==3.0.0
175
+ yarl==1.7.2
176
+ yaspin==2.1.0
177
+ zipp==3.7.0
wandb/run-20220404_181103-88xgr1fg/files/wandb-metadata.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-1043-gcp-x86_64-with-glibc2.29",
3
+ "python": "3.8.10",
4
+ "heartbeatAt": "2022-04-04T18:11:06.449423",
5
+ "startedAt": "2022-04-04T18:11:03.361017",
6
+ "docker": null,
7
+ "cpu_count": 96,
8
+ "cuda": null,
9
+ "args": [
10
+ "--overwrite_output_dir",
11
+ "--freeze_feature_encoder",
12
+ "--predict_with_generate",
13
+ "--do_lower_case",
14
+ "--do_train",
15
+ "--do_eval",
16
+ "--dataset_cache_dir=/home/sanchitgandhi/cache/huggingface/datasets",
17
+ "--dataset_config_name=clean",
18
+ "--dataset_name=librispeech_asr",
19
+ "--eval_split_name=validation",
20
+ "--generation_max_length=40",
21
+ "--generation_num_beams=1",
22
+ "--gradient_accumulation_steps=1",
23
+ "--learning_rate=9.383495031304748e-05",
24
+ "--length_column_name=input_length",
25
+ "--logging_steps=1",
26
+ "--matmul_precision=highest",
27
+ "--max_duration_in_seconds=15",
28
+ "--max_target_length=64",
29
+ "--min_duration_in_seconds=15",
30
+ "--model_name_or_path=./",
31
+ "--num_train_epochs=5",
32
+ "--output_dir=./",
33
+ "--per_device_eval_batch_size=2",
34
+ "--per_device_train_batch_size=2",
35
+ "--preprocessing_num_workers=16",
36
+ "--text_column_name=text",
37
+ "--train_split_name=train.100",
38
+ "--wandb_project=flax-wav2vec2-2-bart-large-cnn",
39
+ "--warmup_steps=500"
40
+ ],
41
+ "state": "running",
42
+ "program": "run_flax_speech_recognition_seq2seq.py",
43
+ "codePath": "run_flax_speech_recognition_seq2seq.py",
44
+ "git": {
45
+ "remote": "https://huggingface.co/sanchit-gandhi/flax-wav2vec2-2-bart-large-cnn",
46
+ "commit": "c09faa6b18ac3f19251c414100acf1bad11c1b16"
47
+ },
48
+ "email": "sanchit@huggingface.co",
49
+ "root": "/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn",
50
+ "host": "t1v-n-4eb331dd-w-0",
51
+ "username": "sanchitgandhi",
52
+ "executable": "/home/sanchitgandhi/venv/bin/python3"
53
+ }
wandb/run-20220404_181103-88xgr1fg/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 23}}
wandb/run-20220404_181103-88xgr1fg/logs/debug-internal.log ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-04-04 18:11:04,131 INFO MainThread:1242765 [internal.py:wandb_internal():92] W&B internal server running at pid: 1242765, started at: 2022-04-04 18:11:04.131243
2
+ 2022-04-04 18:11:04,133 INFO WriterThread:1242765 [datastore.py:open_for_write():77] open: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/run-88xgr1fg.wandb
3
+ 2022-04-04 18:11:04,134 DEBUG SenderThread:1242765 [sender.py:send():235] send: header
4
+ 2022-04-04 18:11:04,134 DEBUG SenderThread:1242765 [sender.py:send():235] send: run
5
+ 2022-04-04 18:11:04,223 INFO SenderThread:1242765 [dir_watcher.py:__init__():169] watching files in: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files
6
+ 2022-04-04 18:11:04,223 INFO SenderThread:1242765 [sender.py:_start_run_threads():812] run started: 88xgr1fg with start time 1649095863
7
+ 2022-04-04 18:11:04,223 DEBUG SenderThread:1242765 [sender.py:send():235] send: summary
8
+ 2022-04-04 18:11:04,223 INFO SenderThread:1242765 [sender.py:_save_file():947] saving file wandb-summary.json with policy end
9
+ 2022-04-04 18:11:04,226 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: check_version
10
+ 2022-04-04 18:11:04,226 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: check_version
11
+ 2022-04-04 18:11:04,286 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: run_start
12
+ 2022-04-04 18:11:05,224 INFO Thread-7 :1242765 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/wandb-summary.json
13
+ 2022-04-04 18:11:06,449 DEBUG HandlerThread:1242765 [meta.py:__init__():37] meta init
14
+ 2022-04-04 18:11:06,449 DEBUG HandlerThread:1242765 [meta.py:__init__():51] meta init done
15
+ 2022-04-04 18:11:06,449 DEBUG HandlerThread:1242765 [meta.py:probe():211] probe
16
+ 2022-04-04 18:11:06,450 DEBUG HandlerThread:1242765 [meta.py:_setup_git():201] setup git
17
+ 2022-04-04 18:11:06,481 DEBUG HandlerThread:1242765 [meta.py:_setup_git():208] setup git done
18
+ 2022-04-04 18:11:06,481 DEBUG HandlerThread:1242765 [meta.py:_save_pip():55] save pip
19
+ 2022-04-04 18:11:06,482 DEBUG HandlerThread:1242765 [meta.py:_save_pip():69] save pip done
20
+ 2022-04-04 18:11:06,482 DEBUG HandlerThread:1242765 [meta.py:probe():249] probe done
21
+ 2022-04-04 18:11:06,485 DEBUG SenderThread:1242765 [sender.py:send():235] send: files
22
+ 2022-04-04 18:11:06,485 INFO SenderThread:1242765 [sender.py:_save_file():947] saving file wandb-metadata.json with policy now
23
+ 2022-04-04 18:11:06,491 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: stop_status
24
+ 2022-04-04 18:11:06,491 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: stop_status
25
+ 2022-04-04 18:11:06,537 DEBUG SenderThread:1242765 [sender.py:send():235] send: telemetry
26
+ 2022-04-04 18:11:06,787 INFO Thread-11 :1242765 [upload_job.py:push():137] Uploaded file /tmp/tmpso3thk5xwandb/16y6egpg-wandb-metadata.json
27
+ 2022-04-04 18:11:07,228 INFO Thread-7 :1242765 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/requirements.txt
28
+ 2022-04-04 18:11:07,228 INFO Thread-7 :1242765 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/output.log
29
+ 2022-04-04 18:11:07,228 INFO Thread-7 :1242765 [dir_watcher.py:_on_file_created():217] file/dir created: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/wandb-metadata.json
30
+ 2022-04-04 18:11:09,229 INFO Thread-7 :1242765 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/output.log
31
+ 2022-04-04 18:11:21,798 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: stop_status
32
+ 2022-04-04 18:11:21,799 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: stop_status
33
+ 2022-04-04 18:11:23,235 INFO Thread-7 :1242765 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/output.log
34
+ 2022-04-04 18:11:25,235 INFO Thread-7 :1242765 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/output.log
35
+ 2022-04-04 18:11:25,880 DEBUG SenderThread:1242765 [sender.py:send():235] send: telemetry
36
+ 2022-04-04 18:11:25,881 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: partial_history
37
+ 2022-04-04 18:11:27,236 INFO Thread-7 :1242765 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/output.log
38
+ 2022-04-04 18:11:28,239 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: poll_exit
39
+ 2022-04-04 18:11:28,240 DEBUG SenderThread:1242765 [sender.py:send():235] send: exit
40
+ 2022-04-04 18:11:28,240 INFO SenderThread:1242765 [sender.py:send_exit():371] handling exit code: 1
41
+ 2022-04-04 18:11:28,240 INFO SenderThread:1242765 [sender.py:send_exit():373] handling runtime: 23
42
+ 2022-04-04 18:11:28,240 INFO SenderThread:1242765 [sender.py:_save_file():947] saving file wandb-summary.json with policy end
43
+ 2022-04-04 18:11:28,241 INFO SenderThread:1242765 [sender.py:send_exit():379] send defer
44
+ 2022-04-04 18:11:28,241 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: poll_exit
45
+ 2022-04-04 18:11:28,241 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
46
+ 2022-04-04 18:11:28,241 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 0
47
+ 2022-04-04 18:11:28,241 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
48
+ 2022-04-04 18:11:28,241 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 0
49
+ 2022-04-04 18:11:28,242 INFO SenderThread:1242765 [sender.py:transition_state():392] send defer: 1
50
+ 2022-04-04 18:11:28,242 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
51
+ 2022-04-04 18:11:28,242 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 1
52
+ 2022-04-04 18:11:28,315 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
53
+ 2022-04-04 18:11:28,315 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 1
54
+ 2022-04-04 18:11:28,315 INFO SenderThread:1242765 [sender.py:transition_state():392] send defer: 2
55
+ 2022-04-04 18:11:28,315 DEBUG SenderThread:1242765 [sender.py:send():235] send: stats
56
+ 2022-04-04 18:11:28,315 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
57
+ 2022-04-04 18:11:28,315 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 2
58
+ 2022-04-04 18:11:28,316 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
59
+ 2022-04-04 18:11:28,316 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 2
60
+ 2022-04-04 18:11:28,316 INFO SenderThread:1242765 [sender.py:transition_state():392] send defer: 3
61
+ 2022-04-04 18:11:28,316 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
62
+ 2022-04-04 18:11:28,316 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 3
63
+ 2022-04-04 18:11:28,316 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
64
+ 2022-04-04 18:11:28,316 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 3
65
+ 2022-04-04 18:11:28,316 INFO SenderThread:1242765 [sender.py:transition_state():392] send defer: 4
66
+ 2022-04-04 18:11:28,316 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
67
+ 2022-04-04 18:11:28,316 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 4
68
+ 2022-04-04 18:11:28,316 DEBUG SenderThread:1242765 [sender.py:send():235] send: summary
69
+ 2022-04-04 18:11:28,317 INFO SenderThread:1242765 [sender.py:_save_file():947] saving file wandb-summary.json with policy end
70
+ 2022-04-04 18:11:28,317 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
71
+ 2022-04-04 18:11:28,317 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 4
72
+ 2022-04-04 18:11:28,317 INFO SenderThread:1242765 [sender.py:transition_state():392] send defer: 5
73
+ 2022-04-04 18:11:28,317 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
74
+ 2022-04-04 18:11:28,317 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 5
75
+ 2022-04-04 18:11:28,317 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
76
+ 2022-04-04 18:11:28,317 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 5
77
+ 2022-04-04 18:11:28,342 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: poll_exit
78
+ 2022-04-04 18:11:28,409 INFO SenderThread:1242765 [sender.py:transition_state():392] send defer: 6
79
+ 2022-04-04 18:11:28,409 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: poll_exit
80
+ 2022-04-04 18:11:28,409 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
81
+ 2022-04-04 18:11:28,410 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 6
82
+ 2022-04-04 18:11:28,410 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
83
+ 2022-04-04 18:11:28,410 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 6
84
+ 2022-04-04 18:11:28,410 INFO SenderThread:1242765 [dir_watcher.py:finish():283] shutting down directory watcher
85
+ 2022-04-04 18:11:28,510 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: poll_exit
86
+ 2022-04-04 18:11:29,237 INFO Thread-7 :1242765 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/config.yaml
87
+ 2022-04-04 18:11:29,237 INFO SenderThread:1242765 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/output.log
88
+ 2022-04-04 18:11:29,238 INFO SenderThread:1242765 [dir_watcher.py:_on_file_modified():230] file/dir modified: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/wandb-summary.json
89
+ 2022-04-04 18:11:29,238 INFO SenderThread:1242765 [dir_watcher.py:finish():313] scan: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files
90
+ 2022-04-04 18:11:29,238 INFO SenderThread:1242765 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/requirements.txt requirements.txt
91
+ 2022-04-04 18:11:29,238 INFO SenderThread:1242765 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/output.log output.log
92
+ 2022-04-04 18:11:29,238 INFO SenderThread:1242765 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/wandb-metadata.json wandb-metadata.json
93
+ 2022-04-04 18:11:29,238 INFO SenderThread:1242765 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/config.yaml config.yaml
94
+ 2022-04-04 18:11:29,238 INFO SenderThread:1242765 [dir_watcher.py:finish():327] scan save: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/wandb-summary.json wandb-summary.json
95
+ 2022-04-04 18:11:29,239 INFO SenderThread:1242765 [sender.py:transition_state():392] send defer: 7
96
+ 2022-04-04 18:11:29,239 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: poll_exit
97
+ 2022-04-04 18:11:29,244 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
98
+ 2022-04-04 18:11:29,244 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 7
99
+ 2022-04-04 18:11:29,247 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
100
+ 2022-04-04 18:11:29,247 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 7
101
+ 2022-04-04 18:11:29,247 INFO SenderThread:1242765 [file_pusher.py:finish():145] shutting down file pusher
102
+ 2022-04-04 18:11:29,348 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: poll_exit
103
+ 2022-04-04 18:11:29,348 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: poll_exit
104
+ 2022-04-04 18:11:29,450 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: poll_exit
105
+ 2022-04-04 18:11:29,450 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: poll_exit
106
+ 2022-04-04 18:11:29,477 INFO Thread-15 :1242765 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/wandb-summary.json
107
+ 2022-04-04 18:11:29,510 INFO Thread-13 :1242765 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/output.log
108
+ 2022-04-04 18:11:29,522 INFO Thread-12 :1242765 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/requirements.txt
109
+ 2022-04-04 18:11:29,523 INFO Thread-14 :1242765 [upload_job.py:push():137] Uploaded file /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/files/config.yaml
110
+ 2022-04-04 18:11:29,551 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: poll_exit
111
+ 2022-04-04 18:11:29,551 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: poll_exit
112
+ 2022-04-04 18:11:29,652 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: poll_exit
113
+ 2022-04-04 18:11:29,653 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: poll_exit
114
+ 2022-04-04 18:11:29,724 INFO Thread-6 :1242765 [sender.py:transition_state():392] send defer: 8
115
+ 2022-04-04 18:11:29,724 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
116
+ 2022-04-04 18:11:29,724 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 8
117
+ 2022-04-04 18:11:29,724 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
118
+ 2022-04-04 18:11:29,724 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 8
119
+ 2022-04-04 18:11:29,754 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: poll_exit
120
+ 2022-04-04 18:11:29,830 INFO SenderThread:1242765 [sender.py:transition_state():392] send defer: 9
121
+ 2022-04-04 18:11:29,830 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: poll_exit
122
+ 2022-04-04 18:11:29,830 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
123
+ 2022-04-04 18:11:29,830 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 9
124
+ 2022-04-04 18:11:29,831 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
125
+ 2022-04-04 18:11:29,831 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 9
126
+ 2022-04-04 18:11:29,831 INFO SenderThread:1242765 [sender.py:transition_state():392] send defer: 10
127
+ 2022-04-04 18:11:29,831 DEBUG SenderThread:1242765 [sender.py:send():235] send: final
128
+ 2022-04-04 18:11:29,831 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: defer
129
+ 2022-04-04 18:11:29,831 INFO HandlerThread:1242765 [handler.py:handle_request_defer():164] handle defer: 10
130
+ 2022-04-04 18:11:29,832 DEBUG SenderThread:1242765 [sender.py:send():235] send: footer
131
+ 2022-04-04 18:11:29,832 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: defer
132
+ 2022-04-04 18:11:29,832 INFO SenderThread:1242765 [sender.py:send_request_defer():388] handle sender defer: 10
133
+ 2022-04-04 18:11:29,931 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: poll_exit
134
+ 2022-04-04 18:11:29,931 DEBUG SenderThread:1242765 [sender.py:send_request():249] send_request: poll_exit
135
+ 2022-04-04 18:11:29,931 INFO SenderThread:1242765 [file_pusher.py:join():150] waiting for file pusher
136
+ 2022-04-04 18:11:30,799 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: sampled_history
137
+ 2022-04-04 18:11:30,799 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: get_summary
138
+ 2022-04-04 18:11:30,800 DEBUG HandlerThread:1242765 [handler.py:handle_request():141] handle_request: shutdown
139
+ 2022-04-04 18:11:30,800 INFO HandlerThread:1242765 [handler.py:finish():778] shutting down handler
140
+ 2022-04-04 18:11:30,832 INFO WriterThread:1242765 [datastore.py:close():281] close: /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/run-88xgr1fg.wandb
141
+ 2022-04-04 18:11:31,698 INFO SenderThread:1242765 [sender.py:finish():1078] shutting down sender
142
+ 2022-04-04 18:11:31,698 INFO SenderThread:1242765 [file_pusher.py:finish():145] shutting down file pusher
143
+ 2022-04-04 18:11:31,698 INFO SenderThread:1242765 [file_pusher.py:join():150] waiting for file pusher
144
+ 2022-04-04 18:11:31,700 INFO MainThread:1242765 [internal.py:handle_exit():82] Internal process exited
wandb/run-20220404_181103-88xgr1fg/logs/debug.log ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-04-04 18:11:03,362 INFO MainThread:1242310 [wandb_setup.py:_flush():75] Loading settings from /home/sanchitgandhi/.config/wandb/settings
2
+ 2022-04-04 18:11:03,362 INFO MainThread:1242310 [wandb_setup.py:_flush():75] Loading settings from wandb/settings
3
+ 2022-04-04 18:11:03,362 INFO MainThread:1242310 [wandb_setup.py:_flush():75] Loading settings from environment variables: {'entity': 'sanchit-gandhi', 'project': 'flax-wav2vec2-2-bart-large-cnn', 'sweep_id': 'a50sny4c', 'root_dir': '/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn', 'run_id': '88xgr1fg', 'sweep_param_path': '/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/sweep-a50sny4c/config-88xgr1fg.yaml'}
4
+ 2022-04-04 18:11:03,362 INFO MainThread:1242310 [wandb_setup.py:_flush():75] Inferring run settings from compute environment: {'program_relpath': 'run_flax_speech_recognition_seq2seq.py', 'program': 'run_flax_speech_recognition_seq2seq.py'}
5
+ 2022-04-04 18:11:03,362 INFO MainThread:1242310 [wandb_init.py:_log_setup():405] Logging user logs to /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/logs/debug.log
6
+ 2022-04-04 18:11:03,362 INFO MainThread:1242310 [wandb_init.py:_log_setup():406] Logging internal logs to /home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn/wandb/run-20220404_181103-88xgr1fg/logs/debug-internal.log
7
+ 2022-04-04 18:11:03,362 INFO MainThread:1242310 [wandb_init.py:init():439] calling init triggers
8
+ 2022-04-04 18:11:03,362 INFO MainThread:1242310 [wandb_init.py:init():442] wandb.init called with sweep_config: {'dataset_cache_dir': '/home/sanchitgandhi/cache/huggingface/datasets', 'dataset_config_name': 'clean', 'dataset_name': 'librispeech_asr', 'eval_split_name': 'validation', 'generation_max_length': 40, 'generation_num_beams': 1, 'gradient_accumulation_steps': 1, 'learning_rate': 9.383495031304748e-05, 'length_column_name': 'input_length', 'logging_steps': 1, 'matmul_precision': 'highest', 'max_duration_in_seconds': 15, 'max_target_length': 64, 'min_duration_in_seconds': 15, 'model_name_or_path': './', 'num_train_epochs': 5, 'output_dir': './', 'per_device_eval_batch_size': 2, 'per_device_train_batch_size': 2, 'preprocessing_num_workers': 16, 'text_column_name': 'text', 'train_split_name': 'train.100', 'wandb_project': 'flax-wav2vec2-2-bart-large-cnn', 'warmup_steps': 500}
9
+ config: {}
10
+ 2022-04-04 18:11:03,362 INFO MainThread:1242310 [wandb_init.py:init():492] starting backend
11
+ 2022-04-04 18:11:03,363 INFO MainThread:1242310 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
12
+ 2022-04-04 18:11:03,375 INFO MainThread:1242310 [backend.py:ensure_launched():219] starting backend process...
13
+ 2022-04-04 18:11:03,383 INFO MainThread:1242310 [backend.py:ensure_launched():224] started backend process with pid: 1242765
14
+ 2022-04-04 18:11:03,384 INFO MainThread:1242310 [wandb_init.py:init():501] backend started and connected
15
+ 2022-04-04 18:11:03,391 INFO MainThread:1242310 [wandb_run.py:_config_callback():992] config_cb None None {'dataset_cache_dir': '/home/sanchitgandhi/cache/huggingface/datasets', 'dataset_config_name': 'clean', 'dataset_name': 'librispeech_asr', 'eval_split_name': 'validation', 'generation_max_length': 40, 'generation_num_beams': 1, 'gradient_accumulation_steps': 1, 'learning_rate': 9.383495031304748e-05, 'length_column_name': 'input_length', 'logging_steps': 1, 'matmul_precision': 'highest', 'max_duration_in_seconds': 15, 'max_target_length': 64, 'min_duration_in_seconds': 15, 'model_name_or_path': './', 'num_train_epochs': 5, 'output_dir': './', 'per_device_eval_batch_size': 2, 'per_device_train_batch_size': 2, 'preprocessing_num_workers': 16, 'text_column_name': 'text', 'train_split_name': 'train.100', 'wandb_project': 'flax-wav2vec2-2-bart-large-cnn', 'warmup_steps': 500}
16
+ 2022-04-04 18:11:03,394 INFO MainThread:1242310 [wandb_init.py:init():565] updated telemetry
17
+ 2022-04-04 18:11:03,414 INFO MainThread:1242310 [wandb_init.py:init():596] communicating run to backend with 30 second timeout
18
+ 2022-04-04 18:11:04,226 INFO MainThread:1242310 [wandb_run.py:_on_init():1759] communicating current version
19
+ 2022-04-04 18:11:04,285 INFO MainThread:1242310 [wandb_run.py:_on_init():1763] got version response
20
+ 2022-04-04 18:11:04,285 INFO MainThread:1242310 [wandb_init.py:init():625] starting run threads in backend
21
+ 2022-04-04 18:11:06,491 INFO MainThread:1242310 [wandb_run.py:_console_start():1733] atexit reg
22
+ 2022-04-04 18:11:06,491 INFO MainThread:1242310 [wandb_run.py:_redirect():1606] redirect: SettingsConsole.REDIRECT
23
+ 2022-04-04 18:11:06,492 INFO MainThread:1242310 [wandb_run.py:_redirect():1611] Redirecting console.
24
+ 2022-04-04 18:11:06,494 INFO MainThread:1242310 [wandb_run.py:_redirect():1667] Redirects installed.
25
+ 2022-04-04 18:11:06,494 INFO MainThread:1242310 [wandb_init.py:init():664] run started, returning control to user process
26
+ 2022-04-04 18:11:25,876 INFO MainThread:1242310 [wandb_run.py:_atexit_cleanup():1702] got exitcode: 1
27
+ 2022-04-04 18:11:25,880 INFO MainThread:1242310 [wandb_run.py:_restore():1674] restore
28
+ 2022-04-04 18:11:28,241 INFO MainThread:1242310 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
29
+ wandb_count: 1
30
+ }
31
+ pusher_stats {
32
+ uploaded_bytes: 1948
33
+ total_bytes: 1948
34
+ }
35
+
36
+ 2022-04-04 18:11:28,410 INFO MainThread:1242310 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
37
+ wandb_count: 1
38
+ }
39
+ pusher_stats {
40
+ uploaded_bytes: 1948
41
+ total_bytes: 1948
42
+ }
43
+
44
+ 2022-04-04 18:11:29,247 INFO MainThread:1242310 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
45
+ wandb_count: 4
46
+ }
47
+ pusher_stats {
48
+ uploaded_bytes: 1948
49
+ total_bytes: 45265
50
+ }
51
+
52
+ 2022-04-04 18:11:29,349 INFO MainThread:1242310 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
53
+ wandb_count: 5
54
+ }
55
+ pusher_stats {
56
+ uploaded_bytes: 45292
57
+ total_bytes: 45292
58
+ }
59
+
60
+ 2022-04-04 18:11:29,450 INFO MainThread:1242310 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
61
+ wandb_count: 5
62
+ }
63
+ pusher_stats {
64
+ uploaded_bytes: 45292
65
+ total_bytes: 45292
66
+ }
67
+
68
+ 2022-04-04 18:11:29,552 INFO MainThread:1242310 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
69
+ wandb_count: 5
70
+ }
71
+ pusher_stats {
72
+ uploaded_bytes: 45292
73
+ total_bytes: 45292
74
+ }
75
+
76
+ 2022-04-04 18:11:29,653 INFO MainThread:1242310 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
77
+ wandb_count: 5
78
+ }
79
+ pusher_stats {
80
+ uploaded_bytes: 45292
81
+ total_bytes: 45292
82
+ }
83
+
84
+ 2022-04-04 18:11:29,831 INFO MainThread:1242310 [wandb_run.py:_on_finish():1831] got exit ret: file_counts {
85
+ wandb_count: 5
86
+ }
87
+ pusher_stats {
88
+ uploaded_bytes: 45292
89
+ total_bytes: 45292
90
+ }
91
+
92
+ 2022-04-04 18:11:30,698 INFO MainThread:1242310 [wandb_run.py:_on_finish():1831] got exit ret: done: true
93
+ exit_result {
94
+ }
95
+ file_counts {
96
+ wandb_count: 5
97
+ }
98
+ pusher_stats {
99
+ uploaded_bytes: 45292
100
+ total_bytes: 45292
101
+ }
102
+ local_info {
103
+ }
104
+
105
+ 2022-04-04 18:11:32,137 INFO MainThread:1242310 [wandb_run.py:_footer_history_summary_info():2865] rendering history
106
+ 2022-04-04 18:11:32,137 INFO MainThread:1242310 [wandb_run.py:_footer_history_summary_info():2894] rendering summary
107
+ 2022-04-04 18:11:32,139 INFO MainThread:1242310 [wandb_run.py:_footer_sync_info():2822] logging synced files
wandb/run-20220404_181103-88xgr1fg/run-88xgr1fg.wandb ADDED
Binary file (40.3 kB). View file
 
wandb/run-20220404_181139-xmgtui21/files/config.yaml ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ cli_version: 0.12.11
7
+ framework: huggingface
8
+ huggingface_version: 4.18.0.dev0
9
+ is_jupyter_run: false
10
+ is_kaggle_kernel: false
11
+ python_version: 3.8.10
12
+ start_time: 1649095899
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 11
19
+ - 12
20
+ - 45
21
+ - 49
22
+ - 51
23
+ 2:
24
+ - 1
25
+ - 2
26
+ - 3
27
+ - 11
28
+ - 12
29
+ - 45
30
+ - 49
31
+ - 51
32
+ 4: 3.8.10
33
+ 5: 0.12.11
34
+ 6: 4.18.0.dev0
35
+ 8:
36
+ - 5
37
+ dataset_cache_dir:
38
+ desc: null
39
+ value: /home/sanchitgandhi/cache/huggingface/datasets
40
+ dataset_config_name:
41
+ desc: null
42
+ value: clean
43
+ dataset_name:
44
+ desc: null
45
+ value: librispeech_asr
46
+ eval_split_name:
47
+ desc: null
48
+ value: validation
49
+ generation_max_length:
50
+ desc: null
51
+ value: 40
52
+ generation_num_beams:
53
+ desc: null
54
+ value: 1
55
+ gradient_accumulation_steps:
56
+ desc: null
57
+ value: 1
58
+ learning_rate:
59
+ desc: null
60
+ value: 7.331199736432637e-05
61
+ length_column_name:
62
+ desc: null
63
+ value: input_length
64
+ logging_steps:
65
+ desc: null
66
+ value: 1
67
+ matmul_precision:
68
+ desc: null
69
+ value: highest
70
+ max_duration_in_seconds:
71
+ desc: null
72
+ value: 15
73
+ max_target_length:
74
+ desc: null
75
+ value: 64
76
+ min_duration_in_seconds:
77
+ desc: null
78
+ value: 15
79
+ model_name_or_path:
80
+ desc: null
81
+ value: ./
82
+ num_train_epochs:
83
+ desc: null
84
+ value: 5
85
+ output_dir:
86
+ desc: null
87
+ value: ./
88
+ per_device_eval_batch_size:
89
+ desc: null
90
+ value: 2
91
+ per_device_train_batch_size:
92
+ desc: null
93
+ value: 2
94
+ preprocessing_num_workers:
95
+ desc: null
96
+ value: 16
97
+ text_column_name:
98
+ desc: null
99
+ value: text
100
+ train_split_name:
101
+ desc: null
102
+ value: train.100
103
+ wandb_project:
104
+ desc: null
105
+ value: flax-wav2vec2-2-bart-large-cnn
106
+ warmup_steps:
107
+ desc: null
108
+ value: 500
wandb/run-20220404_181139-xmgtui21/files/output.log ADDED
@@ -0,0 +1,793 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 04/04/2022 18:11:42 - INFO - __main__ - Training/evaluation parameters FlaxSeq2SeqTrainingArguments(
2
+ _n_gpu=0,
3
+ adafactor=False,
4
+ adam_beta1=0.9,
5
+ adam_beta2=0.999,
6
+ adam_epsilon=1e-08,
7
+ bf16=False,
8
+ bf16_full_eval=False,
9
+ data_seed=None,
10
+ dataloader_drop_last=False,
11
+ dataloader_num_workers=0,
12
+ dataloader_pin_memory=True,
13
+ ddp_bucket_cap_mb=None,
14
+ ddp_find_unused_parameters=None,
15
+ debug=[],
16
+ deepspeed=None,
17
+ disable_tqdm=False,
18
+ do_eval=True,
19
+ do_predict=False,
20
+ do_train=True,
21
+ eval_accumulation_steps=None,
22
+ eval_delay=0,
23
+ eval_steps=None,
24
+ evaluation_strategy=IntervalStrategy.NO,
25
+ fp16=False,
26
+ fp16_backend=auto,
27
+ fp16_full_eval=False,
28
+ fp16_opt_level=O1,
29
+ generation_max_length=40,
30
+ generation_num_beams=1,
31
+ gradient_accumulation_steps=1,
32
+ gradient_checkpointing=False,
33
+ greater_is_better=None,
34
+ group_by_length=False,
35
+ half_precision_backend=auto,
36
+ hub_model_id=None,
37
+ hub_strategy=HubStrategy.EVERY_SAVE,
38
+ hub_token=<HUB_TOKEN>,
39
+ ignore_data_skip=False,
40
+ label_names=None,
41
+ label_smoothing_factor=0.0,
42
+ learning_rate=7.331199736432637e-05,
43
+ length_column_name=input_length,
44
+ load_best_model_at_end=False,
45
+ local_rank=-1,
46
+ log_level=-1,
47
+ log_level_replica=-1,
48
+ log_on_each_node=True,
49
+ logging_dir=./runs/Apr04_18-11-39_t1v-n-4eb331dd-w-0,
50
+ logging_first_step=False,
51
+ logging_nan_inf_filter=True,
52
+ logging_steps=1,
53
+ logging_strategy=IntervalStrategy.STEPS,
54
+ lr_scheduler_type=SchedulerType.LINEAR,
55
+ matmul_precision=highest,
56
+ max_grad_norm=1.0,
57
+ max_steps=-1,
58
+ metric_for_best_model=None,
59
+ mixed_precision=False,
60
+ mp_parameters=,
61
+ no_cuda=False,
62
+ num_train_epochs=5.0,
63
+ optim=OptimizerNames.ADAMW_HF,
64
+ output_dir=./,
65
+ overwrite_output_dir=True,
66
+ past_index=-1,
67
+ per_device_eval_batch_size=2,
68
+ per_device_train_batch_size=2,
69
+ predict_with_generate=True,
70
+ prediction_loss_only=False,
71
+ push_to_hub=False,
72
+ push_to_hub_model_id=None,
73
+ push_to_hub_organization=None,
74
+ push_to_hub_token=<PUSH_TO_HUB_TOKEN>,
75
+ remove_unused_columns=True,
76
+ report_to=['tensorboard', 'wandb'],
77
+ resume_from_checkpoint=None,
78
+ run_name=./,
79
+ save_on_each_node=False,
80
+ save_steps=500,
81
+ save_strategy=IntervalStrategy.STEPS,
82
+ save_total_limit=None,
83
+ seed=42,
84
+ sharded_ddp=[],
85
+ skip_memory_metrics=True,
86
+ sortish_sampler=False,
87
+ tf32=None,
88
+ tpu_metrics_debug=False,
89
+ tpu_num_cores=None,
90
+ use_legacy_prediction_loop=False,
91
+ warmup_ratio=0.0,
92
+ warmup_steps=500,
93
+ weight_decay=0.0,
94
+ xpu_backend=None,
95
+ )
96
+ 04/04/2022 18:11:42 - INFO - __main__ - JAX devices: 1, matmul precision: highest
97
+ 04/04/2022 18:11:42 - WARNING - datasets.builder - Reusing dataset librispeech_asr (/home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c)
98
+ 04/04/2022 18:11:42 - WARNING - datasets.builder - Reusing dataset librispeech_asr (/home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c)
99
+ loading configuration file ./config.json
100
+ Model config SpeechEncoderDecoderConfig {
101
+ "_name_or_path": "./",
102
+ "architectures": [
103
+ "SpeechEncoderDecoderModel"
104
+ ],
105
+ "decoder": {
106
+ "_name_or_path": "facebook/bart-large-cnn",
107
+ "_num_labels": 3,
108
+ "activation_dropout": 0.0,
109
+ "activation_function": "gelu",
110
+ "add_cross_attention": true,
111
+ "add_final_layer_norm": false,
112
+ "architectures": [
113
+ "BartForConditionalGeneration"
114
+ ],
115
+ "attention_dropout": 0.0,
116
+ "bad_words_ids": null,
117
+ "bos_token_id": 0,
118
+ "chunk_size_feed_forward": 0,
119
+ "classif_dropout": 0.0,
120
+ "classifier_dropout": 0.0,
121
+ "cross_attention_hidden_size": null,
122
+ "d_model": 1024,
123
+ "decoder_attention_heads": 16,
124
+ "decoder_ffn_dim": 4096,
125
+ "decoder_layerdrop": 0.0,
126
+ "decoder_layers": 12,
127
+ "decoder_start_token_id": 2,
128
+ "diversity_penalty": 0.0,
129
+ "do_sample": false,
130
+ "dropout": 0.1,
131
+ "early_stopping": true,
132
+ "encoder_attention_heads": 16,
133
+ "encoder_ffn_dim": 4096,
134
+ "encoder_layerdrop": 0.0,
135
+ "encoder_layers": 12,
136
+ "encoder_no_repeat_ngram_size": 0,
137
+ "eos_token_id": 2,
138
+ "exponential_decay_length_penalty": null,
139
+ "finetuning_task": null,
140
+ "force_bos_token_to_be_generated": true,
141
+ "forced_bos_token_id": 0,
142
+ "forced_eos_token_id": 2,
143
+ "gradient_checkpointing": false,
144
+ "id2label": {
145
+ "0": "LABEL_0",
146
+ "1": "LABEL_1",
147
+ "2": "LABEL_2"
148
+ },
149
+ "init_std": 0.02,
150
+ "is_decoder": true,
151
+ "is_encoder_decoder": false,
152
+ "label2id": {
153
+ "LABEL_0": 0,
154
+ "LABEL_1": 1,
155
+ "LABEL_2": 2
156
+ },
157
+ "length_penalty": 2.0,
158
+ "max_length": 142,
159
+ "max_position_embeddings": 1024,
160
+ "min_length": 56,
161
+ "model_type": "bart",
162
+ "no_repeat_ngram_size": 3,
163
+ "normalize_before": false,
164
+ "num_beam_groups": 1,
165
+ "num_beams": 4,
166
+ "num_hidden_layers": 12,
167
+ "num_return_sequences": 1,
168
+ "output_attentions": false,
169
+ "output_hidden_states": false,
170
+ "output_past": true,
171
+ "output_scores": false,
172
+ "pad_token_id": 1,
173
+ "prefix": " ",
174
+ "problem_type": null,
175
+ "pruned_heads": {},
176
+ "remove_invalid_values": false,
177
+ "repetition_penalty": 1.0,
178
+ "return_dict": true,
179
+ "return_dict_in_generate": false,
180
+ "scale_embedding": false,
181
+ "sep_token_id": null,
182
+ "task_specific_params": {
183
+ "summarization": {
184
+ "early_stopping": true,
185
+ "length_penalty": 2.0,
186
+ "max_length": 142,
187
+ "min_length": 56,
188
+ "no_repeat_ngram_size": 3,
189
+ "num_beams": 4
190
+ }
191
+ },
192
+ "temperature": 1.0,
193
+ "tie_encoder_decoder": false,
194
+ "tie_word_embeddings": true,
195
+ "tokenizer_class": null,
196
+ "top_k": 50,
197
+ "top_p": 1.0,
198
+ "torch_dtype": null,
199
+ "torchscript": false,
200
+ "transformers_version": "4.18.0.dev0",
201
+ "typical_p": 1.0,
202
+ "use_bfloat16": false,
203
+ "use_cache": true,
204
+ "vocab_size": 50264
205
+ },
206
+ "decoder_start_token_id": 0,
207
+ "encoder": {
208
+ "_name_or_path": "facebook/wav2vec2-large-lv60",
209
+ "activation_dropout": 0.1,
210
+ "adapter_kernel_size": 3,
211
+ "adapter_stride": 2,
212
+ "add_adapter": true,
213
+ "add_cross_attention": false,
214
+ "apply_spec_augment": true,
215
+ "architectures": [
216
+ "Wav2Vec2ForPreTraining"
217
+ ],
218
+ "attention_dropout": 0.1,
219
+ "bad_words_ids": null,
220
+ "bos_token_id": 1,
221
+ "chunk_size_feed_forward": 0,
222
+ "classifier_proj_size": 256,
223
+ "codevector_dim": 768,
224
+ "contrastive_logits_temperature": 0.1,
225
+ "conv_bias": true,
226
+ "conv_dim": [
227
+ 512,
228
+ 512,
229
+ 512,
230
+ 512,
231
+ 512,
232
+ 512,
233
+ 512
234
+ ],
235
+ "conv_kernel": [
236
+ 10,
237
+ 3,
238
+ 3,
239
+ 3,
240
+ 3,
241
+ 2,
242
+ 2
243
+ ],
244
+ "conv_stride": [
245
+ 5,
246
+ 2,
247
+ 2,
248
+ 2,
249
+ 2,
250
+ 2,
251
+ 2
252
+ ],
253
+ "cross_attention_hidden_size": null,
254
+ "ctc_loss_reduction": "sum",
255
+ "ctc_zero_infinity": false,
256
+ "decoder_start_token_id": null,
257
+ "diversity_loss_weight": 0.1,
258
+ "diversity_penalty": 0.0,
259
+ "do_sample": false,
260
+ "do_stable_layer_norm": true,
261
+ "early_stopping": false,
262
+ "encoder_no_repeat_ngram_size": 0,
263
+ "eos_token_id": 2,
264
+ "exponential_decay_length_penalty": null,
265
+ "feat_extract_activation": "gelu",
266
+ "feat_extract_dropout": 0.0,
267
+ "feat_extract_norm": "layer",
268
+ "feat_proj_dropout": 0.0,
269
+ "feat_quantizer_dropout": 0.0,
270
+ "final_dropout": 0.0,
271
+ "finetuning_task": null,
272
+ "forced_bos_token_id": null,
273
+ "forced_eos_token_id": null,
274
+ "gradient_checkpointing": false,
275
+ "hidden_act": "gelu",
276
+ "hidden_dropout": 0.1,
277
+ "hidden_dropout_prob": 0.1,
278
+ "hidden_size": 1024,
279
+ "id2label": {
280
+ "0": "LABEL_0",
281
+ "1": "LABEL_1"
282
+ },
283
+ "initializer_range": 0.02,
284
+ "intermediate_size": 4096,
285
+ "is_decoder": false,
286
+ "is_encoder_decoder": false,
287
+ "label2id": {
288
+ "LABEL_0": 0,
289
+ "LABEL_1": 1
290
+ },
291
+ "layer_norm_eps": 1e-05,
292
+ "layerdrop": 0.0,
293
+ "length_penalty": 1.0,
294
+ "mask_feature_length": 10,
295
+ "mask_feature_min_masks": 0,
296
+ "mask_feature_prob": 0.0,
297
+ "mask_time_length": 10,
298
+ "mask_time_min_masks": 2,
299
+ "mask_time_prob": 0.1,
300
+ "max_length": 20,
301
+ "min_length": 0,
302
+ "model_type": "wav2vec2",
303
+ "no_repeat_ngram_size": 0,
304
+ "num_adapter_layers": 3,
305
+ "num_attention_heads": 16,
306
+ "num_beam_groups": 1,
307
+ "num_beams": 1,
308
+ "num_codevector_groups": 2,
309
+ "num_codevectors_per_group": 320,
310
+ "num_conv_pos_embedding_groups": 16,
311
+ "num_conv_pos_embeddings": 128,
312
+ "num_feat_extract_layers": 7,
313
+ "num_hidden_layers": 24,
314
+ "num_negatives": 100,
315
+ "num_return_sequences": 1,
316
+ "output_attentions": false,
317
+ "output_hidden_size": 1024,
318
+ "output_hidden_states": false,
319
+ "output_scores": false,
320
+ "pad_token_id": 0,
321
+ "prefix": null,
322
+ "problem_type": null,
323
+ "proj_codevector_dim": 768,
324
+ "pruned_heads": {},
325
+ "remove_invalid_values": false,
326
+ "repetition_penalty": 1.0,
327
+ "return_dict": true,
328
+ "return_dict_in_generate": false,
329
+ "sep_token_id": null,
330
+ "task_specific_params": null,
331
+ "tdnn_dilation": [
332
+ 1,
333
+ 2,
334
+ 3,
335
+ 1,
336
+ 1
337
+ ],
338
+ "tdnn_dim": [
339
+ 512,
340
+ 512,
341
+ 512,
342
+ 512,
343
+ 1500
344
+ ],
345
+ "tdnn_kernel": [
346
+ 5,
347
+ 3,
348
+ 3,
349
+ 1,
350
+ 1
351
+ ],
352
+ "temperature": 1.0,
353
+ "tie_encoder_decoder": false,
354
+ "tie_word_embeddings": true,
355
+ "tokenizer_class": null,
356
+ "top_k": 50,
357
+ "top_p": 1.0,
358
+ "torch_dtype": null,
359
+ "torchscript": false,
360
+ "transformers_version": "4.18.0.dev0",
361
+ "typical_p": 1.0,
362
+ "use_bfloat16": false,
363
+ "use_weighted_layer_sum": false,
364
+ "vocab_size": 32,
365
+ "xvector_output_dim": 512
366
+ },
367
+ "eos_token_id": 2,
368
+ "is_encoder_decoder": true,
369
+ "max_length": 40,
370
+ "model_type": "speech-encoder-decoder",
371
+ "pad_token_id": 1,
372
+ "processor_class": "Wav2Vec2Processor",
373
+ "tie_word_embeddings": false,
374
+ "transformers_version": null,
375
+ "use_cache": false
376
+ }
377
+ loading feature extractor configuration file ./preprocessor_config.json
378
+ Feature extractor Wav2Vec2FeatureExtractor {
379
+ "do_normalize": true,
380
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
381
+ "feature_size": 1,
382
+ "padding_side": "right",
383
+ "padding_value": 0.0,
384
+ "return_attention_mask": true,
385
+ "sampling_rate": 16000
386
+ }
387
+ Didn't find file ./added_tokens.json. We won't load it.
388
+ loading file ./vocab.json
389
+ loading file ./merges.txt
390
+ loading file ./tokenizer.json
391
+ loading file None
392
+ loading file ./special_tokens_map.json
393
+ loading file ./tokenizer_config.json
394
+ loading weights file ./flax_model.msgpack
395
+ tcmalloc: large alloc 2353643520 bytes == 0x9c1e8000 @ 0x7fb0f2c21680 0x7fb0f2c42824 0x5f8a01 0x648cf1 0x5c4676 0x4f290e 0x64f718 0x5048b3 0x56b1da 0x56939a 0x5f6a13 0x50af49 0x5f3547 0x56c8cd 0x56939a 0x50aaa0 0x56c28c 0x56939a 0x5f6a13 0x56b0ae 0x56939a 0x68d047 0x67e351 0x67e3cf 0x67e471 0x67e817 0x6b6fe2 0x6b736d 0x7fb0f2a330b3 0x5fa5ce
396
+ All model checkpoint weights were used when initializing FlaxSpeechEncoderDecoderModel.
397
+ All the weights of FlaxSpeechEncoderDecoderModel were initialized from the model checkpoint at ./.
398
+ If your task is similar to the task the model of the checkpoint was trained on, you can already use FlaxSpeechEncoderDecoderModel for predictions without further training.
399
+ Feature extractor saved in ./preprocessor_config.json
400
+ tokenizer config file saved in ./tokenizer_config.json
401
+ Special tokens file saved in ./special_tokens_map.json
402
+ Configuration saved in ./config.json
403
+ loading feature extractor configuration file ./preprocessor_config.json
404
+ loading configuration file ./config.json
405
+ Model config SpeechEncoderDecoderConfig {
406
+ "_name_or_path": "./",
407
+ "architectures": [
408
+ "SpeechEncoderDecoderModel"
409
+ ],
410
+ "decoder": {
411
+ "_name_or_path": "facebook/bart-large-cnn",
412
+ "_num_labels": 3,
413
+ "activation_dropout": 0.0,
414
+ "activation_function": "gelu",
415
+ "add_cross_attention": true,
416
+ "add_final_layer_norm": false,
417
+ "architectures": [
418
+ "BartForConditionalGeneration"
419
+ ],
420
+ "attention_dropout": 0.0,
421
+ "bad_words_ids": null,
422
+ "bos_token_id": 0,
423
+ "chunk_size_feed_forward": 0,
424
+ "classif_dropout": 0.0,
425
+ "classifier_dropout": 0.0,
426
+ "cross_attention_hidden_size": null,
427
+ "d_model": 1024,
428
+ "decoder_attention_heads": 16,
429
+ "decoder_ffn_dim": 4096,
430
+ "decoder_layerdrop": 0.0,
431
+ "decoder_layers": 12,
432
+ "decoder_start_token_id": 2,
433
+ "diversity_penalty": 0.0,
434
+ "do_sample": false,
435
+ "dropout": 0.1,
436
+ "early_stopping": true,
437
+ "encoder_attention_heads": 16,
438
+ "encoder_ffn_dim": 4096,
439
+ "encoder_layerdrop": 0.0,
440
+ "encoder_layers": 12,
441
+ "encoder_no_repeat_ngram_size": 0,
442
+ "eos_token_id": 2,
443
+ "exponential_decay_length_penalty": null,
444
+ "finetuning_task": null,
445
+ "force_bos_token_to_be_generated": true,
446
+ "forced_bos_token_id": 0,
447
+ "forced_eos_token_id": 2,
448
+ "gradient_checkpointing": false,
449
+ "id2label": {
450
+ "0": "LABEL_0",
451
+ "1": "LABEL_1",
452
+ "2": "LABEL_2"
453
+ },
454
+ "init_std": 0.02,
455
+ "is_decoder": true,
456
+ "is_encoder_decoder": false,
457
+ "label2id": {
458
+ "LABEL_0": 0,
459
+ "LABEL_1": 1,
460
+ "LABEL_2": 2
461
+ },
462
+ "length_penalty": 2.0,
463
+ "max_length": 142,
464
+ "max_position_embeddings": 1024,
465
+ "min_length": 56,
466
+ "model_type": "bart",
467
+ "no_repeat_ngram_size": 3,
468
+ "normalize_before": false,
469
+ "num_beam_groups": 1,
470
+ "num_beams": 4,
471
+ "num_hidden_layers": 12,
472
+ "num_return_sequences": 1,
473
+ "output_attentions": false,
474
+ "output_hidden_states": false,
475
+ "output_past": true,
476
+ "output_scores": false,
477
+ "pad_token_id": 1,
478
+ "prefix": " ",
479
+ "problem_type": null,
480
+ "pruned_heads": {},
481
+ "remove_invalid_values": false,
482
+ "repetition_penalty": 1.0,
483
+ "return_dict": true,
484
+ "return_dict_in_generate": false,
485
+ "scale_embedding": false,
486
+ "sep_token_id": null,
487
+ "task_specific_params": {
488
+ "summarization": {
489
+ "early_stopping": true,
490
+ "length_penalty": 2.0,
491
+ "max_length": 142,
492
+ "min_length": 56,
493
+ "no_repeat_ngram_size": 3,
494
+ "num_beams": 4
495
+ }
496
+ },
497
+ "temperature": 1.0,
498
+ "tie_encoder_decoder": false,
499
+ "tie_word_embeddings": true,
500
+ "tokenizer_class": null,
501
+ "top_k": 50,
502
+ "top_p": 1.0,
503
+ "torch_dtype": null,
504
+ "torchscript": false,
505
+ "transformers_version": "4.18.0.dev0",
506
+ "typical_p": 1.0,
507
+ "use_bfloat16": false,
508
+ "use_cache": true,
509
+ "vocab_size": 50264
510
+ },
511
+ "decoder_start_token_id": 0,
512
+ "encoder": {
513
+ "_name_or_path": "facebook/wav2vec2-large-lv60",
514
+ "activation_dropout": 0.1,
515
+ "adapter_kernel_size": 3,
516
+ "adapter_stride": 2,
517
+ "add_adapter": true,
518
+ "add_cross_attention": false,
519
+ "apply_spec_augment": true,
520
+ "architectures": [
521
+ "Wav2Vec2ForPreTraining"
522
+ ],
523
+ "attention_dropout": 0.1,
524
+ "bad_words_ids": null,
525
+ "bos_token_id": 1,
526
+ "chunk_size_feed_forward": 0,
527
+ "classifier_proj_size": 256,
528
+ "codevector_dim": 768,
529
+ "contrastive_logits_temperature": 0.1,
530
+ "conv_bias": true,
531
+ "conv_dim": [
532
+ 512,
533
+ 512,
534
+ 512,
535
+ 512,
536
+ 512,
537
+ 512,
538
+ 512
539
+ ],
540
+ "conv_kernel": [
541
+ 10,
542
+ 3,
543
+ 3,
544
+ 3,
545
+ 3,
546
+ 2,
547
+ 2
548
+ ],
549
+ "conv_stride": [
550
+ 5,
551
+ 2,
552
+ 2,
553
+ 2,
554
+ 2,
555
+ 2,
556
+ 2
557
+ ],
558
+ "cross_attention_hidden_size": null,
559
+ "ctc_loss_reduction": "sum",
560
+ "ctc_zero_infinity": false,
561
+ "decoder_start_token_id": null,
562
+ "diversity_loss_weight": 0.1,
563
+ "diversity_penalty": 0.0,
564
+ "do_sample": false,
565
+ "do_stable_layer_norm": true,
566
+ "early_stopping": false,
567
+ "encoder_no_repeat_ngram_size": 0,
568
+ "eos_token_id": 2,
569
+ "exponential_decay_length_penalty": null,
570
+ "feat_extract_activation": "gelu",
571
+ "feat_extract_dropout": 0.0,
572
+ "feat_extract_norm": "layer",
573
+ "feat_proj_dropout": 0.0,
574
+ "feat_quantizer_dropout": 0.0,
575
+ "final_dropout": 0.0,
576
+ "finetuning_task": null,
577
+ "forced_bos_token_id": null,
578
+ "forced_eos_token_id": null,
579
+ "gradient_checkpointing": false,
580
+ "hidden_act": "gelu",
581
+ "hidden_dropout": 0.1,
582
+ "hidden_dropout_prob": 0.1,
583
+ "hidden_size": 1024,
584
+ "id2label": {
585
+ "0": "LABEL_0",
586
+ "1": "LABEL_1"
587
+ },
588
+ "initializer_range": 0.02,
589
+ "intermediate_size": 4096,
590
+ "is_decoder": false,
591
+ "is_encoder_decoder": false,
592
+ "label2id": {
593
+ "LABEL_0": 0,
594
+ "LABEL_1": 1
595
+ },
596
+ "layer_norm_eps": 1e-05,
597
+ "layerdrop": 0.0,
598
+ "length_penalty": 1.0,
599
+ "mask_feature_length": 10,
600
+ "mask_feature_min_masks": 0,
601
+ "mask_feature_prob": 0.0,
602
+ "mask_time_length": 10,
603
+ "mask_time_min_masks": 2,
604
+ "mask_time_prob": 0.1,
605
+ "max_length": 20,
606
+ "min_length": 0,
607
+ "model_type": "wav2vec2",
608
+ "no_repeat_ngram_size": 0,
609
+ "num_adapter_layers": 3,
610
+ "num_attention_heads": 16,
611
+ "num_beam_groups": 1,
612
+ "num_beams": 1,
613
+ "num_codevector_groups": 2,
614
+ "num_codevectors_per_group": 320,
615
+ "num_conv_pos_embedding_groups": 16,
616
+ "num_conv_pos_embeddings": 128,
617
+ "num_feat_extract_layers": 7,
618
+ "num_hidden_layers": 24,
619
+ "num_negatives": 100,
620
+ "num_return_sequences": 1,
621
+ "output_attentions": false,
622
+ "output_hidden_size": 1024,
623
+ "output_hidden_states": false,
624
+ "output_scores": false,
625
+ "pad_token_id": 0,
626
+ "prefix": null,
627
+ "problem_type": null,
628
+ "proj_codevector_dim": 768,
629
+ "pruned_heads": {},
630
+ "remove_invalid_values": false,
631
+ "repetition_penalty": 1.0,
632
+ "return_dict": true,
633
+ "return_dict_in_generate": false,
634
+ "sep_token_id": null,
635
+ "task_specific_params": null,
636
+ "tdnn_dilation": [
637
+ 1,
638
+ 2,
639
+ 3,
640
+ 1,
641
+ 1
642
+ ],
643
+ "tdnn_dim": [
644
+ 512,
645
+ 512,
646
+ 512,
647
+ 512,
648
+ 1500
649
+ ],
650
+ "tdnn_kernel": [
651
+ 5,
652
+ 3,
653
+ 3,
654
+ 1,
655
+ 1
656
+ ],
657
+ "temperature": 1.0,
658
+ "tie_encoder_decoder": false,
659
+ "tie_word_embeddings": true,
660
+ "tokenizer_class": null,
661
+ "top_k": 50,
662
+ "top_p": 1.0,
663
+ "torch_dtype": null,
664
+ "torchscript": false,
665
+ "transformers_version": "4.18.0.dev0",
666
+ "typical_p": 1.0,
667
+ "use_bfloat16": false,
668
+ "use_weighted_layer_sum": false,
669
+ "vocab_size": 32,
670
+ "xvector_output_dim": 512
671
+ },
672
+ "eos_token_id": 2,
673
+ "is_encoder_decoder": true,
674
+ "max_length": 40,
675
+ "model_type": "speech-encoder-decoder",
676
+ "pad_token_id": 1,
677
+ "processor_class": "Wav2Vec2Processor",
678
+ "tie_word_embeddings": false,
679
+ "transformers_version": null,
680
+ "use_cache": false
681
+ }
682
+ loading feature extractor configuration file ./preprocessor_config.json
683
+ Feature extractor Wav2Vec2FeatureExtractor {
684
+ "do_normalize": true,
685
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
686
+ "feature_size": 1,
687
+ "padding_side": "right",
688
+ "padding_value": 0.0,
689
+ "return_attention_mask": true,
690
+ "sampling_rate": 16000
691
+ }
692
+ Didn't find file ./added_tokens.json. We won't load it.
693
+ loading file ./vocab.json
694
+ loading file ./merges.txt
695
+ loading file ./tokenizer.json
696
+ loading file None
697
+ loading file ./special_tokens_map.json
698
+ loading file ./tokenizer_config.json
699
+ 2022-04-04 18:12:01.391712: W tensorflow/stream_executor/platform/default/dso_loader.cc:64] Could not load dynamic library 'libcuda.so.1'; dlerror: libcuda.so.1: cannot open shared object file: No such file or directory
700
+ 2022-04-04 18:12:01.391753: W tensorflow/stream_executor/cuda/cuda_driver.cc:269] failed call to cuInit: UNKNOWN ERROR (303)
701
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-eb8e77ddeb6c34dc.arrow
702
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e19eb46cb05f7b14.arrow
703
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-92685008e5c09846.arrow
704
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-0d478bd8b093039a.arrow
705
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-13aa6d940e98be09.arrow
706
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-2d24e62de045a0c8.arrow
707
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-c6e972d87654db26.arrow
708
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-4771f1765b989007.arrow
709
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ab9a2fefe02761db.arrow
710
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-6ce32d0e344f4c06.arrow
711
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-1e0f4ae48a3c8b80.arrow
712
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e6f5ab8655734a9f.arrow
713
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-be6a3a97adc5697a.arrow
714
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-239fcccc5b67f7a4.arrow
715
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ff31cd13241c8bf3.arrow
716
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-e8bf216b19718031.arrow
717
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-03ee7ff6632282a6.arrow
718
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-bb1b7de42a389f66.arrow
719
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-ed92b3340af62815.arrow
720
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-12ce4b2b8350532d.arrow
721
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-75f20b5cb7b89a9d.arrow
722
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-69941cf3a70dde04.arrow
723
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-f4d0102728859e5f.arrow
724
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-1c0eaf46e759e04d.arrow
725
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-9f8c4dccc4fa8f17.arrow
726
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-14df38b848e61b52.arrow
727
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-19c662ecc5b20d66.arrow
728
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-3276de10dacb4963.arrow
729
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-111057e4500c4cc5.arrow
730
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-d1bf30028d638290.arrow
731
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cc50c08263e878fc.arrow
732
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-f1813a8ae49a67d6.arrow
733
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00000_of_00016.arrow
734
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00001_of_00016.arrow
735
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00002_of_00016.arrow
736
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00003_of_00016.arrow
737
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00004_of_00016.arrow
738
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00005_of_00016.arrow
739
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00006_of_00016.arrow
740
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00007_of_00016.arrow
741
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00008_of_00016.arrow
742
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00009_of_00016.arrow
743
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00010_of_00016.arrow
744
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00011_of_00016.arrow
745
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00012_of_00016.arrow
746
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00013_of_00016.arrow
747
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00014_of_00016.arrow
748
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-a668e959fdd4c3af_00015_of_00016.arrow
749
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00000_of_00016.arrow
750
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00001_of_00016.arrow
751
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00002_of_00016.arrow
752
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00003_of_00016.arrow
753
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00004_of_00016.arrow
754
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00005_of_00016.arrow
755
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00006_of_00016.arrow
756
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00007_of_00016.arrow
757
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00008_of_00016.arrow
758
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00009_of_00016.arrow
759
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00010_of_00016.arrow
760
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00011_of_00016.arrow
761
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00012_of_00016.arrow
762
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00013_of_00016.arrow
763
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00014_of_00016.arrow
764
+ 04/04/2022 18:12:00 - WARNING - datasets.arrow_dataset - Loading cached processed dataset at /home/sanchitgandhi/cache/huggingface/datasets/librispeech_asr/clean/2.1.0/1f4602f6b5fed8d3ab3e3382783173f2e12d9877e98775e34d7780881175096c/cache-cea415e4aeb9dd74_00015_of_00016.arrow
765
+ 04/04/2022 18:12:04 - INFO - __main__ - ***** Running training *****
766
+ 04/04/2022 18:12:04 - INFO - __main__ - Num examples = 0
767
+ 04/04/2022 18:12:04 - INFO - __main__ - Num Epochs = 5
768
+ 04/04/2022 18:12:04 - INFO - __main__ - Instantaneous batch size per device = 2
769
+ 04/04/2022 18:12:04 - INFO - __main__ - Num gradient accumulation steps = 1
770
+ 04/04/2022 18:12:04 - INFO - __main__ - Total train batch size (w. parallel & distributed) = 2
771
+ 04/04/2022 18:12:04 - INFO - __main__ - Total optimization steps = 0
772
+ Epoch ... (1/5): 0%| | 0/5 [00:00<?, ?it/s]
773
+ Traceback (most recent call last):
774
+ File "run_flax_speech_recognition_seq2seq.py", line 1133, in <module>
775
+ main()
776
+ File "run_flax_speech_recognition_seq2seq.py", line 1050, in main
777
+ train_samples_idx = get_grouped_indices(vectorized_datasets["train"], batch_size_per_update, input_rng)
778
+ File "run_flax_speech_recognition_seq2seq.py", line 453, in get_grouped_indices
779
+ lengths = dataset["input_length"]
780
+ File "/home/sanchitgandhi/datasets/src/datasets/arrow_dataset.py", line 1765, in __getitem__
781
+ return self._getitem(
782
+ File "/home/sanchitgandhi/datasets/src/datasets/arrow_dataset.py", line 1749, in _getitem
783
+ pa_subtable = query_table(self._data, key, indices=self._indices if self._indices is not None else None)
784
+ File "/home/sanchitgandhi/datasets/src/datasets/formatting/formatting.py", line 491, in query_table
785
+ pa_subtable = _query_table_with_indices_mapping(table, key, indices=indices)
786
+ File "/home/sanchitgandhi/datasets/src/datasets/formatting/formatting.py", line 69, in _query_table_with_indices_mapping
787
+ return _query_table(table, indices.column(0).to_pylist())
788
+ File "/home/sanchitgandhi/datasets/src/datasets/table.py", line 343, in column
789
+ return self.table.column(*args, **kwargs)
790
+ File "pyarrow/table.pxi", line 2115, in pyarrow.lib.Table.column
791
+ File "pyarrow/table.pxi", line 2130, in pyarrow.lib.Table._column
792
+ File "pyarrow/array.pxi", line 561, in pyarrow.lib._normalize_index
793
+ IndexError: index out of bounds
wandb/run-20220404_181139-xmgtui21/files/requirements.txt ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.0.0
2
+ aiohttp==3.8.1
3
+ aiosignal==1.2.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asttokens==2.0.5
9
+ astunparse==1.6.3
10
+ async-timeout==4.0.2
11
+ attrs==21.4.0
12
+ audioread==2.1.9
13
+ babel==2.9.1
14
+ backcall==0.2.0
15
+ beautifulsoup4==4.10.0
16
+ bleach==4.1.0
17
+ cachetools==5.0.0
18
+ certifi==2021.10.8
19
+ cffi==1.15.0
20
+ charset-normalizer==2.0.12
21
+ chex==0.1.1
22
+ click==8.1.0
23
+ cycler==0.11.0
24
+ datasets==2.0.1.dev0
25
+ debugpy==1.6.0
26
+ decorator==5.1.1
27
+ defusedxml==0.7.1
28
+ dill==0.3.4
29
+ dm-tree==0.1.6
30
+ docker-pycreds==0.4.0
31
+ entrypoints==0.4
32
+ executing==0.8.3
33
+ filelock==3.6.0
34
+ flatbuffers==2.0
35
+ flax==0.4.1
36
+ fonttools==4.31.2
37
+ frozenlist==1.3.0
38
+ fsspec==2022.2.0
39
+ gast==0.5.3
40
+ gitdb==4.0.9
41
+ gitpython==3.1.27
42
+ google-auth-oauthlib==0.4.6
43
+ google-auth==2.6.2
44
+ google-pasta==0.2.0
45
+ grpcio==1.44.0
46
+ h5py==3.6.0
47
+ huggingface-hub==0.4.0
48
+ idna==3.3
49
+ importlib-metadata==4.11.3
50
+ importlib-resources==5.6.0
51
+ ipdb==0.13.9
52
+ ipykernel==6.10.0
53
+ ipython-genutils==0.2.0
54
+ ipython==8.2.0
55
+ jax==0.3.4
56
+ jaxlib==0.3.2
57
+ jedi==0.18.1
58
+ jinja2==3.1.1
59
+ jiwer==2.3.0
60
+ joblib==1.1.0
61
+ json5==0.9.6
62
+ jsonschema==4.4.0
63
+ jupyter-client==7.2.1
64
+ jupyter-core==4.9.2
65
+ jupyter-server==1.16.0
66
+ jupyterlab-pygments==0.1.2
67
+ jupyterlab-server==2.12.0
68
+ jupyterlab==3.3.2
69
+ keras-preprocessing==1.1.2
70
+ keras==2.8.0
71
+ kiwisolver==1.4.2
72
+ libclang==13.0.0
73
+ librosa==0.9.1
74
+ libtpu-nightly==0.1.dev20220315
75
+ llvmlite==0.38.0
76
+ markdown==3.3.6
77
+ markupsafe==2.1.1
78
+ matplotlib-inline==0.1.3
79
+ matplotlib==3.5.1
80
+ mistune==0.8.4
81
+ msgpack==1.0.3
82
+ multidict==6.0.2
83
+ multiprocess==0.70.12.2
84
+ nbclassic==0.3.7
85
+ nbclient==0.5.13
86
+ nbconvert==6.4.5
87
+ nbformat==5.2.0
88
+ nest-asyncio==1.5.4
89
+ notebook-shim==0.1.0
90
+ notebook==6.4.10
91
+ numba==0.55.1
92
+ numpy==1.21.0
93
+ oauthlib==3.2.0
94
+ opt-einsum==3.3.0
95
+ optax==0.1.1
96
+ packaging==21.3
97
+ pandas==1.4.1
98
+ pandocfilters==1.5.0
99
+ parso==0.8.3
100
+ pathtools==0.1.2
101
+ pexpect==4.8.0
102
+ pickleshare==0.7.5
103
+ pillow==9.0.1
104
+ pip==20.0.2
105
+ pkg-resources==0.0.0
106
+ pooch==1.6.0
107
+ prometheus-client==0.13.1
108
+ promise==2.3
109
+ prompt-toolkit==3.0.28
110
+ protobuf==3.19.4
111
+ psutil==5.9.0
112
+ ptyprocess==0.7.0
113
+ pure-eval==0.2.2
114
+ pyarrow==7.0.0
115
+ pyasn1-modules==0.2.8
116
+ pyasn1==0.4.8
117
+ pycparser==2.21
118
+ pygments==2.11.2
119
+ pyparsing==3.0.7
120
+ pyrsistent==0.18.1
121
+ python-dateutil==2.8.2
122
+ python-levenshtein==0.12.2
123
+ pytz==2022.1
124
+ pyyaml==6.0
125
+ pyzmq==22.3.0
126
+ regex==2022.3.15
127
+ requests-oauthlib==1.3.1
128
+ requests==2.27.1
129
+ resampy==0.2.2
130
+ responses==0.18.0
131
+ rsa==4.8
132
+ sacremoses==0.0.49
133
+ scikit-learn==1.0.2
134
+ scipy==1.8.0
135
+ send2trash==1.8.0
136
+ sentry-sdk==1.5.8
137
+ setproctitle==1.2.2
138
+ setuptools==44.0.0
139
+ shortuuid==1.0.8
140
+ six==1.16.0
141
+ smmap==5.0.0
142
+ sniffio==1.2.0
143
+ soundfile==0.10.3.post1
144
+ soupsieve==2.3.1
145
+ stack-data==0.2.0
146
+ tensorboard-data-server==0.6.1
147
+ tensorboard-plugin-wit==1.8.1
148
+ tensorboard==2.8.0
149
+ tensorflow-io-gcs-filesystem==0.24.0
150
+ tensorflow==2.8.0
151
+ termcolor==1.1.0
152
+ terminado==0.13.3
153
+ testpath==0.6.0
154
+ tf-estimator-nightly==2.8.0.dev2021122109
155
+ threadpoolctl==3.1.0
156
+ tokenizers==0.11.6
157
+ toml==0.10.2
158
+ toolz==0.11.2
159
+ torch==1.11.0+cpu
160
+ torchaudio==0.11.0+cpu
161
+ tornado==6.1
162
+ tqdm==4.63.1
163
+ traitlets==5.1.1
164
+ transformers==4.18.0.dev0
165
+ typing-extensions==4.1.1
166
+ urllib3==1.26.9
167
+ wandb==0.12.11
168
+ wcwidth==0.2.5
169
+ webencodings==0.5.1
170
+ websocket-client==1.3.2
171
+ werkzeug==2.1.0
172
+ wheel==0.37.1
173
+ wrapt==1.14.0
174
+ xxhash==3.0.0
175
+ yarl==1.7.2
176
+ yaspin==2.1.0
177
+ zipp==3.7.0
wandb/run-20220404_181139-xmgtui21/files/wandb-metadata.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.4.0-1043-gcp-x86_64-with-glibc2.29",
3
+ "python": "3.8.10",
4
+ "heartbeatAt": "2022-04-04T18:11:42.230539",
5
+ "startedAt": "2022-04-04T18:11:39.127973",
6
+ "docker": null,
7
+ "cpu_count": 96,
8
+ "cuda": null,
9
+ "args": [
10
+ "--overwrite_output_dir",
11
+ "--freeze_feature_encoder",
12
+ "--predict_with_generate",
13
+ "--do_lower_case",
14
+ "--do_train",
15
+ "--do_eval",
16
+ "--dataset_cache_dir=/home/sanchitgandhi/cache/huggingface/datasets",
17
+ "--dataset_config_name=clean",
18
+ "--dataset_name=librispeech_asr",
19
+ "--eval_split_name=validation",
20
+ "--generation_max_length=40",
21
+ "--generation_num_beams=1",
22
+ "--gradient_accumulation_steps=1",
23
+ "--learning_rate=7.331199736432637e-05",
24
+ "--length_column_name=input_length",
25
+ "--logging_steps=1",
26
+ "--matmul_precision=highest",
27
+ "--max_duration_in_seconds=15",
28
+ "--max_target_length=64",
29
+ "--min_duration_in_seconds=15",
30
+ "--model_name_or_path=./",
31
+ "--num_train_epochs=5",
32
+ "--output_dir=./",
33
+ "--per_device_eval_batch_size=2",
34
+ "--per_device_train_batch_size=2",
35
+ "--preprocessing_num_workers=16",
36
+ "--text_column_name=text",
37
+ "--train_split_name=train.100",
38
+ "--wandb_project=flax-wav2vec2-2-bart-large-cnn",
39
+ "--warmup_steps=500"
40
+ ],
41
+ "state": "running",
42
+ "program": "run_flax_speech_recognition_seq2seq.py",
43
+ "codePath": "run_flax_speech_recognition_seq2seq.py",
44
+ "git": {
45
+ "remote": "https://huggingface.co/sanchit-gandhi/flax-wav2vec2-2-bart-large-cnn",
46
+ "commit": "c09faa6b18ac3f19251c414100acf1bad11c1b16"
47
+ },
48
+ "email": "sanchit@huggingface.co",
49
+ "root": "/home/sanchitgandhi/flax-wav2vec2-2-bart-large-cnn",
50
+ "host": "t1v-n-4eb331dd-w-0",
51
+ "username": "sanchitgandhi",
52
+ "executable": "/home/sanchitgandhi/venv/bin/python3"
53
+ }