AlexN commited on
Commit
ef5fb31
1 Parent(s): 9d07727
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .ipynb_checkpoints/added_tokens-checkpoint.json +0 -1
  2. .ipynb_checkpoints/all_results-checkpoint.json +14 -0
  3. .ipynb_checkpoints/run-checkpoint.sh +7 -7
  4. .ipynb_checkpoints/run_speech_recognition_ctc-checkpoint.py +2 -1
  5. .ipynb_checkpoints/special_tokens_map-checkpoint.json +0 -1
  6. .ipynb_checkpoints/tokenizer_config-checkpoint.json +0 -1
  7. .ipynb_checkpoints/train_results-checkpoint.json +8 -0
  8. .ipynb_checkpoints/trainer_state-checkpoint.json +1096 -0
  9. added_tokens.json +0 -1
  10. config.json +5 -5
  11. run.sh +3 -3
  12. run_speech_recognition_ctc.py +2 -2
  13. special_tokens_map.json +0 -1
  14. tokenizer_config.json +0 -1
  15. wandb/debug-internal.log +1 -1
  16. wandb/debug.log +1 -1
  17. wandb/latest-run +1 -1
  18. wandb/run-20220130_163246-20w0bl60/files/conda-environment.yaml +0 -0
  19. wandb/run-20220130_163246-20w0bl60/files/config.yaml +654 -0
  20. wandb/run-20220130_163246-20w0bl60/files/output.log +19 -0
  21. wandb/run-20220130_163246-20w0bl60/files/requirements.txt +180 -0
  22. wandb/run-20220130_163246-20w0bl60/files/wandb-metadata.json +65 -0
  23. wandb/run-20220130_163246-20w0bl60/files/wandb-summary.json +1 -0
  24. wandb/run-20220130_163246-20w0bl60/logs/debug-internal.log +147 -0
  25. wandb/run-20220130_163246-20w0bl60/logs/debug.log +128 -0
  26. wandb/run-20220130_163246-20w0bl60/run-20w0bl60.wandb +0 -0
  27. wandb/run-20220130_174749-kzuygp3o/files/conda-environment.yaml +0 -0
  28. wandb/run-20220130_174749-kzuygp3o/files/config.yaml +654 -0
  29. wandb/run-20220130_174749-kzuygp3o/files/output.log +19 -0
  30. wandb/run-20220130_174749-kzuygp3o/files/requirements.txt +180 -0
  31. wandb/run-20220130_174749-kzuygp3o/files/wandb-metadata.json +66 -0
  32. wandb/run-20220130_174749-kzuygp3o/files/wandb-summary.json +1 -0
  33. wandb/run-20220130_174749-kzuygp3o/logs/debug-internal.log +146 -0
  34. wandb/run-20220130_174749-kzuygp3o/logs/debug.log +128 -0
  35. wandb/run-20220130_174749-kzuygp3o/run-kzuygp3o.wandb +0 -0
  36. wandb/run-20220130_175135-mz3dbiz4/files/conda-environment.yaml +0 -0
  37. wandb/run-20220130_175135-mz3dbiz4/files/config.yaml +654 -0
  38. wandb/run-20220130_175135-mz3dbiz4/files/output.log +56 -0
  39. wandb/run-20220130_175135-mz3dbiz4/files/requirements.txt +180 -0
  40. wandb/run-20220130_175135-mz3dbiz4/files/wandb-metadata.json +66 -0
  41. wandb/run-20220130_175135-mz3dbiz4/files/wandb-summary.json +1 -0
  42. wandb/run-20220130_175135-mz3dbiz4/logs/debug-internal.log +192 -0
  43. wandb/run-20220130_175135-mz3dbiz4/logs/debug.log +128 -0
  44. wandb/run-20220130_175135-mz3dbiz4/run-mz3dbiz4.wandb +0 -0
  45. wandb/run-20220130_175417-3n44neww/files/conda-environment.yaml +0 -0
  46. wandb/run-20220130_175417-3n44neww/files/config.yaml +686 -0
  47. wandb/run-20220130_175417-3n44neww/files/output.log +67 -0
  48. wandb/run-20220130_175417-3n44neww/files/requirements.txt +180 -0
  49. wandb/run-20220130_175417-3n44neww/files/wandb-metadata.json +66 -0
  50. wandb/run-20220130_175417-3n44neww/files/wandb-summary.json +1 -0
.ipynb_checkpoints/added_tokens-checkpoint.json DELETED
@@ -1 +0,0 @@
1
- {"<s>": 216, "</s>": 217}
 
 
.ipynb_checkpoints/all_results-checkpoint.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "eval_loss": 0.23875188827514648,
4
+ "eval_runtime": 294.1776,
5
+ "eval_samples": 5792,
6
+ "eval_samples_per_second": 19.689,
7
+ "eval_steps_per_second": 0.309,
8
+ "eval_wer": 0.3680797679950471,
9
+ "train_loss": 1.442369053426242,
10
+ "train_runtime": 53680.5392,
11
+ "train_samples": 442265,
12
+ "train_samples_per_second": 16.478,
13
+ "train_steps_per_second": 0.257
14
+ }
.ipynb_checkpoints/run-checkpoint.sh CHANGED
@@ -6,12 +6,12 @@ python run_speech_recognition_ctc.py \
6
  --tokenizer_name_or_path="./" \
7
  --output_dir="./" \
8
  --overwrite_output_dir \
9
- --num_train_epochs="2" \
10
  --per_device_train_batch_size="64" \
11
  --per_device_eval_batch_size="64" \
12
  --gradient_accumulation_steps="1" \
13
- --learning_rate="1e-4" \
14
- --warmup_steps="1500" \
15
  --length_column_name="input_length" \
16
  --evaluation_strategy="steps" \
17
  --text_column_name="sentence" \
@@ -23,15 +23,15 @@ python run_speech_recognition_ctc.py \
23
  --save_total_limit="2" \
24
  --freeze_feature_encoder \
25
  --feat_proj_dropout="0.0" \
26
- --mask_time_prob="0.75" \
27
  --mask_time_length="10" \
28
- --mask_feature_prob="0.4" \
29
  --mask_feature_length="10" \
30
  --gradient_checkpointing \
31
  --report_to="wandb" \
32
  --run_name="xls-r-300m-fr" \
33
- --max_eval_samples="6000" \
34
- --max_duration_in_seconds="9" \
35
  --use_auth_token \
36
  --fp16 \
37
  --group_by_length \
 
6
  --tokenizer_name_or_path="./" \
7
  --output_dir="./" \
8
  --overwrite_output_dir \
9
+ --num_train_epochs="4" \
10
  --per_device_train_batch_size="64" \
11
  --per_device_eval_batch_size="64" \
12
  --gradient_accumulation_steps="1" \
13
+ --learning_rate="1.5e-4" \
14
+ --warmup_steps="2700" \
15
  --length_column_name="input_length" \
16
  --evaluation_strategy="steps" \
17
  --text_column_name="sentence" \
 
23
  --save_total_limit="2" \
24
  --freeze_feature_encoder \
25
  --feat_proj_dropout="0.0" \
26
+ --mask_time_prob="0.6" \
27
  --mask_time_length="10" \
28
+ --mask_feature_prob="0.25" \
29
  --mask_feature_length="10" \
30
  --gradient_checkpointing \
31
  --report_to="wandb" \
32
  --run_name="xls-r-300m-fr" \
33
+ --max_eval_samples="4500" \
34
+ --max_duration_in_seconds="10" \
35
  --use_auth_token \
36
  --fp16 \
37
  --group_by_length \
.ipynb_checkpoints/run_speech_recognition_ctc-checkpoint.py CHANGED
@@ -642,7 +642,8 @@ def main():
642
 
643
  pred.label_ids[pred.label_ids == -100] = tokenizer.pad_token_id
644
 
645
- pred_str = tokenizer.batch_decode(pred_ids, skip_special_tokens=True)#being sure to remove <s> from the output
 
646
  # we do not want to group tokens when computing the metrics
647
  label_str = tokenizer.batch_decode(pred.label_ids, group_tokens=False)
648
 
 
642
 
643
  pred.label_ids[pred.label_ids == -100] = tokenizer.pad_token_id
644
 
645
+ pred_str = tokenizer.batch_decode(pred_ids)#, skip_special_tokens=True)#being sure to remove <s> from the output
646
+ print(pred_str)
647
  # we do not want to group tokens when computing the metrics
648
  label_str = tokenizer.batch_decode(pred.label_ids, group_tokens=False)
649
 
.ipynb_checkpoints/special_tokens_map-checkpoint.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
 
 
.ipynb_checkpoints/tokenizer_config-checkpoint.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "config": null, "tokenizer_type": "wav2vec2", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "./", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
 
 
.ipynb_checkpoints/train_results-checkpoint.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "train_loss": 1.442369053426242,
4
+ "train_runtime": 53680.5392,
5
+ "train_samples": 442265,
6
+ "train_samples_per_second": 16.478,
7
+ "train_steps_per_second": 0.257
8
+ }
.ipynb_checkpoints/trainer_state-checkpoint.json ADDED
@@ -0,0 +1,1096 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.23875188827514648,
3
+ "best_model_checkpoint": "./checkpoint-13500",
4
+ "epoch": 2.0,
5
+ "global_step": 13822,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.01,
12
+ "learning_rate": 6.533333333333333e-06,
13
+ "loss": 17.2403,
14
+ "step": 100
15
+ },
16
+ {
17
+ "epoch": 0.03,
18
+ "learning_rate": 1.32e-05,
19
+ "loss": 10.2311,
20
+ "step": 200
21
+ },
22
+ {
23
+ "epoch": 0.04,
24
+ "learning_rate": 1.9800000000000004e-05,
25
+ "loss": 7.834,
26
+ "step": 300
27
+ },
28
+ {
29
+ "epoch": 0.06,
30
+ "learning_rate": 2.646666666666667e-05,
31
+ "loss": 6.0656,
32
+ "step": 400
33
+ },
34
+ {
35
+ "epoch": 0.07,
36
+ "learning_rate": 3.313333333333333e-05,
37
+ "loss": 4.3748,
38
+ "step": 500
39
+ },
40
+ {
41
+ "epoch": 0.07,
42
+ "eval_loss": 3.878422975540161,
43
+ "eval_runtime": 285.8223,
44
+ "eval_samples_per_second": 20.264,
45
+ "eval_steps_per_second": 0.318,
46
+ "eval_wer": 1.0,
47
+ "step": 500
48
+ },
49
+ {
50
+ "epoch": 0.09,
51
+ "learning_rate": 3.9800000000000005e-05,
52
+ "loss": 3.2923,
53
+ "step": 600
54
+ },
55
+ {
56
+ "epoch": 0.1,
57
+ "learning_rate": 4.646666666666667e-05,
58
+ "loss": 2.9475,
59
+ "step": 700
60
+ },
61
+ {
62
+ "epoch": 0.12,
63
+ "learning_rate": 5.3133333333333335e-05,
64
+ "loss": 2.8639,
65
+ "step": 800
66
+ },
67
+ {
68
+ "epoch": 0.13,
69
+ "learning_rate": 5.9800000000000003e-05,
70
+ "loss": 2.8265,
71
+ "step": 900
72
+ },
73
+ {
74
+ "epoch": 0.14,
75
+ "learning_rate": 6.646666666666667e-05,
76
+ "loss": 2.8068,
77
+ "step": 1000
78
+ },
79
+ {
80
+ "epoch": 0.14,
81
+ "eval_loss": 2.828850746154785,
82
+ "eval_runtime": 292.3877,
83
+ "eval_samples_per_second": 19.809,
84
+ "eval_steps_per_second": 0.311,
85
+ "eval_wer": 0.9826485059793412,
86
+ "step": 1000
87
+ },
88
+ {
89
+ "epoch": 0.16,
90
+ "learning_rate": 7.306666666666668e-05,
91
+ "loss": 2.779,
92
+ "step": 1100
93
+ },
94
+ {
95
+ "epoch": 0.17,
96
+ "learning_rate": 7.973333333333334e-05,
97
+ "loss": 2.6402,
98
+ "step": 1200
99
+ },
100
+ {
101
+ "epoch": 0.19,
102
+ "learning_rate": 8.64e-05,
103
+ "loss": 2.1119,
104
+ "step": 1300
105
+ },
106
+ {
107
+ "epoch": 0.2,
108
+ "learning_rate": 9.306666666666667e-05,
109
+ "loss": 1.7965,
110
+ "step": 1400
111
+ },
112
+ {
113
+ "epoch": 0.22,
114
+ "learning_rate": 9.973333333333334e-05,
115
+ "loss": 1.6698,
116
+ "step": 1500
117
+ },
118
+ {
119
+ "epoch": 0.22,
120
+ "eval_loss": 0.881136417388916,
121
+ "eval_runtime": 297.1806,
122
+ "eval_samples_per_second": 19.49,
123
+ "eval_steps_per_second": 0.306,
124
+ "eval_wer": 0.7127472384241911,
125
+ "step": 1500
126
+ },
127
+ {
128
+ "epoch": 0.23,
129
+ "learning_rate": 9.92209056971271e-05,
130
+ "loss": 1.5882,
131
+ "step": 1600
132
+ },
133
+ {
134
+ "epoch": 0.25,
135
+ "learning_rate": 9.840934913163448e-05,
136
+ "loss": 1.5172,
137
+ "step": 1700
138
+ },
139
+ {
140
+ "epoch": 0.26,
141
+ "learning_rate": 9.759779256614186e-05,
142
+ "loss": 1.4579,
143
+ "step": 1800
144
+ },
145
+ {
146
+ "epoch": 0.27,
147
+ "learning_rate": 9.678623600064926e-05,
148
+ "loss": 1.3829,
149
+ "step": 1900
150
+ },
151
+ {
152
+ "epoch": 0.29,
153
+ "learning_rate": 9.597467943515663e-05,
154
+ "loss": 1.3488,
155
+ "step": 2000
156
+ },
157
+ {
158
+ "epoch": 0.29,
159
+ "eval_loss": 0.516592800617218,
160
+ "eval_runtime": 301.2842,
161
+ "eval_samples_per_second": 19.224,
162
+ "eval_steps_per_second": 0.302,
163
+ "eval_wer": 0.5369024731988661,
164
+ "step": 2000
165
+ },
166
+ {
167
+ "epoch": 0.3,
168
+ "learning_rate": 9.516312286966402e-05,
169
+ "loss": 1.2981,
170
+ "step": 2100
171
+ },
172
+ {
173
+ "epoch": 0.32,
174
+ "learning_rate": 9.43515663041714e-05,
175
+ "loss": 1.2845,
176
+ "step": 2200
177
+ },
178
+ {
179
+ "epoch": 0.33,
180
+ "learning_rate": 9.354000973867879e-05,
181
+ "loss": 1.2459,
182
+ "step": 2300
183
+ },
184
+ {
185
+ "epoch": 0.35,
186
+ "learning_rate": 9.272845317318618e-05,
187
+ "loss": 1.2255,
188
+ "step": 2400
189
+ },
190
+ {
191
+ "epoch": 0.36,
192
+ "learning_rate": 9.191689660769356e-05,
193
+ "loss": 1.2239,
194
+ "step": 2500
195
+ },
196
+ {
197
+ "epoch": 0.36,
198
+ "eval_loss": 0.4104757010936737,
199
+ "eval_runtime": 299.1395,
200
+ "eval_samples_per_second": 19.362,
201
+ "eval_steps_per_second": 0.304,
202
+ "eval_wer": 0.474111245071524,
203
+ "step": 2500
204
+ },
205
+ {
206
+ "epoch": 0.38,
207
+ "learning_rate": 9.110534004220094e-05,
208
+ "loss": 1.2024,
209
+ "step": 2600
210
+ },
211
+ {
212
+ "epoch": 0.39,
213
+ "learning_rate": 9.030189904236326e-05,
214
+ "loss": 1.1851,
215
+ "step": 2700
216
+ },
217
+ {
218
+ "epoch": 0.41,
219
+ "learning_rate": 8.949034247687063e-05,
220
+ "loss": 1.1768,
221
+ "step": 2800
222
+ },
223
+ {
224
+ "epoch": 0.42,
225
+ "learning_rate": 8.867878591137803e-05,
226
+ "loss": 1.1641,
227
+ "step": 2900
228
+ },
229
+ {
230
+ "epoch": 0.43,
231
+ "learning_rate": 8.786722934588541e-05,
232
+ "loss": 1.1537,
233
+ "step": 3000
234
+ },
235
+ {
236
+ "epoch": 0.43,
237
+ "eval_loss": 0.35850802063941956,
238
+ "eval_runtime": 299.662,
239
+ "eval_samples_per_second": 19.328,
240
+ "eval_steps_per_second": 0.304,
241
+ "eval_wer": 0.4448499462348073,
242
+ "step": 3000
243
+ },
244
+ {
245
+ "epoch": 0.45,
246
+ "learning_rate": 8.70556727803928e-05,
247
+ "loss": 1.1449,
248
+ "step": 3100
249
+ },
250
+ {
251
+ "epoch": 0.46,
252
+ "learning_rate": 8.624411621490018e-05,
253
+ "loss": 1.1379,
254
+ "step": 3200
255
+ },
256
+ {
257
+ "epoch": 0.48,
258
+ "learning_rate": 8.543255964940758e-05,
259
+ "loss": 1.1331,
260
+ "step": 3300
261
+ },
262
+ {
263
+ "epoch": 0.49,
264
+ "learning_rate": 8.462100308391495e-05,
265
+ "loss": 1.1205,
266
+ "step": 3400
267
+ },
268
+ {
269
+ "epoch": 0.51,
270
+ "learning_rate": 8.380944651842234e-05,
271
+ "loss": 1.1184,
272
+ "step": 3500
273
+ },
274
+ {
275
+ "epoch": 0.51,
276
+ "eval_loss": 0.333638072013855,
277
+ "eval_runtime": 297.0402,
278
+ "eval_samples_per_second": 19.499,
279
+ "eval_steps_per_second": 0.306,
280
+ "eval_wer": 0.42922545537489004,
281
+ "step": 3500
282
+ },
283
+ {
284
+ "epoch": 0.52,
285
+ "learning_rate": 8.299788995292971e-05,
286
+ "loss": 1.1014,
287
+ "step": 3600
288
+ },
289
+ {
290
+ "epoch": 0.54,
291
+ "learning_rate": 8.218633338743711e-05,
292
+ "loss": 1.1114,
293
+ "step": 3700
294
+ },
295
+ {
296
+ "epoch": 0.55,
297
+ "learning_rate": 8.13747768219445e-05,
298
+ "loss": 1.117,
299
+ "step": 3800
300
+ },
301
+ {
302
+ "epoch": 0.56,
303
+ "learning_rate": 8.056322025645188e-05,
304
+ "loss": 1.102,
305
+ "step": 3900
306
+ },
307
+ {
308
+ "epoch": 0.58,
309
+ "learning_rate": 7.975166369095926e-05,
310
+ "loss": 1.0968,
311
+ "step": 4000
312
+ },
313
+ {
314
+ "epoch": 0.58,
315
+ "eval_loss": 0.31949570775032043,
316
+ "eval_runtime": 296.0172,
317
+ "eval_samples_per_second": 19.566,
318
+ "eval_steps_per_second": 0.307,
319
+ "eval_wer": 0.4180162273127179,
320
+ "step": 4000
321
+ },
322
+ {
323
+ "epoch": 0.59,
324
+ "learning_rate": 7.894822269112158e-05,
325
+ "loss": 1.0942,
326
+ "step": 4100
327
+ },
328
+ {
329
+ "epoch": 0.61,
330
+ "learning_rate": 7.813666612562897e-05,
331
+ "loss": 1.0859,
332
+ "step": 4200
333
+ },
334
+ {
335
+ "epoch": 0.62,
336
+ "learning_rate": 7.732510956013635e-05,
337
+ "loss": 1.0767,
338
+ "step": 4300
339
+ },
340
+ {
341
+ "epoch": 0.64,
342
+ "learning_rate": 7.652166856029866e-05,
343
+ "loss": 1.0766,
344
+ "step": 4400
345
+ },
346
+ {
347
+ "epoch": 0.65,
348
+ "learning_rate": 7.571011199480604e-05,
349
+ "loss": 1.0737,
350
+ "step": 4500
351
+ },
352
+ {
353
+ "epoch": 0.65,
354
+ "eval_loss": 0.30754634737968445,
355
+ "eval_runtime": 296.2378,
356
+ "eval_samples_per_second": 19.552,
357
+ "eval_steps_per_second": 0.307,
358
+ "eval_wer": 0.41408973899442797,
359
+ "step": 4500
360
+ },
361
+ {
362
+ "epoch": 0.67,
363
+ "learning_rate": 7.489855542931342e-05,
364
+ "loss": 1.0807,
365
+ "step": 4600
366
+ },
367
+ {
368
+ "epoch": 0.68,
369
+ "learning_rate": 7.40869988638208e-05,
370
+ "loss": 1.071,
371
+ "step": 4700
372
+ },
373
+ {
374
+ "epoch": 0.69,
375
+ "learning_rate": 7.32754422983282e-05,
376
+ "loss": 1.0613,
377
+ "step": 4800
378
+ },
379
+ {
380
+ "epoch": 0.71,
381
+ "learning_rate": 7.246388573283557e-05,
382
+ "loss": 1.0635,
383
+ "step": 4900
384
+ },
385
+ {
386
+ "epoch": 0.72,
387
+ "learning_rate": 7.165232916734297e-05,
388
+ "loss": 1.0677,
389
+ "step": 5000
390
+ },
391
+ {
392
+ "epoch": 0.72,
393
+ "eval_loss": 0.30150118470191956,
394
+ "eval_runtime": 297.4742,
395
+ "eval_samples_per_second": 19.471,
396
+ "eval_steps_per_second": 0.306,
397
+ "eval_wer": 0.4089250219948516,
398
+ "step": 5000
399
+ },
400
+ {
401
+ "epoch": 0.74,
402
+ "learning_rate": 7.084077260185034e-05,
403
+ "loss": 1.0707,
404
+ "step": 5100
405
+ },
406
+ {
407
+ "epoch": 0.75,
408
+ "learning_rate": 7.002921603635774e-05,
409
+ "loss": 1.0617,
410
+ "step": 5200
411
+ },
412
+ {
413
+ "epoch": 0.77,
414
+ "learning_rate": 6.921765947086512e-05,
415
+ "loss": 1.0566,
416
+ "step": 5300
417
+ },
418
+ {
419
+ "epoch": 0.78,
420
+ "learning_rate": 6.84061029053725e-05,
421
+ "loss": 1.0518,
422
+ "step": 5400
423
+ },
424
+ {
425
+ "epoch": 0.8,
426
+ "learning_rate": 6.760266190553481e-05,
427
+ "loss": 1.0462,
428
+ "step": 5500
429
+ },
430
+ {
431
+ "epoch": 0.8,
432
+ "eval_loss": 0.297052800655365,
433
+ "eval_runtime": 296.1592,
434
+ "eval_samples_per_second": 19.557,
435
+ "eval_steps_per_second": 0.307,
436
+ "eval_wer": 0.4077193782788621,
437
+ "step": 5500
438
+ },
439
+ {
440
+ "epoch": 0.81,
441
+ "learning_rate": 6.679110534004221e-05,
442
+ "loss": 1.0514,
443
+ "step": 5600
444
+ },
445
+ {
446
+ "epoch": 0.82,
447
+ "learning_rate": 6.597954877454959e-05,
448
+ "loss": 1.0446,
449
+ "step": 5700
450
+ },
451
+ {
452
+ "epoch": 0.84,
453
+ "learning_rate": 6.516799220905698e-05,
454
+ "loss": 1.0358,
455
+ "step": 5800
456
+ },
457
+ {
458
+ "epoch": 0.85,
459
+ "learning_rate": 6.435643564356436e-05,
460
+ "loss": 1.0364,
461
+ "step": 5900
462
+ },
463
+ {
464
+ "epoch": 0.87,
465
+ "learning_rate": 6.354487907807174e-05,
466
+ "loss": 1.0392,
467
+ "step": 6000
468
+ },
469
+ {
470
+ "epoch": 0.87,
471
+ "eval_loss": 0.2870033383369446,
472
+ "eval_runtime": 295.9814,
473
+ "eval_samples_per_second": 19.569,
474
+ "eval_steps_per_second": 0.307,
475
+ "eval_wer": 0.3997034768157972,
476
+ "step": 6000
477
+ },
478
+ {
479
+ "epoch": 0.88,
480
+ "learning_rate": 6.273332251257913e-05,
481
+ "loss": 1.0375,
482
+ "step": 6100
483
+ },
484
+ {
485
+ "epoch": 0.9,
486
+ "learning_rate": 6.192176594708652e-05,
487
+ "loss": 1.0408,
488
+ "step": 6200
489
+ },
490
+ {
491
+ "epoch": 0.91,
492
+ "learning_rate": 6.11102093815939e-05,
493
+ "loss": 1.0382,
494
+ "step": 6300
495
+ },
496
+ {
497
+ "epoch": 0.93,
498
+ "learning_rate": 6.0298652816101284e-05,
499
+ "loss": 1.0335,
500
+ "step": 6400
501
+ },
502
+ {
503
+ "epoch": 0.94,
504
+ "learning_rate": 5.948709625060867e-05,
505
+ "loss": 1.0178,
506
+ "step": 6500
507
+ },
508
+ {
509
+ "epoch": 0.94,
510
+ "eval_loss": 0.28046590089797974,
511
+ "eval_runtime": 297.8045,
512
+ "eval_samples_per_second": 19.449,
513
+ "eval_steps_per_second": 0.306,
514
+ "eval_wer": 0.39629834794225943,
515
+ "step": 6500
516
+ },
517
+ {
518
+ "epoch": 0.95,
519
+ "learning_rate": 5.867553968511605e-05,
520
+ "loss": 1.0312,
521
+ "step": 6600
522
+ },
523
+ {
524
+ "epoch": 0.97,
525
+ "learning_rate": 5.786398311962344e-05,
526
+ "loss": 1.033,
527
+ "step": 6700
528
+ },
529
+ {
530
+ "epoch": 0.98,
531
+ "learning_rate": 5.7052426554130825e-05,
532
+ "loss": 1.0289,
533
+ "step": 6800
534
+ },
535
+ {
536
+ "epoch": 1.0,
537
+ "learning_rate": 5.624086998863821e-05,
538
+ "loss": 1.027,
539
+ "step": 6900
540
+ },
541
+ {
542
+ "epoch": 1.01,
543
+ "learning_rate": 5.54293134231456e-05,
544
+ "loss": 0.992,
545
+ "step": 7000
546
+ },
547
+ {
548
+ "epoch": 1.01,
549
+ "eval_loss": 0.2747785747051239,
550
+ "eval_runtime": 298.4841,
551
+ "eval_samples_per_second": 19.405,
552
+ "eval_steps_per_second": 0.305,
553
+ "eval_wer": 0.39352862589201343,
554
+ "step": 7000
555
+ },
556
+ {
557
+ "epoch": 1.03,
558
+ "learning_rate": 5.461775685765298e-05,
559
+ "loss": 1.0025,
560
+ "step": 7100
561
+ },
562
+ {
563
+ "epoch": 1.04,
564
+ "learning_rate": 5.3806200292160366e-05,
565
+ "loss": 1.0122,
566
+ "step": 7200
567
+ },
568
+ {
569
+ "epoch": 1.06,
570
+ "learning_rate": 5.299464372666775e-05,
571
+ "loss": 1.018,
572
+ "step": 7300
573
+ },
574
+ {
575
+ "epoch": 1.07,
576
+ "learning_rate": 5.218308716117514e-05,
577
+ "loss": 0.9936,
578
+ "step": 7400
579
+ },
580
+ {
581
+ "epoch": 1.09,
582
+ "learning_rate": 5.137153059568252e-05,
583
+ "loss": 1.0197,
584
+ "step": 7500
585
+ },
586
+ {
587
+ "epoch": 1.09,
588
+ "eval_loss": 0.26907604932785034,
589
+ "eval_runtime": 298.796,
590
+ "eval_samples_per_second": 19.384,
591
+ "eval_steps_per_second": 0.305,
592
+ "eval_wer": 0.3884453713056796,
593
+ "step": 7500
594
+ },
595
+ {
596
+ "epoch": 1.1,
597
+ "learning_rate": 5.055997403018991e-05,
598
+ "loss": 1.008,
599
+ "step": 7600
600
+ },
601
+ {
602
+ "epoch": 1.11,
603
+ "learning_rate": 4.97484174646973e-05,
604
+ "loss": 1.0028,
605
+ "step": 7700
606
+ },
607
+ {
608
+ "epoch": 1.13,
609
+ "learning_rate": 4.893686089920468e-05,
610
+ "loss": 0.9929,
611
+ "step": 7800
612
+ },
613
+ {
614
+ "epoch": 1.14,
615
+ "learning_rate": 4.8125304333712064e-05,
616
+ "loss": 0.995,
617
+ "step": 7900
618
+ },
619
+ {
620
+ "epoch": 1.16,
621
+ "learning_rate": 4.731374776821945e-05,
622
+ "loss": 1.0056,
623
+ "step": 8000
624
+ },
625
+ {
626
+ "epoch": 1.16,
627
+ "eval_loss": 0.26817116141319275,
628
+ "eval_runtime": 298.1504,
629
+ "eval_samples_per_second": 19.426,
630
+ "eval_steps_per_second": 0.305,
631
+ "eval_wer": 0.3888526833718922,
632
+ "step": 8000
633
+ },
634
+ {
635
+ "epoch": 1.17,
636
+ "learning_rate": 4.6510306768381754e-05,
637
+ "loss": 0.9971,
638
+ "step": 8100
639
+ },
640
+ {
641
+ "epoch": 1.19,
642
+ "learning_rate": 4.5698750202889145e-05,
643
+ "loss": 0.9976,
644
+ "step": 8200
645
+ },
646
+ {
647
+ "epoch": 1.2,
648
+ "learning_rate": 4.488719363739653e-05,
649
+ "loss": 1.0014,
650
+ "step": 8300
651
+ },
652
+ {
653
+ "epoch": 1.22,
654
+ "learning_rate": 4.407563707190391e-05,
655
+ "loss": 0.9835,
656
+ "step": 8400
657
+ },
658
+ {
659
+ "epoch": 1.23,
660
+ "learning_rate": 4.3272196072066225e-05,
661
+ "loss": 0.9826,
662
+ "step": 8500
663
+ },
664
+ {
665
+ "epoch": 1.23,
666
+ "eval_loss": 0.26473307609558105,
667
+ "eval_runtime": 299.0161,
668
+ "eval_samples_per_second": 19.37,
669
+ "eval_steps_per_second": 0.304,
670
+ "eval_wer": 0.38675095311023494,
671
+ "step": 8500
672
+ },
673
+ {
674
+ "epoch": 1.24,
675
+ "learning_rate": 4.2460639506573615e-05,
676
+ "loss": 0.9838,
677
+ "step": 8600
678
+ },
679
+ {
680
+ "epoch": 1.26,
681
+ "learning_rate": 4.1649082941081e-05,
682
+ "loss": 0.9836,
683
+ "step": 8700
684
+ },
685
+ {
686
+ "epoch": 1.27,
687
+ "learning_rate": 4.0845641941243305e-05,
688
+ "loss": 0.9824,
689
+ "step": 8800
690
+ },
691
+ {
692
+ "epoch": 1.29,
693
+ "learning_rate": 4.003408537575069e-05,
694
+ "loss": 0.9715,
695
+ "step": 8900
696
+ },
697
+ {
698
+ "epoch": 1.3,
699
+ "learning_rate": 3.922252881025807e-05,
700
+ "loss": 0.9815,
701
+ "step": 9000
702
+ },
703
+ {
704
+ "epoch": 1.3,
705
+ "eval_loss": 0.26034271717071533,
706
+ "eval_runtime": 299.6782,
707
+ "eval_samples_per_second": 19.327,
708
+ "eval_steps_per_second": 0.304,
709
+ "eval_wer": 0.3831828994102121,
710
+ "step": 9000
711
+ },
712
+ {
713
+ "epoch": 1.32,
714
+ "learning_rate": 3.841097224476546e-05,
715
+ "loss": 0.9757,
716
+ "step": 9100
717
+ },
718
+ {
719
+ "epoch": 1.33,
720
+ "learning_rate": 3.7599415679272846e-05,
721
+ "loss": 0.9689,
722
+ "step": 9200
723
+ },
724
+ {
725
+ "epoch": 1.35,
726
+ "learning_rate": 3.678785911378023e-05,
727
+ "loss": 0.9778,
728
+ "step": 9300
729
+ },
730
+ {
731
+ "epoch": 1.36,
732
+ "learning_rate": 3.5976302548287614e-05,
733
+ "loss": 0.9794,
734
+ "step": 9400
735
+ },
736
+ {
737
+ "epoch": 1.37,
738
+ "learning_rate": 3.5164745982795e-05,
739
+ "loss": 0.9717,
740
+ "step": 9500
741
+ },
742
+ {
743
+ "epoch": 1.37,
744
+ "eval_loss": 0.25609487295150757,
745
+ "eval_runtime": 299.6976,
746
+ "eval_samples_per_second": 19.326,
747
+ "eval_steps_per_second": 0.304,
748
+ "eval_wer": 0.3807064420476392,
749
+ "step": 9500
750
+ },
751
+ {
752
+ "epoch": 1.39,
753
+ "learning_rate": 3.435318941730239e-05,
754
+ "loss": 0.9752,
755
+ "step": 9600
756
+ },
757
+ {
758
+ "epoch": 1.4,
759
+ "learning_rate": 3.354163285180977e-05,
760
+ "loss": 0.965,
761
+ "step": 9700
762
+ },
763
+ {
764
+ "epoch": 1.42,
765
+ "learning_rate": 3.2730076286317155e-05,
766
+ "loss": 0.9522,
767
+ "step": 9800
768
+ },
769
+ {
770
+ "epoch": 1.43,
771
+ "learning_rate": 3.191851972082454e-05,
772
+ "loss": 0.9718,
773
+ "step": 9900
774
+ },
775
+ {
776
+ "epoch": 1.45,
777
+ "learning_rate": 3.110696315533193e-05,
778
+ "loss": 0.9605,
779
+ "step": 10000
780
+ },
781
+ {
782
+ "epoch": 1.45,
783
+ "eval_loss": 0.25231894850730896,
784
+ "eval_runtime": 297.5796,
785
+ "eval_samples_per_second": 19.464,
786
+ "eval_steps_per_second": 0.306,
787
+ "eval_wer": 0.3782951546156603,
788
+ "step": 10000
789
+ },
790
+ {
791
+ "epoch": 1.46,
792
+ "learning_rate": 3.0295406589839315e-05,
793
+ "loss": 0.9635,
794
+ "step": 10100
795
+ },
796
+ {
797
+ "epoch": 1.48,
798
+ "learning_rate": 2.94838500243467e-05,
799
+ "loss": 0.9632,
800
+ "step": 10200
801
+ },
802
+ {
803
+ "epoch": 1.49,
804
+ "learning_rate": 2.8672293458854082e-05,
805
+ "loss": 0.9548,
806
+ "step": 10300
807
+ },
808
+ {
809
+ "epoch": 1.5,
810
+ "learning_rate": 2.7868852459016392e-05,
811
+ "loss": 0.9554,
812
+ "step": 10400
813
+ },
814
+ {
815
+ "epoch": 1.52,
816
+ "learning_rate": 2.7057295893523783e-05,
817
+ "loss": 0.96,
818
+ "step": 10500
819
+ },
820
+ {
821
+ "epoch": 1.52,
822
+ "eval_loss": 0.24939315021038055,
823
+ "eval_runtime": 300.036,
824
+ "eval_samples_per_second": 19.304,
825
+ "eval_steps_per_second": 0.303,
826
+ "eval_wer": 0.3787513441298185,
827
+ "step": 10500
828
+ },
829
+ {
830
+ "epoch": 1.53,
831
+ "learning_rate": 2.6245739328031166e-05,
832
+ "loss": 0.9611,
833
+ "step": 10600
834
+ },
835
+ {
836
+ "epoch": 1.55,
837
+ "learning_rate": 2.543418276253855e-05,
838
+ "loss": 0.9594,
839
+ "step": 10700
840
+ },
841
+ {
842
+ "epoch": 1.56,
843
+ "learning_rate": 2.4622626197045937e-05,
844
+ "loss": 0.9589,
845
+ "step": 10800
846
+ },
847
+ {
848
+ "epoch": 1.58,
849
+ "learning_rate": 2.381106963155332e-05,
850
+ "loss": 0.9441,
851
+ "step": 10900
852
+ },
853
+ {
854
+ "epoch": 1.59,
855
+ "learning_rate": 2.2999513066060704e-05,
856
+ "loss": 0.9442,
857
+ "step": 11000
858
+ },
859
+ {
860
+ "epoch": 1.59,
861
+ "eval_loss": 0.24783751368522644,
862
+ "eval_runtime": 297.7741,
863
+ "eval_samples_per_second": 19.451,
864
+ "eval_steps_per_second": 0.306,
865
+ "eval_wer": 0.3760142070448695,
866
+ "step": 11000
867
+ },
868
+ {
869
+ "epoch": 1.61,
870
+ "learning_rate": 2.218795650056809e-05,
871
+ "loss": 0.9496,
872
+ "step": 11100
873
+ },
874
+ {
875
+ "epoch": 1.62,
876
+ "learning_rate": 2.1376399935075474e-05,
877
+ "loss": 0.9486,
878
+ "step": 11200
879
+ },
880
+ {
881
+ "epoch": 1.64,
882
+ "learning_rate": 2.056484336958286e-05,
883
+ "loss": 0.9558,
884
+ "step": 11300
885
+ },
886
+ {
887
+ "epoch": 1.65,
888
+ "learning_rate": 1.9753286804090245e-05,
889
+ "loss": 0.9486,
890
+ "step": 11400
891
+ },
892
+ {
893
+ "epoch": 1.66,
894
+ "learning_rate": 1.8941730238597632e-05,
895
+ "loss": 0.9564,
896
+ "step": 11500
897
+ },
898
+ {
899
+ "epoch": 1.66,
900
+ "eval_loss": 0.2454409897327423,
901
+ "eval_runtime": 296.0265,
902
+ "eval_samples_per_second": 19.566,
903
+ "eval_steps_per_second": 0.307,
904
+ "eval_wer": 0.3733096549252175,
905
+ "step": 11500
906
+ },
907
+ {
908
+ "epoch": 1.68,
909
+ "learning_rate": 1.8130173673105015e-05,
910
+ "loss": 0.9427,
911
+ "step": 11600
912
+ },
913
+ {
914
+ "epoch": 1.69,
915
+ "learning_rate": 1.73186171076124e-05,
916
+ "loss": 0.9423,
917
+ "step": 11700
918
+ },
919
+ {
920
+ "epoch": 1.71,
921
+ "learning_rate": 1.6507060542119786e-05,
922
+ "loss": 0.9503,
923
+ "step": 11800
924
+ },
925
+ {
926
+ "epoch": 1.72,
927
+ "learning_rate": 1.5695503976627173e-05,
928
+ "loss": 0.9383,
929
+ "step": 11900
930
+ },
931
+ {
932
+ "epoch": 1.74,
933
+ "learning_rate": 1.4883947411134558e-05,
934
+ "loss": 0.9436,
935
+ "step": 12000
936
+ },
937
+ {
938
+ "epoch": 1.74,
939
+ "eval_loss": 0.24390804767608643,
940
+ "eval_runtime": 295.4584,
941
+ "eval_samples_per_second": 19.603,
942
+ "eval_steps_per_second": 0.308,
943
+ "eval_wer": 0.37467822346769203,
944
+ "step": 12000
945
+ },
946
+ {
947
+ "epoch": 1.75,
948
+ "learning_rate": 1.4072390845641942e-05,
949
+ "loss": 0.9491,
950
+ "step": 12100
951
+ },
952
+ {
953
+ "epoch": 1.77,
954
+ "learning_rate": 1.3260834280149325e-05,
955
+ "loss": 0.9419,
956
+ "step": 12200
957
+ },
958
+ {
959
+ "epoch": 1.78,
960
+ "learning_rate": 1.2449277714656712e-05,
961
+ "loss": 0.9517,
962
+ "step": 12300
963
+ },
964
+ {
965
+ "epoch": 1.79,
966
+ "learning_rate": 1.1637721149164097e-05,
967
+ "loss": 0.9367,
968
+ "step": 12400
969
+ },
970
+ {
971
+ "epoch": 1.81,
972
+ "learning_rate": 1.0826164583671483e-05,
973
+ "loss": 0.938,
974
+ "step": 12500
975
+ },
976
+ {
977
+ "epoch": 1.81,
978
+ "eval_loss": 0.24111612141132355,
979
+ "eval_runtime": 296.9314,
980
+ "eval_samples_per_second": 19.506,
981
+ "eval_steps_per_second": 0.306,
982
+ "eval_wer": 0.37159894424712436,
983
+ "step": 12500
984
+ },
985
+ {
986
+ "epoch": 1.82,
987
+ "learning_rate": 1.0014608018178868e-05,
988
+ "loss": 0.9337,
989
+ "step": 12600
990
+ },
991
+ {
992
+ "epoch": 1.84,
993
+ "learning_rate": 9.203051452686253e-06,
994
+ "loss": 0.9284,
995
+ "step": 12700
996
+ },
997
+ {
998
+ "epoch": 1.85,
999
+ "learning_rate": 8.391494887193638e-06,
1000
+ "loss": 0.938,
1001
+ "step": 12800
1002
+ },
1003
+ {
1004
+ "epoch": 1.87,
1005
+ "learning_rate": 7.579938321701023e-06,
1006
+ "loss": 0.9365,
1007
+ "step": 12900
1008
+ },
1009
+ {
1010
+ "epoch": 1.88,
1011
+ "learning_rate": 6.768381756208409e-06,
1012
+ "loss": 0.9353,
1013
+ "step": 13000
1014
+ },
1015
+ {
1016
+ "epoch": 1.88,
1017
+ "eval_loss": 0.23965783417224884,
1018
+ "eval_runtime": 296.6078,
1019
+ "eval_samples_per_second": 19.527,
1020
+ "eval_steps_per_second": 0.307,
1021
+ "eval_wer": 0.3697904786731402,
1022
+ "step": 13000
1023
+ },
1024
+ {
1025
+ "epoch": 1.9,
1026
+ "learning_rate": 5.956825190715793e-06,
1027
+ "loss": 0.9413,
1028
+ "step": 13100
1029
+ },
1030
+ {
1031
+ "epoch": 1.91,
1032
+ "learning_rate": 5.153384190878104e-06,
1033
+ "loss": 0.9356,
1034
+ "step": 13200
1035
+ },
1036
+ {
1037
+ "epoch": 1.92,
1038
+ "learning_rate": 4.34182762538549e-06,
1039
+ "loss": 0.9209,
1040
+ "step": 13300
1041
+ },
1042
+ {
1043
+ "epoch": 1.94,
1044
+ "learning_rate": 3.530271059892875e-06,
1045
+ "loss": 0.9362,
1046
+ "step": 13400
1047
+ },
1048
+ {
1049
+ "epoch": 1.95,
1050
+ "learning_rate": 2.71871449440026e-06,
1051
+ "loss": 0.9271,
1052
+ "step": 13500
1053
+ },
1054
+ {
1055
+ "epoch": 1.95,
1056
+ "eval_loss": 0.23875188827514648,
1057
+ "eval_runtime": 296.9414,
1058
+ "eval_samples_per_second": 19.506,
1059
+ "eval_steps_per_second": 0.306,
1060
+ "eval_wer": 0.3680797679950471,
1061
+ "step": 13500
1062
+ },
1063
+ {
1064
+ "epoch": 1.97,
1065
+ "learning_rate": 1.907157928907645e-06,
1066
+ "loss": 0.9288,
1067
+ "step": 13600
1068
+ },
1069
+ {
1070
+ "epoch": 1.98,
1071
+ "learning_rate": 1.0956013634150302e-06,
1072
+ "loss": 0.9345,
1073
+ "step": 13700
1074
+ },
1075
+ {
1076
+ "epoch": 2.0,
1077
+ "learning_rate": 2.840447979224152e-07,
1078
+ "loss": 0.9326,
1079
+ "step": 13800
1080
+ },
1081
+ {
1082
+ "epoch": 2.0,
1083
+ "step": 13822,
1084
+ "total_flos": 1.2600843645735263e+20,
1085
+ "train_loss": 1.442369053426242,
1086
+ "train_runtime": 53680.5392,
1087
+ "train_samples_per_second": 16.478,
1088
+ "train_steps_per_second": 0.257
1089
+ }
1090
+ ],
1091
+ "max_steps": 13822,
1092
+ "num_train_epochs": 2,
1093
+ "total_flos": 1.2600843645735263e+20,
1094
+ "trial_name": null,
1095
+ "trial_params": null
1096
+ }
added_tokens.json DELETED
@@ -1 +0,0 @@
1
- {}
 
 
config.json CHANGED
@@ -6,7 +6,7 @@
6
  "add_adapter": false,
7
  "apply_spec_augment": true,
8
  "architectures": [
9
- "Wav2Vec2ForCTC"
10
  ],
11
  "attention_dropout": 0.0,
12
  "bos_token_id": 1,
@@ -61,10 +61,10 @@
61
  "layerdrop": 0.0,
62
  "mask_feature_length": 10,
63
  "mask_feature_min_masks": 0,
64
- "mask_feature_prob": 0.4,
65
  "mask_time_length": 10,
66
  "mask_time_min_masks": 2,
67
- "mask_time_prob": 0.75,
68
  "model_type": "wav2vec2",
69
  "num_adapter_layers": 3,
70
  "num_attention_heads": 16,
@@ -76,7 +76,7 @@
76
  "num_hidden_layers": 24,
77
  "num_negatives": 100,
78
  "output_hidden_size": 1024,
79
- "pad_token_id": 216,
80
  "proj_codevector_dim": 768,
81
  "tdnn_dilation": [
82
  1,
@@ -102,6 +102,6 @@
102
  "torch_dtype": "float32",
103
  "transformers_version": "4.17.0.dev0",
104
  "use_weighted_layer_sum": false,
105
- "vocab_size": 218,
106
  "xvector_output_dim": 512
107
  }
 
6
  "add_adapter": false,
7
  "apply_spec_augment": true,
8
  "architectures": [
9
+ "Wav2Vec2ForPreTraining"
10
  ],
11
  "attention_dropout": 0.0,
12
  "bos_token_id": 1,
 
61
  "layerdrop": 0.0,
62
  "mask_feature_length": 10,
63
  "mask_feature_min_masks": 0,
64
+ "mask_feature_prob": 0.25,
65
  "mask_time_length": 10,
66
  "mask_time_min_masks": 2,
67
+ "mask_time_prob": 0.6,
68
  "model_type": "wav2vec2",
69
  "num_adapter_layers": 3,
70
  "num_attention_heads": 16,
 
76
  "num_hidden_layers": 24,
77
  "num_negatives": 100,
78
  "output_hidden_size": 1024,
79
+ "pad_token_id": 218,
80
  "proj_codevector_dim": 768,
81
  "tdnn_dilation": [
82
  1,
 
102
  "torch_dtype": "float32",
103
  "transformers_version": "4.17.0.dev0",
104
  "use_weighted_layer_sum": false,
105
+ "vocab_size": 219,
106
  "xvector_output_dim": 512
107
  }
run.sh CHANGED
@@ -10,8 +10,8 @@ python run_speech_recognition_ctc.py \
10
  --per_device_train_batch_size="64" \
11
  --per_device_eval_batch_size="64" \
12
  --gradient_accumulation_steps="1" \
13
- --learning_rate="2e-4" \
14
- --warmup_steps="2500" \
15
  --length_column_name="input_length" \
16
  --evaluation_strategy="steps" \
17
  --text_column_name="sentence" \
@@ -30,7 +30,7 @@ python run_speech_recognition_ctc.py \
30
  --gradient_checkpointing \
31
  --report_to="wandb" \
32
  --run_name="xls-r-300m-fr" \
33
- --max_eval_samples="6000" \
34
  --max_duration_in_seconds="10" \
35
  --use_auth_token \
36
  --fp16 \
 
10
  --per_device_train_batch_size="64" \
11
  --per_device_eval_batch_size="64" \
12
  --gradient_accumulation_steps="1" \
13
+ --learning_rate="1.5e-4" \
14
+ --warmup_steps="2700" \
15
  --length_column_name="input_length" \
16
  --evaluation_strategy="steps" \
17
  --text_column_name="sentence" \
 
30
  --gradient_checkpointing \
31
  --report_to="wandb" \
32
  --run_name="xls-r-300m-fr" \
33
+ --max_eval_samples="4500" \
34
  --max_duration_in_seconds="10" \
35
  --use_auth_token \
36
  --fp16 \
run_speech_recognition_ctc.py CHANGED
@@ -524,7 +524,7 @@ def main():
524
  # load feature_extractor and tokenizer
525
  tokenizer = Wav2Vec2CTCTokenizer.from_pretrained(
526
  tokenizer_name_or_path,
527
- use_auth_token=data_args.use_auth_token, eos_token=None, bos_token=None,
528
  **tokenizer_kwargs,
529
  )
530
  feature_extractor = AutoFeatureExtractor.from_pretrained(
@@ -729,7 +729,7 @@ def main():
729
  kwargs = {
730
  "finetuned_from": model_args.model_name_or_path,
731
  "tasks": "speech-recognition",
732
- "tags": ["automatic-speech-recognition", data_args.dataset_name],
733
  "dataset_args": f"Config: {config_name}, Training split: {data_args.train_split_name}, Eval split: {data_args.eval_split_name}",
734
  "dataset": f"{data_args.dataset_name.upper()} - {config_name.upper()}",
735
  }
 
524
  # load feature_extractor and tokenizer
525
  tokenizer = Wav2Vec2CTCTokenizer.from_pretrained(
526
  tokenizer_name_or_path,
527
+ use_auth_token=data_args.use_auth_token,
528
  **tokenizer_kwargs,
529
  )
530
  feature_extractor = AutoFeatureExtractor.from_pretrained(
 
729
  kwargs = {
730
  "finetuned_from": model_args.model_name_or_path,
731
  "tasks": "speech-recognition",
732
+ "tags": ["automatic-speech-recognition", "robust-speech-event", data_args.dataset_name],
733
  "dataset_args": f"Config: {config_name}, Training split: {data_args.train_split_name}, Eval split: {data_args.eval_split_name}",
734
  "dataset": f"{data_args.dataset_name.upper()} - {config_name.upper()}",
735
  }
special_tokens_map.json DELETED
@@ -1 +0,0 @@
1
- {"bos_token": null, "eos_token": null, "unk_token": "[UNK]", "pad_token": "[PAD]"}
 
 
tokenizer_config.json DELETED
@@ -1 +0,0 @@
1
- {"unk_token": "[UNK]", "bos_token": null, "eos_token": null, "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "config": null, "tokenizer_type": "wav2vec2", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "./", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
 
 
wandb/debug-internal.log CHANGED
@@ -1 +1 @@
1
- run-20220129_215451-1vipdbow/logs/debug-internal.log
 
1
+ run-20220130_175417-3n44neww/logs/debug-internal.log
wandb/debug.log CHANGED
@@ -1 +1 @@
1
- run-20220129_215451-1vipdbow/logs/debug.log
 
1
+ run-20220130_175417-3n44neww/logs/debug.log
wandb/latest-run CHANGED
@@ -1 +1 @@
1
- run-20220129_215451-1vipdbow
 
1
+ run-20220130_175417-3n44neww
wandb/run-20220130_163246-20w0bl60/files/conda-environment.yaml ADDED
File without changes
wandb/run-20220130_163246-20w0bl60/files/config.yaml ADDED
@@ -0,0 +1,654 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: facebook/wav2vec2-xls-r-300m
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.9
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.8.8
22
+ start_time: 1643560367
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.8.8
37
+ 5: 0.12.9
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ activation_dropout:
42
+ desc: null
43
+ value: 0.05
44
+ adafactor:
45
+ desc: null
46
+ value: false
47
+ adam_beta1:
48
+ desc: null
49
+ value: 0.9
50
+ adam_beta2:
51
+ desc: null
52
+ value: 0.999
53
+ adam_epsilon:
54
+ desc: null
55
+ value: 1.0e-08
56
+ adapter_kernel_size:
57
+ desc: null
58
+ value: 3
59
+ adapter_stride:
60
+ desc: null
61
+ value: 2
62
+ add_adapter:
63
+ desc: null
64
+ value: false
65
+ add_cross_attention:
66
+ desc: null
67
+ value: false
68
+ apply_spec_augment:
69
+ desc: null
70
+ value: true
71
+ architectures:
72
+ desc: null
73
+ value:
74
+ - Wav2Vec2ForPreTraining
75
+ attention_dropout:
76
+ desc: null
77
+ value: 0.0
78
+ bad_words_ids:
79
+ desc: null
80
+ value: null
81
+ bf16:
82
+ desc: null
83
+ value: false
84
+ bf16_full_eval:
85
+ desc: null
86
+ value: false
87
+ bos_token_id:
88
+ desc: null
89
+ value: 1
90
+ chunk_size_feed_forward:
91
+ desc: null
92
+ value: 0
93
+ classifier_proj_size:
94
+ desc: null
95
+ value: 256
96
+ codevector_dim:
97
+ desc: null
98
+ value: 768
99
+ contrastive_logits_temperature:
100
+ desc: null
101
+ value: 0.1
102
+ conv_bias:
103
+ desc: null
104
+ value: true
105
+ conv_dim:
106
+ desc: null
107
+ value:
108
+ - 512
109
+ - 512
110
+ - 512
111
+ - 512
112
+ - 512
113
+ - 512
114
+ - 512
115
+ conv_kernel:
116
+ desc: null
117
+ value:
118
+ - 10
119
+ - 3
120
+ - 3
121
+ - 3
122
+ - 3
123
+ - 2
124
+ - 2
125
+ conv_stride:
126
+ desc: null
127
+ value:
128
+ - 5
129
+ - 2
130
+ - 2
131
+ - 2
132
+ - 2
133
+ - 2
134
+ - 2
135
+ cross_attention_hidden_size:
136
+ desc: null
137
+ value: null
138
+ ctc_loss_reduction:
139
+ desc: null
140
+ value: mean
141
+ ctc_zero_infinity:
142
+ desc: null
143
+ value: false
144
+ dataloader_drop_last:
145
+ desc: null
146
+ value: false
147
+ dataloader_num_workers:
148
+ desc: null
149
+ value: 0
150
+ dataloader_pin_memory:
151
+ desc: null
152
+ value: true
153
+ ddp_bucket_cap_mb:
154
+ desc: null
155
+ value: None
156
+ ddp_find_unused_parameters:
157
+ desc: null
158
+ value: None
159
+ debug:
160
+ desc: null
161
+ value: '[]'
162
+ decoder_start_token_id:
163
+ desc: null
164
+ value: null
165
+ deepspeed:
166
+ desc: null
167
+ value: None
168
+ disable_tqdm:
169
+ desc: null
170
+ value: false
171
+ diversity_loss_weight:
172
+ desc: null
173
+ value: 0.1
174
+ diversity_penalty:
175
+ desc: null
176
+ value: 0.0
177
+ do_eval:
178
+ desc: null
179
+ value: true
180
+ do_predict:
181
+ desc: null
182
+ value: false
183
+ do_sample:
184
+ desc: null
185
+ value: false
186
+ do_stable_layer_norm:
187
+ desc: null
188
+ value: true
189
+ do_train:
190
+ desc: null
191
+ value: true
192
+ early_stopping:
193
+ desc: null
194
+ value: false
195
+ encoder_no_repeat_ngram_size:
196
+ desc: null
197
+ value: 0
198
+ eos_token_id:
199
+ desc: null
200
+ value: 2
201
+ eval_accumulation_steps:
202
+ desc: null
203
+ value: None
204
+ eval_batch_size:
205
+ desc: null
206
+ value: 64
207
+ eval_steps:
208
+ desc: null
209
+ value: 500
210
+ evaluation_strategy:
211
+ desc: null
212
+ value: steps
213
+ feat_extract_activation:
214
+ desc: null
215
+ value: gelu
216
+ feat_extract_dropout:
217
+ desc: null
218
+ value: 0.0
219
+ feat_extract_norm:
220
+ desc: null
221
+ value: layer
222
+ feat_proj_dropout:
223
+ desc: null
224
+ value: 0.0
225
+ feat_quantizer_dropout:
226
+ desc: null
227
+ value: 0.0
228
+ final_dropout:
229
+ desc: null
230
+ value: 0.0
231
+ finetuning_task:
232
+ desc: null
233
+ value: null
234
+ forced_bos_token_id:
235
+ desc: null
236
+ value: null
237
+ forced_eos_token_id:
238
+ desc: null
239
+ value: null
240
+ fp16:
241
+ desc: null
242
+ value: true
243
+ fp16_backend:
244
+ desc: null
245
+ value: auto
246
+ fp16_full_eval:
247
+ desc: null
248
+ value: false
249
+ fp16_opt_level:
250
+ desc: null
251
+ value: O1
252
+ gradient_accumulation_steps:
253
+ desc: null
254
+ value: 1
255
+ gradient_checkpointing:
256
+ desc: null
257
+ value: true
258
+ greater_is_better:
259
+ desc: null
260
+ value: false
261
+ group_by_length:
262
+ desc: null
263
+ value: true
264
+ half_precision_backend:
265
+ desc: null
266
+ value: amp
267
+ hidden_act:
268
+ desc: null
269
+ value: gelu
270
+ hidden_dropout:
271
+ desc: null
272
+ value: 0.0
273
+ hidden_size:
274
+ desc: null
275
+ value: 1024
276
+ hub_model_id:
277
+ desc: null
278
+ value: None
279
+ hub_strategy:
280
+ desc: null
281
+ value: every_save
282
+ hub_token:
283
+ desc: null
284
+ value: <HUB_TOKEN>
285
+ id2label:
286
+ desc: null
287
+ value:
288
+ '0': LABEL_0
289
+ '1': LABEL_1
290
+ ignore_data_skip:
291
+ desc: null
292
+ value: false
293
+ initializer_range:
294
+ desc: null
295
+ value: 0.02
296
+ intermediate_size:
297
+ desc: null
298
+ value: 4096
299
+ is_decoder:
300
+ desc: null
301
+ value: false
302
+ is_encoder_decoder:
303
+ desc: null
304
+ value: false
305
+ label2id:
306
+ desc: null
307
+ value:
308
+ LABEL_0: 0
309
+ LABEL_1: 1
310
+ label_names:
311
+ desc: null
312
+ value: None
313
+ label_smoothing_factor:
314
+ desc: null
315
+ value: 0.0
316
+ layer_norm_eps:
317
+ desc: null
318
+ value: 1.0e-05
319
+ layerdrop:
320
+ desc: null
321
+ value: 0.0
322
+ learning_rate:
323
+ desc: null
324
+ value: 0.0002
325
+ length_column_name:
326
+ desc: null
327
+ value: input_length
328
+ length_penalty:
329
+ desc: null
330
+ value: 1.0
331
+ load_best_model_at_end:
332
+ desc: null
333
+ value: true
334
+ local_rank:
335
+ desc: null
336
+ value: -1
337
+ log_level:
338
+ desc: null
339
+ value: -1
340
+ log_level_replica:
341
+ desc: null
342
+ value: -1
343
+ log_on_each_node:
344
+ desc: null
345
+ value: true
346
+ logging_dir:
347
+ desc: null
348
+ value: ./runs/Jan30_14-50-09_job-1abccd0a-3293-4ffe-8274-9e8f841f653f
349
+ logging_first_step:
350
+ desc: null
351
+ value: false
352
+ logging_nan_inf_filter:
353
+ desc: null
354
+ value: true
355
+ logging_steps:
356
+ desc: null
357
+ value: 100
358
+ logging_strategy:
359
+ desc: null
360
+ value: steps
361
+ lr_scheduler_type:
362
+ desc: null
363
+ value: linear
364
+ mask_feature_length:
365
+ desc: null
366
+ value: 10
367
+ mask_feature_min_masks:
368
+ desc: null
369
+ value: 0
370
+ mask_feature_prob:
371
+ desc: null
372
+ value: 0.25
373
+ mask_time_length:
374
+ desc: null
375
+ value: 10
376
+ mask_time_min_masks:
377
+ desc: null
378
+ value: 2
379
+ mask_time_prob:
380
+ desc: null
381
+ value: 0.6
382
+ max_grad_norm:
383
+ desc: null
384
+ value: 1.0
385
+ max_length:
386
+ desc: null
387
+ value: 20
388
+ max_steps:
389
+ desc: null
390
+ value: -1
391
+ metric_for_best_model:
392
+ desc: null
393
+ value: loss
394
+ min_length:
395
+ desc: null
396
+ value: 0
397
+ model_type:
398
+ desc: null
399
+ value: wav2vec2
400
+ mp_parameters:
401
+ desc: null
402
+ value: ''
403
+ no_cuda:
404
+ desc: null
405
+ value: false
406
+ no_repeat_ngram_size:
407
+ desc: null
408
+ value: 0
409
+ num_adapter_layers:
410
+ desc: null
411
+ value: 3
412
+ num_attention_heads:
413
+ desc: null
414
+ value: 16
415
+ num_beam_groups:
416
+ desc: null
417
+ value: 1
418
+ num_beams:
419
+ desc: null
420
+ value: 1
421
+ num_codevector_groups:
422
+ desc: null
423
+ value: 2
424
+ num_codevectors_per_group:
425
+ desc: null
426
+ value: 320
427
+ num_conv_pos_embedding_groups:
428
+ desc: null
429
+ value: 16
430
+ num_conv_pos_embeddings:
431
+ desc: null
432
+ value: 128
433
+ num_feat_extract_layers:
434
+ desc: null
435
+ value: 7
436
+ num_hidden_layers:
437
+ desc: null
438
+ value: 24
439
+ num_negatives:
440
+ desc: null
441
+ value: 100
442
+ num_return_sequences:
443
+ desc: null
444
+ value: 1
445
+ num_train_epochs:
446
+ desc: null
447
+ value: 4.0
448
+ optim:
449
+ desc: null
450
+ value: adamw_hf
451
+ output_attentions:
452
+ desc: null
453
+ value: false
454
+ output_dir:
455
+ desc: null
456
+ value: ./
457
+ output_hidden_size:
458
+ desc: null
459
+ value: 1024
460
+ output_hidden_states:
461
+ desc: null
462
+ value: false
463
+ output_scores:
464
+ desc: null
465
+ value: false
466
+ overwrite_output_dir:
467
+ desc: null
468
+ value: true
469
+ pad_token_id:
470
+ desc: null
471
+ value: 216
472
+ past_index:
473
+ desc: null
474
+ value: -1
475
+ per_device_eval_batch_size:
476
+ desc: null
477
+ value: 64
478
+ per_device_train_batch_size:
479
+ desc: null
480
+ value: 64
481
+ per_gpu_eval_batch_size:
482
+ desc: null
483
+ value: None
484
+ per_gpu_train_batch_size:
485
+ desc: null
486
+ value: None
487
+ prediction_loss_only:
488
+ desc: null
489
+ value: false
490
+ prefix:
491
+ desc: null
492
+ value: null
493
+ problem_type:
494
+ desc: null
495
+ value: null
496
+ proj_codevector_dim:
497
+ desc: null
498
+ value: 768
499
+ pruned_heads:
500
+ desc: null
501
+ value: {}
502
+ push_to_hub:
503
+ desc: null
504
+ value: true
505
+ push_to_hub_model_id:
506
+ desc: null
507
+ value: None
508
+ push_to_hub_organization:
509
+ desc: null
510
+ value: None
511
+ push_to_hub_token:
512
+ desc: null
513
+ value: <PUSH_TO_HUB_TOKEN>
514
+ remove_invalid_values:
515
+ desc: null
516
+ value: false
517
+ remove_unused_columns:
518
+ desc: null
519
+ value: true
520
+ repetition_penalty:
521
+ desc: null
522
+ value: 1.0
523
+ report_to:
524
+ desc: null
525
+ value: '[''wandb'']'
526
+ resume_from_checkpoint:
527
+ desc: null
528
+ value: None
529
+ return_dict:
530
+ desc: null
531
+ value: true
532
+ return_dict_in_generate:
533
+ desc: null
534
+ value: false
535
+ run_name:
536
+ desc: null
537
+ value: xls-r-300m-fr
538
+ save_on_each_node:
539
+ desc: null
540
+ value: false
541
+ save_steps:
542
+ desc: null
543
+ value: 500
544
+ save_strategy:
545
+ desc: null
546
+ value: steps
547
+ save_total_limit:
548
+ desc: null
549
+ value: 2
550
+ seed:
551
+ desc: null
552
+ value: 42
553
+ sep_token_id:
554
+ desc: null
555
+ value: null
556
+ sharded_ddp:
557
+ desc: null
558
+ value: '[]'
559
+ skip_memory_metrics:
560
+ desc: null
561
+ value: true
562
+ task_specific_params:
563
+ desc: null
564
+ value: null
565
+ tdnn_dilation:
566
+ desc: null
567
+ value:
568
+ - 1
569
+ - 2
570
+ - 3
571
+ - 1
572
+ - 1
573
+ tdnn_dim:
574
+ desc: null
575
+ value:
576
+ - 512
577
+ - 512
578
+ - 512
579
+ - 512
580
+ - 1500
581
+ tdnn_kernel:
582
+ desc: null
583
+ value:
584
+ - 5
585
+ - 3
586
+ - 3
587
+ - 1
588
+ - 1
589
+ temperature:
590
+ desc: null
591
+ value: 1.0
592
+ tf32:
593
+ desc: null
594
+ value: None
595
+ tie_encoder_decoder:
596
+ desc: null
597
+ value: false
598
+ tie_word_embeddings:
599
+ desc: null
600
+ value: true
601
+ tokenizer_class:
602
+ desc: null
603
+ value: null
604
+ top_k:
605
+ desc: null
606
+ value: 50
607
+ top_p:
608
+ desc: null
609
+ value: 1.0
610
+ torch_dtype:
611
+ desc: null
612
+ value: float32
613
+ torchscript:
614
+ desc: null
615
+ value: false
616
+ tpu_metrics_debug:
617
+ desc: null
618
+ value: false
619
+ tpu_num_cores:
620
+ desc: null
621
+ value: None
622
+ train_batch_size:
623
+ desc: null
624
+ value: 64
625
+ transformers_version:
626
+ desc: null
627
+ value: 4.17.0.dev0
628
+ use_bfloat16:
629
+ desc: null
630
+ value: false
631
+ use_legacy_prediction_loop:
632
+ desc: null
633
+ value: false
634
+ use_weighted_layer_sum:
635
+ desc: null
636
+ value: false
637
+ vocab_size:
638
+ desc: null
639
+ value: 216
640
+ warmup_ratio:
641
+ desc: null
642
+ value: 0.0
643
+ warmup_steps:
644
+ desc: null
645
+ value: 2500
646
+ weight_decay:
647
+ desc: null
648
+ value: 0.0
649
+ xpu_backend:
650
+ desc: null
651
+ value: None
652
+ xvector_output_dim:
653
+ desc: null
654
+ value: 512
wandb/run-20220130_163246-20w0bl60/files/output.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ 0%| | 0/27860 [00:00<?, ?it/s]Traceback (most recent call last):
3
+ File "run_speech_recognition_ctc.py", line 748, in <module>
4
+ main()
5
+ File "run_speech_recognition_ctc.py", line 699, in main
6
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
7
+ File "/opt/conda/lib/python3.8/site-packages/transformers/trainer.py", line 1373, in train
8
+ tr_loss_step = self.training_step(model, inputs)
9
+ File "/opt/conda/lib/python3.8/site-packages/transformers/trainer.py", line 1948, in training_step
10
+ loss = self.compute_loss(model, inputs)
11
+ File "/opt/conda/lib/python3.8/site-packages/transformers/trainer.py", line 1980, in compute_loss
12
+ outputs = model(**inputs)
13
+ File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
14
+ return forward_call(*input, **kwargs)
15
+ File "/opt/conda/lib/python3.8/site-packages/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1777, in forward
16
+ loss = nn.functional.ctc_loss(
17
+ File "/opt/conda/lib/python3.8/site-packages/torch/nn/functional.py", line 2460, in ctc_loss
18
+ return torch.ctc_loss(
19
+ RuntimeError: blank must be in label range
wandb/run-20220130_163246-20w0bl60/files/requirements.txt ADDED
@@ -0,0 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ aiohttp==3.8.1
2
+ aiosignal==1.2.0
3
+ analytics-python==1.4.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asgiref==3.5.0
9
+ asttokens==2.0.5
10
+ async-timeout==4.0.2
11
+ attrs==21.4.0
12
+ audioread==2.1.9
13
+ backcall==0.2.0
14
+ backoff==1.10.0
15
+ bcrypt==3.2.0
16
+ beautifulsoup4==4.9.3
17
+ black==21.12b0
18
+ bleach==4.1.0
19
+ brotlipy==0.7.0
20
+ certifi==2020.12.5
21
+ cffi==1.14.3
22
+ chardet==3.0.4
23
+ charset-normalizer==2.0.10
24
+ click==8.0.3
25
+ conda-build==3.21.4
26
+ conda-package-handling==1.7.2
27
+ conda==4.9.2
28
+ configparser==5.2.0
29
+ cryptography==3.2.1
30
+ cycler==0.11.0
31
+ datasets==1.18.2.dev0
32
+ debugpy==1.5.1
33
+ decorator==4.4.2
34
+ defusedxml==0.7.1
35
+ dill==0.3.4
36
+ dnspython==2.1.0
37
+ docker-pycreds==0.4.0
38
+ entrypoints==0.3
39
+ executing==0.8.2
40
+ fastapi==0.73.0
41
+ ffmpy==0.3.0
42
+ filelock==3.0.12
43
+ fonttools==4.29.0
44
+ frozenlist==1.3.0
45
+ fsspec==2022.1.0
46
+ gitdb==4.0.9
47
+ gitpython==3.1.26
48
+ glob2==0.7
49
+ gradio==2.7.5.2
50
+ h11==0.13.0
51
+ huggingface-hub==0.4.0
52
+ idna==2.10
53
+ importlib-resources==5.4.0
54
+ ipykernel==6.7.0
55
+ ipython-genutils==0.2.0
56
+ ipython==8.0.1
57
+ ipywidgets==7.6.3
58
+ jedi==0.17.0
59
+ jinja2==2.11.3
60
+ jiwer==2.3.0
61
+ joblib==1.1.0
62
+ json5==0.9.6
63
+ jsonschema==4.4.0
64
+ jupyter-client==7.1.2
65
+ jupyter-core==4.9.1
66
+ jupyterlab-pygments==0.1.2
67
+ jupyterlab-server==1.2.0
68
+ jupyterlab-widgets==1.0.2
69
+ jupyterlab==2.2.9
70
+ kiwisolver==1.3.2
71
+ libarchive-c==2.9
72
+ librosa==0.8.1
73
+ llvmlite==0.38.0
74
+ markdown2==2.4.2
75
+ markupsafe==1.1.1
76
+ matplotlib-inline==0.1.3
77
+ matplotlib==3.5.1
78
+ mistune==0.8.4
79
+ mkl-fft==1.3.0
80
+ mkl-random==1.1.1
81
+ mkl-service==2.3.0
82
+ monotonic==1.6
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nano==0.10.0
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.19.2
94
+ olefile==0.46
95
+ packaging==21.3
96
+ pandas==1.4.0
97
+ pandocfilters==1.5.0
98
+ paramiko==2.9.2
99
+ parso==0.8.1
100
+ pathspec==0.9.0
101
+ pathtools==0.1.2
102
+ pexpect==4.8.0
103
+ pickleshare==0.7.5
104
+ pillow==8.1.2
105
+ pip==21.3.1
106
+ pkginfo==1.7.0
107
+ platformdirs==2.4.1
108
+ pooch==1.6.0
109
+ prometheus-client==0.13.0
110
+ promise==2.3
111
+ prompt-toolkit==3.0.8
112
+ protobuf==3.19.4
113
+ psutil==5.8.0
114
+ ptyprocess==0.7.0
115
+ pure-eval==0.2.2
116
+ pyarrow==6.0.1
117
+ pycosat==0.6.3
118
+ pycparser==2.20
119
+ pycryptodome==3.13.0
120
+ pydantic==1.9.0
121
+ pydub==0.25.1
122
+ pygments==2.8.0
123
+ pynacl==1.5.0
124
+ pyopenssl==19.1.0
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ pysocks==1.7.1
128
+ python-dateutil==2.8.2
129
+ python-etcd==0.4.5
130
+ python-levenshtein==0.12.2
131
+ python-multipart==0.0.5
132
+ pytz==2021.1
133
+ pyyaml==5.4.1
134
+ pyzmq==22.3.0
135
+ regex==2022.1.18
136
+ requests==2.24.0
137
+ resampy==0.2.2
138
+ ruamel-yaml==0.15.87
139
+ sacremoses==0.0.47
140
+ scikit-learn==1.0.2
141
+ scipy==1.7.3
142
+ send2trash==1.8.0
143
+ sentry-sdk==1.5.4
144
+ setuptools==50.3.1.post20201107
145
+ shortuuid==1.0.8
146
+ six==1.15.0
147
+ smmap==5.0.0
148
+ sniffio==1.2.0
149
+ soundfile==0.10.3.post1
150
+ soupsieve==2.2
151
+ stack-data==0.1.4
152
+ starlette==0.17.1
153
+ subprocess32==3.5.4
154
+ termcolor==1.1.0
155
+ terminado==0.13.1
156
+ testpath==0.5.0
157
+ threadpoolctl==3.0.0
158
+ tokenizers==0.11.4
159
+ tomli==1.2.3
160
+ torch==1.10.2
161
+ torchaudio==0.10.2
162
+ torchelastic==0.2.2
163
+ torchtext==0.9.1
164
+ torchvision==0.9.1
165
+ tornado==6.1
166
+ tqdm==4.62.3
167
+ traitlets==5.1.1
168
+ transformers==4.17.0.dev0
169
+ typing-extensions==4.0.1
170
+ urllib3==1.25.11
171
+ uvicorn==0.17.1
172
+ wandb==0.12.9
173
+ wcwidth==0.2.5
174
+ webencodings==0.5.1
175
+ wheel==0.35.1
176
+ widgetsnbextension==3.5.2
177
+ xxhash==2.0.2
178
+ yarl==1.7.2
179
+ yaspin==2.1.0
180
+ zipp==3.7.0
wandb/run-20220130_163246-20w0bl60/files/wandb-metadata.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.15.0-151-generic-x86_64-with-glibc2.10",
3
+ "python": "3.8.8",
4
+ "heartbeatAt": "2022-01-30T16:32:48.885264",
5
+ "startedAt": "2022-01-30T16:32:46.431923",
6
+ "docker": null,
7
+ "gpu": "Tesla V100S-PCIE-32GB",
8
+ "gpu_count": 1,
9
+ "cpu_count": 60,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=mozilla-foundation/common_voice_8_0",
13
+ "--model_name_or_path=facebook/wav2vec2-xls-r-300m",
14
+ "--dataset_config_name=fr",
15
+ "--tokenizer_name_or_path=./",
16
+ "--output_dir=./",
17
+ "--overwrite_output_dir",
18
+ "--num_train_epochs=4",
19
+ "--per_device_train_batch_size=64",
20
+ "--per_device_eval_batch_size=64",
21
+ "--gradient_accumulation_steps=1",
22
+ "--learning_rate=2e-4",
23
+ "--warmup_steps=2500",
24
+ "--length_column_name=input_length",
25
+ "--evaluation_strategy=steps",
26
+ "--text_column_name=sentence",
27
+ "--save_steps=500",
28
+ "--eval_steps=500",
29
+ "--logging_steps=100",
30
+ "--layerdrop=0.0",
31
+ "--activation_dropout=0.05",
32
+ "--save_total_limit=2",
33
+ "--freeze_feature_encoder",
34
+ "--feat_proj_dropout=0.0",
35
+ "--mask_time_prob=0.6",
36
+ "--mask_time_length=10",
37
+ "--mask_feature_prob=0.25",
38
+ "--mask_feature_length=10",
39
+ "--gradient_checkpointing",
40
+ "--report_to=wandb",
41
+ "--run_name=xls-r-300m-fr",
42
+ "--max_eval_samples=6000",
43
+ "--max_duration_in_seconds=10",
44
+ "--use_auth_token",
45
+ "--fp16",
46
+ "--group_by_length",
47
+ "--preprocessing_num_workers=64",
48
+ "--do_train",
49
+ "--do_eval",
50
+ "--load_best_model_at_end",
51
+ "--push_to_hub"
52
+ ],
53
+ "state": "running",
54
+ "program": "run_speech_recognition_ctc.py",
55
+ "codePath": "run_speech_recognition_ctc.py",
56
+ "git": {
57
+ "remote": "https://huggingface.co/AlexN/xls-r-300m-fr",
58
+ "commit": "9d07727def2251b0909520344156eb711ff096b9"
59
+ },
60
+ "email": "alexandrenap@hotmail.fr",
61
+ "root": "/workspace/xls-r-300m-fr",
62
+ "host": "job-1abccd0a-3293-4ffe-8274-9e8f841f653f",
63
+ "username": "ovh",
64
+ "executable": "/opt/conda/bin/python"
65
+ }
wandb/run-20220130_163246-20w0bl60/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 7}}
wandb/run-20220130_163246-20w0bl60/logs/debug-internal.log ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-01-30 16:32:48,452 INFO MainThread:52211 [internal.py:wandb_internal():87] W&B internal server running at pid: 52211, started at: 2022-01-30 16:32:48.452114
2
+ 2022-01-30 16:32:48,456 INFO WriterThread:52211 [datastore.py:open_for_write():77] open: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/run-20w0bl60.wandb
3
+ 2022-01-30 16:32:48,457 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: check_version
4
+ 2022-01-30 16:32:48,460 DEBUG SenderThread:52211 [sender.py:send():234] send: header
5
+ 2022-01-30 16:32:48,460 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: check_version
6
+ 2022-01-30 16:32:48,657 DEBUG SenderThread:52211 [sender.py:send():234] send: run
7
+ 2022-01-30 16:32:48,873 INFO SenderThread:52211 [dir_watcher.py:__init__():169] watching files in: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files
8
+ 2022-01-30 16:32:48,873 INFO SenderThread:52211 [sender.py:_start_run_threads():804] run started: 20w0bl60 with start time 1643560367
9
+ 2022-01-30 16:32:48,873 DEBUG SenderThread:52211 [sender.py:send():234] send: summary
10
+ 2022-01-30 16:32:48,874 INFO SenderThread:52211 [sender.py:_save_file():939] saving file wandb-summary.json with policy end
11
+ 2022-01-30 16:32:48,876 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: run_start
12
+ 2022-01-30 16:32:48,884 DEBUG HandlerThread:52211 [meta.py:__init__():40] meta init
13
+ 2022-01-30 16:32:48,885 DEBUG HandlerThread:52211 [meta.py:__init__():54] meta init done
14
+ 2022-01-30 16:32:48,885 DEBUG HandlerThread:52211 [meta.py:probe():214] probe
15
+ 2022-01-30 16:32:48,894 DEBUG HandlerThread:52211 [meta.py:_setup_git():204] setup git
16
+ 2022-01-30 16:32:48,923 DEBUG HandlerThread:52211 [meta.py:_setup_git():211] setup git done
17
+ 2022-01-30 16:32:48,923 DEBUG HandlerThread:52211 [meta.py:_save_pip():58] save pip
18
+ 2022-01-30 16:32:48,924 DEBUG HandlerThread:52211 [meta.py:_save_pip():72] save pip done
19
+ 2022-01-30 16:32:48,924 DEBUG HandlerThread:52211 [meta.py:_save_conda():79] save conda
20
+ 2022-01-30 16:32:49,449 DEBUG HandlerThread:52211 [meta.py:_save_conda():89] save conda done
21
+ 2022-01-30 16:32:49,450 DEBUG HandlerThread:52211 [meta.py:probe():252] probe done
22
+ 2022-01-30 16:32:49,458 DEBUG SenderThread:52211 [sender.py:send():234] send: files
23
+ 2022-01-30 16:32:49,459 INFO SenderThread:52211 [sender.py:_save_file():939] saving file wandb-metadata.json with policy now
24
+ 2022-01-30 16:32:49,475 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: stop_status
25
+ 2022-01-30 16:32:49,476 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: stop_status
26
+ 2022-01-30 16:32:49,635 DEBUG SenderThread:52211 [sender.py:send():234] send: config
27
+ 2022-01-30 16:32:49,637 DEBUG SenderThread:52211 [sender.py:send():234] send: metric
28
+ 2022-01-30 16:32:49,639 DEBUG SenderThread:52211 [sender.py:send():234] send: metric
29
+ 2022-01-30 16:32:49,639 WARNING SenderThread:52211 [sender.py:send_metric():897] Seen metric with glob (shouldnt happen)
30
+ 2022-01-30 16:32:49,876 INFO Thread-8 :52211 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/wandb-summary.json
31
+ 2022-01-30 16:32:49,877 INFO Thread-8 :52211 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/output.log
32
+ 2022-01-30 16:32:49,877 INFO Thread-8 :52211 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/wandb-metadata.json
33
+ 2022-01-30 16:32:49,878 INFO Thread-8 :52211 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/requirements.txt
34
+ 2022-01-30 16:32:49,878 INFO Thread-8 :52211 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/conda-environment.yaml
35
+ 2022-01-30 16:32:50,007 INFO Thread-11 :52211 [upload_job.py:push():137] Uploaded file /tmp/tmphz4zp49fwandb/20m3u1mt-wandb-metadata.json
36
+ 2022-01-30 16:32:51,875 INFO Thread-8 :52211 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/output.log
37
+ 2022-01-30 16:32:55,878 INFO Thread-8 :52211 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/output.log
38
+ 2022-01-30 16:32:56,702 DEBUG SenderThread:52211 [sender.py:send():234] send: telemetry
39
+ 2022-01-30 16:32:56,703 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
40
+ 2022-01-30 16:32:56,703 DEBUG SenderThread:52211 [sender.py:send():234] send: exit
41
+ 2022-01-30 16:32:56,703 INFO SenderThread:52211 [sender.py:send_exit():366] handling exit code: 1
42
+ 2022-01-30 16:32:56,704 INFO SenderThread:52211 [sender.py:send_exit():368] handling runtime: 7
43
+ 2022-01-30 16:32:56,704 INFO SenderThread:52211 [sender.py:_save_file():939] saving file wandb-summary.json with policy end
44
+ 2022-01-30 16:32:56,705 INFO SenderThread:52211 [sender.py:send_exit():374] send defer
45
+ 2022-01-30 16:32:56,705 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
46
+ 2022-01-30 16:32:56,706 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
47
+ 2022-01-30 16:32:56,707 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 0
48
+ 2022-01-30 16:32:56,707 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
49
+ 2022-01-30 16:32:56,707 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 0
50
+ 2022-01-30 16:32:56,707 INFO SenderThread:52211 [sender.py:transition_state():387] send defer: 1
51
+ 2022-01-30 16:32:56,708 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
52
+ 2022-01-30 16:32:56,708 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 1
53
+ 2022-01-30 16:32:56,780 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
54
+ 2022-01-30 16:32:56,781 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 1
55
+ 2022-01-30 16:32:56,781 INFO SenderThread:52211 [sender.py:transition_state():387] send defer: 2
56
+ 2022-01-30 16:32:56,781 DEBUG SenderThread:52211 [sender.py:send():234] send: stats
57
+ 2022-01-30 16:32:56,782 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
58
+ 2022-01-30 16:32:56,782 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 2
59
+ 2022-01-30 16:32:56,783 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
60
+ 2022-01-30 16:32:56,783 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 2
61
+ 2022-01-30 16:32:56,783 INFO SenderThread:52211 [sender.py:transition_state():387] send defer: 3
62
+ 2022-01-30 16:32:56,784 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
63
+ 2022-01-30 16:32:56,784 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 3
64
+ 2022-01-30 16:32:56,784 DEBUG SenderThread:52211 [sender.py:send():234] send: summary
65
+ 2022-01-30 16:32:56,785 INFO SenderThread:52211 [sender.py:_save_file():939] saving file wandb-summary.json with policy end
66
+ 2022-01-30 16:32:56,785 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
67
+ 2022-01-30 16:32:56,785 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 3
68
+ 2022-01-30 16:32:56,785 INFO SenderThread:52211 [sender.py:transition_state():387] send defer: 4
69
+ 2022-01-30 16:32:56,786 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
70
+ 2022-01-30 16:32:56,786 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 4
71
+ 2022-01-30 16:32:56,787 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
72
+ 2022-01-30 16:32:56,787 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 4
73
+ 2022-01-30 16:32:56,808 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
74
+ 2022-01-30 16:32:56,879 INFO Thread-8 :52211 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/output.log
75
+ 2022-01-30 16:32:56,879 INFO Thread-8 :52211 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/wandb-summary.json
76
+ 2022-01-30 16:32:57,074 INFO SenderThread:52211 [sender.py:transition_state():387] send defer: 5
77
+ 2022-01-30 16:32:57,074 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
78
+ 2022-01-30 16:32:57,075 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
79
+ 2022-01-30 16:32:57,076 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 5
80
+ 2022-01-30 16:32:57,076 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
81
+ 2022-01-30 16:32:57,076 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 5
82
+ 2022-01-30 16:32:57,076 INFO SenderThread:52211 [dir_watcher.py:finish():283] shutting down directory watcher
83
+ 2022-01-30 16:32:57,177 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
84
+ 2022-01-30 16:32:57,879 INFO Thread-8 :52211 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/config.yaml
85
+ 2022-01-30 16:32:57,881 INFO SenderThread:52211 [dir_watcher.py:finish():313] scan: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files
86
+ 2022-01-30 16:32:57,881 INFO SenderThread:52211 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/output.log output.log
87
+ 2022-01-30 16:32:57,882 INFO SenderThread:52211 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/wandb-metadata.json wandb-metadata.json
88
+ 2022-01-30 16:32:57,882 INFO SenderThread:52211 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/config.yaml config.yaml
89
+ 2022-01-30 16:32:57,888 INFO SenderThread:52211 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/conda-environment.yaml conda-environment.yaml
90
+ 2022-01-30 16:32:57,896 INFO SenderThread:52211 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/wandb-summary.json wandb-summary.json
91
+ 2022-01-30 16:32:57,897 INFO SenderThread:52211 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/requirements.txt requirements.txt
92
+ 2022-01-30 16:32:57,904 INFO SenderThread:52211 [sender.py:transition_state():387] send defer: 6
93
+ 2022-01-30 16:32:57,905 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
94
+ 2022-01-30 16:32:57,907 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
95
+ 2022-01-30 16:32:57,912 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 6
96
+ 2022-01-30 16:32:57,913 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
97
+ 2022-01-30 16:32:57,915 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 6
98
+ 2022-01-30 16:32:57,915 INFO SenderThread:52211 [file_pusher.py:finish():177] shutting down file pusher
99
+ 2022-01-30 16:32:58,007 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
100
+ 2022-01-30 16:32:58,008 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
101
+ 2022-01-30 16:32:58,111 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
102
+ 2022-01-30 16:32:58,111 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
103
+ 2022-01-30 16:32:58,214 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
104
+ 2022-01-30 16:32:58,214 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
105
+ 2022-01-30 16:32:58,317 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
106
+ 2022-01-30 16:32:58,318 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
107
+ 2022-01-30 16:32:58,419 INFO Thread-13 :52211 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/config.yaml
108
+ 2022-01-30 16:32:58,424 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
109
+ 2022-01-30 16:32:58,427 INFO Thread-12 :52211 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/output.log
110
+ 2022-01-30 16:32:58,428 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
111
+ 2022-01-30 16:32:58,471 INFO Thread-15 :52211 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/requirements.txt
112
+ 2022-01-30 16:32:58,515 INFO Thread-14 :52211 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/files/wandb-summary.json
113
+ 2022-01-30 16:32:58,531 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
114
+ 2022-01-30 16:32:58,532 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
115
+ 2022-01-30 16:32:58,634 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
116
+ 2022-01-30 16:32:58,634 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
117
+ 2022-01-30 16:32:58,716 INFO Thread-7 :52211 [sender.py:transition_state():387] send defer: 7
118
+ 2022-01-30 16:32:58,717 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
119
+ 2022-01-30 16:32:58,717 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 7
120
+ 2022-01-30 16:32:58,718 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
121
+ 2022-01-30 16:32:58,718 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 7
122
+ 2022-01-30 16:32:58,736 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
123
+ 2022-01-30 16:32:58,862 INFO SenderThread:52211 [sender.py:transition_state():387] send defer: 8
124
+ 2022-01-30 16:32:58,862 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
125
+ 2022-01-30 16:32:58,863 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
126
+ 2022-01-30 16:32:58,864 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 8
127
+ 2022-01-30 16:32:58,864 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
128
+ 2022-01-30 16:32:58,864 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 8
129
+ 2022-01-30 16:32:58,864 INFO SenderThread:52211 [sender.py:transition_state():387] send defer: 9
130
+ 2022-01-30 16:32:58,866 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: defer
131
+ 2022-01-30 16:32:58,866 DEBUG SenderThread:52211 [sender.py:send():234] send: final
132
+ 2022-01-30 16:32:58,866 INFO HandlerThread:52211 [handler.py:handle_request_defer():147] handle defer: 9
133
+ 2022-01-30 16:32:58,866 DEBUG SenderThread:52211 [sender.py:send():234] send: footer
134
+ 2022-01-30 16:32:58,867 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: defer
135
+ 2022-01-30 16:32:58,867 INFO SenderThread:52211 [sender.py:send_request_defer():383] handle sender defer: 9
136
+ 2022-01-30 16:32:58,965 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: poll_exit
137
+ 2022-01-30 16:32:58,966 DEBUG SenderThread:52211 [sender.py:send_request():248] send_request: poll_exit
138
+ 2022-01-30 16:32:58,967 INFO SenderThread:52211 [file_pusher.py:join():182] waiting for file pusher
139
+ 2022-01-30 16:32:59,277 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: get_summary
140
+ 2022-01-30 16:32:59,279 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: sampled_history
141
+ 2022-01-30 16:32:59,280 DEBUG HandlerThread:52211 [handler.py:handle_request():130] handle_request: shutdown
142
+ 2022-01-30 16:32:59,280 INFO HandlerThread:52211 [handler.py:finish():731] shutting down handler
143
+ 2022-01-30 16:32:59,866 INFO WriterThread:52211 [datastore.py:close():281] close: /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/run-20w0bl60.wandb
144
+ 2022-01-30 16:33:00,275 INFO SenderThread:52211 [sender.py:finish():1070] shutting down sender
145
+ 2022-01-30 16:33:00,275 INFO SenderThread:52211 [file_pusher.py:finish():177] shutting down file pusher
146
+ 2022-01-30 16:33:00,275 INFO SenderThread:52211 [file_pusher.py:join():182] waiting for file pusher
147
+ 2022-01-30 16:33:00,279 INFO MainThread:52211 [internal.py:handle_exit():77] Internal process exited
wandb/run-20220130_163246-20w0bl60/logs/debug.log ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-01-30 16:32:46,436 INFO MainThread:12552 [wandb_setup.py:_flush():71] setting env: {}
2
+ 2022-01-30 16:32:46,437 INFO MainThread:12552 [wandb_setup.py:_flush():71] setting login settings: {}
3
+ 2022-01-30 16:32:46,437 INFO MainThread:12552 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/logs/debug.log
4
+ 2022-01-30 16:32:46,437 INFO MainThread:12552 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220130_163246-20w0bl60/logs/debug-internal.log
5
+ 2022-01-30 16:32:46,438 INFO MainThread:12552 [wandb_init.py:init():404] calling init triggers
6
+ 2022-01-30 16:32:46,438 INFO MainThread:12552 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
7
+ config: {}
8
+ 2022-01-30 16:32:46,438 INFO MainThread:12552 [wandb_init.py:init():460] starting backend
9
+ 2022-01-30 16:32:46,438 INFO MainThread:12552 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
10
+ 2022-01-30 16:32:46,803 INFO MainThread:12552 [backend.py:ensure_launched():216] starting backend process...
11
+ 2022-01-30 16:32:47,202 INFO MainThread:12552 [backend.py:ensure_launched():221] started backend process with pid: 52211
12
+ 2022-01-30 16:32:47,205 INFO MainThread:12552 [wandb_init.py:init():469] backend started and connected
13
+ 2022-01-30 16:32:47,215 INFO MainThread:12552 [wandb_init.py:init():533] updated telemetry
14
+ 2022-01-30 16:32:48,051 INFO MainThread:12552 [wandb_init.py:init():563] communicating current version
15
+ 2022-01-30 16:32:48,654 INFO MainThread:12552 [wandb_init.py:init():568] got version response
16
+ 2022-01-30 16:32:48,655 INFO MainThread:12552 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
17
+ 2022-01-30 16:32:48,875 INFO MainThread:12552 [wandb_init.py:init():606] starting run threads in backend
18
+ 2022-01-30 16:32:49,475 INFO MainThread:12552 [wandb_run.py:_console_start():1810] atexit reg
19
+ 2022-01-30 16:32:49,476 INFO MainThread:12552 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
20
+ 2022-01-30 16:32:49,477 INFO MainThread:12552 [wandb_run.py:_redirect():1689] Redirecting console.
21
+ 2022-01-30 16:32:49,483 INFO MainThread:12552 [wandb_run.py:_redirect():1745] Redirects installed.
22
+ 2022-01-30 16:32:49,483 INFO MainThread:12552 [wandb_init.py:init():633] run started, returning control to user process
23
+ 2022-01-30 16:32:49,485 INFO MainThread:12552 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 216, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 216, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.6, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 4.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jan30_14-50-09_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
24
+ 2022-01-30 16:32:49,490 INFO MainThread:12552 [wandb_watch.py:watch():43] Watching
25
+ 2022-01-30 16:32:54,620 INFO MainThread:12552 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 1
26
+ 2022-01-30 16:32:54,623 INFO MainThread:12552 [wandb_run.py:_restore():1752] restore
27
+ 2022-01-30 16:32:56,707 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
28
+ wandb_count: 1
29
+ }
30
+ pusher_stats {
31
+ uploaded_bytes: 2180
32
+ total_bytes: 2180
33
+ }
34
+
35
+ 2022-01-30 16:32:57,075 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
36
+ wandb_count: 1
37
+ }
38
+ pusher_stats {
39
+ uploaded_bytes: 2180
40
+ total_bytes: 2180
41
+ }
42
+
43
+ 2022-01-30 16:32:57,906 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
44
+ wandb_count: 4
45
+ }
46
+ pusher_stats {
47
+ uploaded_bytes: 2180
48
+ total_bytes: 12932
49
+ }
50
+
51
+ 2022-01-30 16:32:58,009 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
52
+ wandb_count: 5
53
+ }
54
+ pusher_stats {
55
+ uploaded_bytes: 2180
56
+ total_bytes: 16065
57
+ }
58
+
59
+ 2022-01-30 16:32:58,112 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
60
+ wandb_count: 5
61
+ }
62
+ pusher_stats {
63
+ uploaded_bytes: 3508
64
+ total_bytes: 16065
65
+ }
66
+
67
+ 2022-01-30 16:32:58,216 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
68
+ wandb_count: 5
69
+ }
70
+ pusher_stats {
71
+ uploaded_bytes: 16065
72
+ total_bytes: 16065
73
+ }
74
+
75
+ 2022-01-30 16:32:58,319 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
76
+ wandb_count: 5
77
+ }
78
+ pusher_stats {
79
+ uploaded_bytes: 16065
80
+ total_bytes: 16065
81
+ }
82
+
83
+ 2022-01-30 16:32:58,430 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
84
+ wandb_count: 5
85
+ }
86
+ pusher_stats {
87
+ uploaded_bytes: 16065
88
+ total_bytes: 16065
89
+ }
90
+
91
+ 2022-01-30 16:32:58,533 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
92
+ wandb_count: 5
93
+ }
94
+ pusher_stats {
95
+ uploaded_bytes: 16065
96
+ total_bytes: 16065
97
+ }
98
+
99
+ 2022-01-30 16:32:58,635 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
100
+ wandb_count: 5
101
+ }
102
+ pusher_stats {
103
+ uploaded_bytes: 16065
104
+ total_bytes: 16065
105
+ }
106
+
107
+ 2022-01-30 16:32:58,864 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
108
+ wandb_count: 5
109
+ }
110
+ pusher_stats {
111
+ uploaded_bytes: 16065
112
+ total_bytes: 16065
113
+ }
114
+
115
+ 2022-01-30 16:32:59,276 INFO MainThread:12552 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
116
+ exit_result {
117
+ }
118
+ file_counts {
119
+ wandb_count: 5
120
+ }
121
+ pusher_stats {
122
+ uploaded_bytes: 16065
123
+ total_bytes: 16065
124
+ }
125
+ local_info {
126
+ }
127
+
128
+ 2022-01-30 16:33:00,418 INFO MainThread:12552 [wandb_run.py:_append_files():2180] logging synced files
wandb/run-20220130_163246-20w0bl60/run-20w0bl60.wandb ADDED
Binary file (7.91 kB). View file
 
wandb/run-20220130_174749-kzuygp3o/files/conda-environment.yaml ADDED
File without changes
wandb/run-20220130_174749-kzuygp3o/files/config.yaml ADDED
@@ -0,0 +1,654 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: facebook/wav2vec2-xls-r-300m
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.9
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.8.8
22
+ start_time: 1643564869
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.8.8
37
+ 5: 0.12.9
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ activation_dropout:
42
+ desc: null
43
+ value: 0.05
44
+ adafactor:
45
+ desc: null
46
+ value: false
47
+ adam_beta1:
48
+ desc: null
49
+ value: 0.9
50
+ adam_beta2:
51
+ desc: null
52
+ value: 0.999
53
+ adam_epsilon:
54
+ desc: null
55
+ value: 1.0e-08
56
+ adapter_kernel_size:
57
+ desc: null
58
+ value: 3
59
+ adapter_stride:
60
+ desc: null
61
+ value: 2
62
+ add_adapter:
63
+ desc: null
64
+ value: false
65
+ add_cross_attention:
66
+ desc: null
67
+ value: false
68
+ apply_spec_augment:
69
+ desc: null
70
+ value: true
71
+ architectures:
72
+ desc: null
73
+ value:
74
+ - Wav2Vec2ForPreTraining
75
+ attention_dropout:
76
+ desc: null
77
+ value: 0.0
78
+ bad_words_ids:
79
+ desc: null
80
+ value: null
81
+ bf16:
82
+ desc: null
83
+ value: false
84
+ bf16_full_eval:
85
+ desc: null
86
+ value: false
87
+ bos_token_id:
88
+ desc: null
89
+ value: 1
90
+ chunk_size_feed_forward:
91
+ desc: null
92
+ value: 0
93
+ classifier_proj_size:
94
+ desc: null
95
+ value: 256
96
+ codevector_dim:
97
+ desc: null
98
+ value: 768
99
+ contrastive_logits_temperature:
100
+ desc: null
101
+ value: 0.1
102
+ conv_bias:
103
+ desc: null
104
+ value: true
105
+ conv_dim:
106
+ desc: null
107
+ value:
108
+ - 512
109
+ - 512
110
+ - 512
111
+ - 512
112
+ - 512
113
+ - 512
114
+ - 512
115
+ conv_kernel:
116
+ desc: null
117
+ value:
118
+ - 10
119
+ - 3
120
+ - 3
121
+ - 3
122
+ - 3
123
+ - 2
124
+ - 2
125
+ conv_stride:
126
+ desc: null
127
+ value:
128
+ - 5
129
+ - 2
130
+ - 2
131
+ - 2
132
+ - 2
133
+ - 2
134
+ - 2
135
+ cross_attention_hidden_size:
136
+ desc: null
137
+ value: null
138
+ ctc_loss_reduction:
139
+ desc: null
140
+ value: mean
141
+ ctc_zero_infinity:
142
+ desc: null
143
+ value: false
144
+ dataloader_drop_last:
145
+ desc: null
146
+ value: false
147
+ dataloader_num_workers:
148
+ desc: null
149
+ value: 0
150
+ dataloader_pin_memory:
151
+ desc: null
152
+ value: true
153
+ ddp_bucket_cap_mb:
154
+ desc: null
155
+ value: None
156
+ ddp_find_unused_parameters:
157
+ desc: null
158
+ value: None
159
+ debug:
160
+ desc: null
161
+ value: '[]'
162
+ decoder_start_token_id:
163
+ desc: null
164
+ value: null
165
+ deepspeed:
166
+ desc: null
167
+ value: None
168
+ disable_tqdm:
169
+ desc: null
170
+ value: false
171
+ diversity_loss_weight:
172
+ desc: null
173
+ value: 0.1
174
+ diversity_penalty:
175
+ desc: null
176
+ value: 0.0
177
+ do_eval:
178
+ desc: null
179
+ value: true
180
+ do_predict:
181
+ desc: null
182
+ value: false
183
+ do_sample:
184
+ desc: null
185
+ value: false
186
+ do_stable_layer_norm:
187
+ desc: null
188
+ value: true
189
+ do_train:
190
+ desc: null
191
+ value: true
192
+ early_stopping:
193
+ desc: null
194
+ value: false
195
+ encoder_no_repeat_ngram_size:
196
+ desc: null
197
+ value: 0
198
+ eos_token_id:
199
+ desc: null
200
+ value: 2
201
+ eval_accumulation_steps:
202
+ desc: null
203
+ value: None
204
+ eval_batch_size:
205
+ desc: null
206
+ value: 64
207
+ eval_steps:
208
+ desc: null
209
+ value: 500
210
+ evaluation_strategy:
211
+ desc: null
212
+ value: steps
213
+ feat_extract_activation:
214
+ desc: null
215
+ value: gelu
216
+ feat_extract_dropout:
217
+ desc: null
218
+ value: 0.0
219
+ feat_extract_norm:
220
+ desc: null
221
+ value: layer
222
+ feat_proj_dropout:
223
+ desc: null
224
+ value: 0.0
225
+ feat_quantizer_dropout:
226
+ desc: null
227
+ value: 0.0
228
+ final_dropout:
229
+ desc: null
230
+ value: 0.0
231
+ finetuning_task:
232
+ desc: null
233
+ value: null
234
+ forced_bos_token_id:
235
+ desc: null
236
+ value: null
237
+ forced_eos_token_id:
238
+ desc: null
239
+ value: null
240
+ fp16:
241
+ desc: null
242
+ value: true
243
+ fp16_backend:
244
+ desc: null
245
+ value: auto
246
+ fp16_full_eval:
247
+ desc: null
248
+ value: false
249
+ fp16_opt_level:
250
+ desc: null
251
+ value: O1
252
+ gradient_accumulation_steps:
253
+ desc: null
254
+ value: 1
255
+ gradient_checkpointing:
256
+ desc: null
257
+ value: true
258
+ greater_is_better:
259
+ desc: null
260
+ value: false
261
+ group_by_length:
262
+ desc: null
263
+ value: true
264
+ half_precision_backend:
265
+ desc: null
266
+ value: amp
267
+ hidden_act:
268
+ desc: null
269
+ value: gelu
270
+ hidden_dropout:
271
+ desc: null
272
+ value: 0.0
273
+ hidden_size:
274
+ desc: null
275
+ value: 1024
276
+ hub_model_id:
277
+ desc: null
278
+ value: None
279
+ hub_strategy:
280
+ desc: null
281
+ value: every_save
282
+ hub_token:
283
+ desc: null
284
+ value: <HUB_TOKEN>
285
+ id2label:
286
+ desc: null
287
+ value:
288
+ '0': LABEL_0
289
+ '1': LABEL_1
290
+ ignore_data_skip:
291
+ desc: null
292
+ value: false
293
+ initializer_range:
294
+ desc: null
295
+ value: 0.02
296
+ intermediate_size:
297
+ desc: null
298
+ value: 4096
299
+ is_decoder:
300
+ desc: null
301
+ value: false
302
+ is_encoder_decoder:
303
+ desc: null
304
+ value: false
305
+ label2id:
306
+ desc: null
307
+ value:
308
+ LABEL_0: 0
309
+ LABEL_1: 1
310
+ label_names:
311
+ desc: null
312
+ value: None
313
+ label_smoothing_factor:
314
+ desc: null
315
+ value: 0.0
316
+ layer_norm_eps:
317
+ desc: null
318
+ value: 1.0e-05
319
+ layerdrop:
320
+ desc: null
321
+ value: 0.0
322
+ learning_rate:
323
+ desc: null
324
+ value: 0.0002
325
+ length_column_name:
326
+ desc: null
327
+ value: input_length
328
+ length_penalty:
329
+ desc: null
330
+ value: 1.0
331
+ load_best_model_at_end:
332
+ desc: null
333
+ value: true
334
+ local_rank:
335
+ desc: null
336
+ value: -1
337
+ log_level:
338
+ desc: null
339
+ value: -1
340
+ log_level_replica:
341
+ desc: null
342
+ value: -1
343
+ log_on_each_node:
344
+ desc: null
345
+ value: true
346
+ logging_dir:
347
+ desc: null
348
+ value: ./runs/Jan30_17-46-00_job-1abccd0a-3293-4ffe-8274-9e8f841f653f
349
+ logging_first_step:
350
+ desc: null
351
+ value: false
352
+ logging_nan_inf_filter:
353
+ desc: null
354
+ value: true
355
+ logging_steps:
356
+ desc: null
357
+ value: 100
358
+ logging_strategy:
359
+ desc: null
360
+ value: steps
361
+ lr_scheduler_type:
362
+ desc: null
363
+ value: linear
364
+ mask_feature_length:
365
+ desc: null
366
+ value: 10
367
+ mask_feature_min_masks:
368
+ desc: null
369
+ value: 0
370
+ mask_feature_prob:
371
+ desc: null
372
+ value: 0.25
373
+ mask_time_length:
374
+ desc: null
375
+ value: 10
376
+ mask_time_min_masks:
377
+ desc: null
378
+ value: 2
379
+ mask_time_prob:
380
+ desc: null
381
+ value: 0.6
382
+ max_grad_norm:
383
+ desc: null
384
+ value: 1.0
385
+ max_length:
386
+ desc: null
387
+ value: 20
388
+ max_steps:
389
+ desc: null
390
+ value: -1
391
+ metric_for_best_model:
392
+ desc: null
393
+ value: loss
394
+ min_length:
395
+ desc: null
396
+ value: 0
397
+ model_type:
398
+ desc: null
399
+ value: wav2vec2
400
+ mp_parameters:
401
+ desc: null
402
+ value: ''
403
+ no_cuda:
404
+ desc: null
405
+ value: false
406
+ no_repeat_ngram_size:
407
+ desc: null
408
+ value: 0
409
+ num_adapter_layers:
410
+ desc: null
411
+ value: 3
412
+ num_attention_heads:
413
+ desc: null
414
+ value: 16
415
+ num_beam_groups:
416
+ desc: null
417
+ value: 1
418
+ num_beams:
419
+ desc: null
420
+ value: 1
421
+ num_codevector_groups:
422
+ desc: null
423
+ value: 2
424
+ num_codevectors_per_group:
425
+ desc: null
426
+ value: 320
427
+ num_conv_pos_embedding_groups:
428
+ desc: null
429
+ value: 16
430
+ num_conv_pos_embeddings:
431
+ desc: null
432
+ value: 128
433
+ num_feat_extract_layers:
434
+ desc: null
435
+ value: 7
436
+ num_hidden_layers:
437
+ desc: null
438
+ value: 24
439
+ num_negatives:
440
+ desc: null
441
+ value: 100
442
+ num_return_sequences:
443
+ desc: null
444
+ value: 1
445
+ num_train_epochs:
446
+ desc: null
447
+ value: 4.0
448
+ optim:
449
+ desc: null
450
+ value: adamw_hf
451
+ output_attentions:
452
+ desc: null
453
+ value: false
454
+ output_dir:
455
+ desc: null
456
+ value: ./
457
+ output_hidden_size:
458
+ desc: null
459
+ value: 1024
460
+ output_hidden_states:
461
+ desc: null
462
+ value: false
463
+ output_scores:
464
+ desc: null
465
+ value: false
466
+ overwrite_output_dir:
467
+ desc: null
468
+ value: true
469
+ pad_token_id:
470
+ desc: null
471
+ value: 216
472
+ past_index:
473
+ desc: null
474
+ value: -1
475
+ per_device_eval_batch_size:
476
+ desc: null
477
+ value: 64
478
+ per_device_train_batch_size:
479
+ desc: null
480
+ value: 64
481
+ per_gpu_eval_batch_size:
482
+ desc: null
483
+ value: None
484
+ per_gpu_train_batch_size:
485
+ desc: null
486
+ value: None
487
+ prediction_loss_only:
488
+ desc: null
489
+ value: false
490
+ prefix:
491
+ desc: null
492
+ value: null
493
+ problem_type:
494
+ desc: null
495
+ value: null
496
+ proj_codevector_dim:
497
+ desc: null
498
+ value: 768
499
+ pruned_heads:
500
+ desc: null
501
+ value: {}
502
+ push_to_hub:
503
+ desc: null
504
+ value: true
505
+ push_to_hub_model_id:
506
+ desc: null
507
+ value: None
508
+ push_to_hub_organization:
509
+ desc: null
510
+ value: None
511
+ push_to_hub_token:
512
+ desc: null
513
+ value: <PUSH_TO_HUB_TOKEN>
514
+ remove_invalid_values:
515
+ desc: null
516
+ value: false
517
+ remove_unused_columns:
518
+ desc: null
519
+ value: true
520
+ repetition_penalty:
521
+ desc: null
522
+ value: 1.0
523
+ report_to:
524
+ desc: null
525
+ value: '[''wandb'']'
526
+ resume_from_checkpoint:
527
+ desc: null
528
+ value: None
529
+ return_dict:
530
+ desc: null
531
+ value: true
532
+ return_dict_in_generate:
533
+ desc: null
534
+ value: false
535
+ run_name:
536
+ desc: null
537
+ value: xls-r-300m-fr
538
+ save_on_each_node:
539
+ desc: null
540
+ value: false
541
+ save_steps:
542
+ desc: null
543
+ value: 500
544
+ save_strategy:
545
+ desc: null
546
+ value: steps
547
+ save_total_limit:
548
+ desc: null
549
+ value: 2
550
+ seed:
551
+ desc: null
552
+ value: 42
553
+ sep_token_id:
554
+ desc: null
555
+ value: null
556
+ sharded_ddp:
557
+ desc: null
558
+ value: '[]'
559
+ skip_memory_metrics:
560
+ desc: null
561
+ value: true
562
+ task_specific_params:
563
+ desc: null
564
+ value: null
565
+ tdnn_dilation:
566
+ desc: null
567
+ value:
568
+ - 1
569
+ - 2
570
+ - 3
571
+ - 1
572
+ - 1
573
+ tdnn_dim:
574
+ desc: null
575
+ value:
576
+ - 512
577
+ - 512
578
+ - 512
579
+ - 512
580
+ - 1500
581
+ tdnn_kernel:
582
+ desc: null
583
+ value:
584
+ - 5
585
+ - 3
586
+ - 3
587
+ - 1
588
+ - 1
589
+ temperature:
590
+ desc: null
591
+ value: 1.0
592
+ tf32:
593
+ desc: null
594
+ value: None
595
+ tie_encoder_decoder:
596
+ desc: null
597
+ value: false
598
+ tie_word_embeddings:
599
+ desc: null
600
+ value: true
601
+ tokenizer_class:
602
+ desc: null
603
+ value: null
604
+ top_k:
605
+ desc: null
606
+ value: 50
607
+ top_p:
608
+ desc: null
609
+ value: 1.0
610
+ torch_dtype:
611
+ desc: null
612
+ value: float32
613
+ torchscript:
614
+ desc: null
615
+ value: false
616
+ tpu_metrics_debug:
617
+ desc: null
618
+ value: false
619
+ tpu_num_cores:
620
+ desc: null
621
+ value: None
622
+ train_batch_size:
623
+ desc: null
624
+ value: 64
625
+ transformers_version:
626
+ desc: null
627
+ value: 4.17.0.dev0
628
+ use_bfloat16:
629
+ desc: null
630
+ value: false
631
+ use_legacy_prediction_loop:
632
+ desc: null
633
+ value: false
634
+ use_weighted_layer_sum:
635
+ desc: null
636
+ value: false
637
+ vocab_size:
638
+ desc: null
639
+ value: 216
640
+ warmup_ratio:
641
+ desc: null
642
+ value: 0.0
643
+ warmup_steps:
644
+ desc: null
645
+ value: 2500
646
+ weight_decay:
647
+ desc: null
648
+ value: 0.0
649
+ xpu_backend:
650
+ desc: null
651
+ value: None
652
+ xvector_output_dim:
653
+ desc: null
654
+ value: 512
wandb/run-20220130_174749-kzuygp3o/files/output.log ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ 0%| | 0/64 [00:00<?, ?it/s]Traceback (most recent call last):
3
+ File "run_speech_recognition_ctc.py", line 748, in <module>
4
+ main()
5
+ File "run_speech_recognition_ctc.py", line 699, in main
6
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
7
+ File "/opt/conda/lib/python3.8/site-packages/transformers/trainer.py", line 1373, in train
8
+ tr_loss_step = self.training_step(model, inputs)
9
+ File "/opt/conda/lib/python3.8/site-packages/transformers/trainer.py", line 1948, in training_step
10
+ loss = self.compute_loss(model, inputs)
11
+ File "/opt/conda/lib/python3.8/site-packages/transformers/trainer.py", line 1980, in compute_loss
12
+ outputs = model(**inputs)
13
+ File "/opt/conda/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1102, in _call_impl
14
+ return forward_call(*input, **kwargs)
15
+ File "/opt/conda/lib/python3.8/site-packages/transformers/models/wav2vec2/modeling_wav2vec2.py", line 1777, in forward
16
+ loss = nn.functional.ctc_loss(
17
+ File "/opt/conda/lib/python3.8/site-packages/torch/nn/functional.py", line 2460, in ctc_loss
18
+ return torch.ctc_loss(
19
+ RuntimeError: blank must be in label range
wandb/run-20220130_174749-kzuygp3o/files/requirements.txt ADDED
@@ -0,0 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ aiohttp==3.8.1
2
+ aiosignal==1.2.0
3
+ analytics-python==1.4.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asgiref==3.5.0
9
+ asttokens==2.0.5
10
+ async-timeout==4.0.2
11
+ attrs==21.4.0
12
+ audioread==2.1.9
13
+ backcall==0.2.0
14
+ backoff==1.10.0
15
+ bcrypt==3.2.0
16
+ beautifulsoup4==4.9.3
17
+ black==21.12b0
18
+ bleach==4.1.0
19
+ brotlipy==0.7.0
20
+ certifi==2020.12.5
21
+ cffi==1.14.3
22
+ chardet==3.0.4
23
+ charset-normalizer==2.0.10
24
+ click==8.0.3
25
+ conda-build==3.21.4
26
+ conda-package-handling==1.7.2
27
+ conda==4.9.2
28
+ configparser==5.2.0
29
+ cryptography==3.2.1
30
+ cycler==0.11.0
31
+ datasets==1.18.2.dev0
32
+ debugpy==1.5.1
33
+ decorator==4.4.2
34
+ defusedxml==0.7.1
35
+ dill==0.3.4
36
+ dnspython==2.1.0
37
+ docker-pycreds==0.4.0
38
+ entrypoints==0.3
39
+ executing==0.8.2
40
+ fastapi==0.73.0
41
+ ffmpy==0.3.0
42
+ filelock==3.0.12
43
+ fonttools==4.29.0
44
+ frozenlist==1.3.0
45
+ fsspec==2022.1.0
46
+ gitdb==4.0.9
47
+ gitpython==3.1.26
48
+ glob2==0.7
49
+ gradio==2.7.5.2
50
+ h11==0.13.0
51
+ huggingface-hub==0.4.0
52
+ idna==2.10
53
+ importlib-resources==5.4.0
54
+ ipykernel==6.7.0
55
+ ipython-genutils==0.2.0
56
+ ipython==8.0.1
57
+ ipywidgets==7.6.3
58
+ jedi==0.17.0
59
+ jinja2==2.11.3
60
+ jiwer==2.3.0
61
+ joblib==1.1.0
62
+ json5==0.9.6
63
+ jsonschema==4.4.0
64
+ jupyter-client==7.1.2
65
+ jupyter-core==4.9.1
66
+ jupyterlab-pygments==0.1.2
67
+ jupyterlab-server==1.2.0
68
+ jupyterlab-widgets==1.0.2
69
+ jupyterlab==2.2.9
70
+ kiwisolver==1.3.2
71
+ libarchive-c==2.9
72
+ librosa==0.8.1
73
+ llvmlite==0.38.0
74
+ markdown2==2.4.2
75
+ markupsafe==1.1.1
76
+ matplotlib-inline==0.1.3
77
+ matplotlib==3.5.1
78
+ mistune==0.8.4
79
+ mkl-fft==1.3.0
80
+ mkl-random==1.1.1
81
+ mkl-service==2.3.0
82
+ monotonic==1.6
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nano==0.10.0
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.19.2
94
+ olefile==0.46
95
+ packaging==21.3
96
+ pandas==1.4.0
97
+ pandocfilters==1.5.0
98
+ paramiko==2.9.2
99
+ parso==0.8.1
100
+ pathspec==0.9.0
101
+ pathtools==0.1.2
102
+ pexpect==4.8.0
103
+ pickleshare==0.7.5
104
+ pillow==8.1.2
105
+ pip==21.3.1
106
+ pkginfo==1.7.0
107
+ platformdirs==2.4.1
108
+ pooch==1.6.0
109
+ prometheus-client==0.13.0
110
+ promise==2.3
111
+ prompt-toolkit==3.0.8
112
+ protobuf==3.19.4
113
+ psutil==5.8.0
114
+ ptyprocess==0.7.0
115
+ pure-eval==0.2.2
116
+ pyarrow==6.0.1
117
+ pycosat==0.6.3
118
+ pycparser==2.20
119
+ pycryptodome==3.13.0
120
+ pydantic==1.9.0
121
+ pydub==0.25.1
122
+ pygments==2.8.0
123
+ pynacl==1.5.0
124
+ pyopenssl==19.1.0
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ pysocks==1.7.1
128
+ python-dateutil==2.8.2
129
+ python-etcd==0.4.5
130
+ python-levenshtein==0.12.2
131
+ python-multipart==0.0.5
132
+ pytz==2021.1
133
+ pyyaml==5.4.1
134
+ pyzmq==22.3.0
135
+ regex==2022.1.18
136
+ requests==2.24.0
137
+ resampy==0.2.2
138
+ ruamel-yaml==0.15.87
139
+ sacremoses==0.0.47
140
+ scikit-learn==1.0.2
141
+ scipy==1.7.3
142
+ send2trash==1.8.0
143
+ sentry-sdk==1.5.4
144
+ setuptools==50.3.1.post20201107
145
+ shortuuid==1.0.8
146
+ six==1.15.0
147
+ smmap==5.0.0
148
+ sniffio==1.2.0
149
+ soundfile==0.10.3.post1
150
+ soupsieve==2.2
151
+ stack-data==0.1.4
152
+ starlette==0.17.1
153
+ subprocess32==3.5.4
154
+ termcolor==1.1.0
155
+ terminado==0.13.1
156
+ testpath==0.5.0
157
+ threadpoolctl==3.0.0
158
+ tokenizers==0.11.4
159
+ tomli==1.2.3
160
+ torch==1.10.2
161
+ torchaudio==0.10.2
162
+ torchelastic==0.2.2
163
+ torchtext==0.9.1
164
+ torchvision==0.9.1
165
+ tornado==6.1
166
+ tqdm==4.62.3
167
+ traitlets==5.1.1
168
+ transformers==4.17.0.dev0
169
+ typing-extensions==4.0.1
170
+ urllib3==1.25.11
171
+ uvicorn==0.17.1
172
+ wandb==0.12.9
173
+ wcwidth==0.2.5
174
+ webencodings==0.5.1
175
+ wheel==0.35.1
176
+ widgetsnbextension==3.5.2
177
+ xxhash==2.0.2
178
+ yarl==1.7.2
179
+ yaspin==2.1.0
180
+ zipp==3.7.0
wandb/run-20220130_174749-kzuygp3o/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.15.0-151-generic-x86_64-with-glibc2.10",
3
+ "python": "3.8.8",
4
+ "heartbeatAt": "2022-01-30T17:47:50.968083",
5
+ "startedAt": "2022-01-30T17:47:49.159655",
6
+ "docker": null,
7
+ "gpu": "Tesla V100S-PCIE-32GB",
8
+ "gpu_count": 1,
9
+ "cpu_count": 60,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=mozilla-foundation/common_voice_8_0",
13
+ "--model_name_or_path=facebook/wav2vec2-xls-r-300m",
14
+ "--dataset_config_name=fr",
15
+ "--tokenizer_name_or_path=./",
16
+ "--output_dir=./",
17
+ "--overwrite_output_dir",
18
+ "--num_train_epochs=4",
19
+ "--per_device_train_batch_size=64",
20
+ "--per_device_eval_batch_size=64",
21
+ "--gradient_accumulation_steps=1",
22
+ "--learning_rate=2e-4",
23
+ "--warmup_steps=2500",
24
+ "--length_column_name=input_length",
25
+ "--evaluation_strategy=steps",
26
+ "--text_column_name=sentence",
27
+ "--save_steps=500",
28
+ "--eval_steps=500",
29
+ "--logging_steps=100",
30
+ "--layerdrop=0.0",
31
+ "--activation_dropout=0.05",
32
+ "--save_total_limit=2",
33
+ "--freeze_feature_encoder",
34
+ "--feat_proj_dropout=0.0",
35
+ "--mask_time_prob=0.6",
36
+ "--mask_time_length=10",
37
+ "--mask_feature_prob=0.25",
38
+ "--mask_feature_length=10",
39
+ "--gradient_checkpointing",
40
+ "--report_to=wandb",
41
+ "--run_name=xls-r-300m-fr",
42
+ "--max_train_samples=1000",
43
+ "--max_eval_samples=6000",
44
+ "--max_duration_in_seconds=10",
45
+ "--use_auth_token",
46
+ "--fp16",
47
+ "--group_by_length",
48
+ "--preprocessing_num_workers=64",
49
+ "--do_train",
50
+ "--do_eval",
51
+ "--load_best_model_at_end",
52
+ "--push_to_hub"
53
+ ],
54
+ "state": "running",
55
+ "program": "run_speech_recognition_ctc.py",
56
+ "codePath": "run_speech_recognition_ctc.py",
57
+ "git": {
58
+ "remote": "https://huggingface.co/AlexN/xls-r-300m-fr",
59
+ "commit": "9d07727def2251b0909520344156eb711ff096b9"
60
+ },
61
+ "email": "alexandrenap@hotmail.fr",
62
+ "root": "/workspace/xls-r-300m-fr",
63
+ "host": "job-1abccd0a-3293-4ffe-8274-9e8f841f653f",
64
+ "username": "ovh",
65
+ "executable": "/opt/conda/bin/python"
66
+ }
wandb/run-20220130_174749-kzuygp3o/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 6}}
wandb/run-20220130_174749-kzuygp3o/logs/debug-internal.log ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-01-30 17:47:50,131 INFO MainThread:19599 [internal.py:wandb_internal():87] W&B internal server running at pid: 19599, started at: 2022-01-30 17:47:50.130774
2
+ 2022-01-30 17:47:50,135 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: check_version
3
+ 2022-01-30 17:47:50,135 INFO WriterThread:19599 [datastore.py:open_for_write():77] open: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/run-kzuygp3o.wandb
4
+ 2022-01-30 17:47:50,139 DEBUG SenderThread:19599 [sender.py:send():234] send: header
5
+ 2022-01-30 17:47:50,139 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: check_version
6
+ 2022-01-30 17:47:50,215 DEBUG SenderThread:19599 [sender.py:send():234] send: run
7
+ 2022-01-30 17:47:50,957 INFO SenderThread:19599 [dir_watcher.py:__init__():169] watching files in: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files
8
+ 2022-01-30 17:47:50,957 INFO SenderThread:19599 [sender.py:_start_run_threads():804] run started: kzuygp3o with start time 1643564869
9
+ 2022-01-30 17:47:50,958 DEBUG SenderThread:19599 [sender.py:send():234] send: summary
10
+ 2022-01-30 17:47:50,958 INFO SenderThread:19599 [sender.py:_save_file():939] saving file wandb-summary.json with policy end
11
+ 2022-01-30 17:47:50,959 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: run_start
12
+ 2022-01-30 17:47:50,967 DEBUG HandlerThread:19599 [meta.py:__init__():40] meta init
13
+ 2022-01-30 17:47:50,967 DEBUG HandlerThread:19599 [meta.py:__init__():54] meta init done
14
+ 2022-01-30 17:47:50,968 DEBUG HandlerThread:19599 [meta.py:probe():214] probe
15
+ 2022-01-30 17:47:50,976 DEBUG HandlerThread:19599 [meta.py:_setup_git():204] setup git
16
+ 2022-01-30 17:47:51,010 DEBUG HandlerThread:19599 [meta.py:_setup_git():211] setup git done
17
+ 2022-01-30 17:47:51,010 DEBUG HandlerThread:19599 [meta.py:_save_pip():58] save pip
18
+ 2022-01-30 17:47:51,011 DEBUG HandlerThread:19599 [meta.py:_save_pip():72] save pip done
19
+ 2022-01-30 17:47:51,011 DEBUG HandlerThread:19599 [meta.py:_save_conda():79] save conda
20
+ 2022-01-30 17:47:51,503 DEBUG HandlerThread:19599 [meta.py:_save_conda():89] save conda done
21
+ 2022-01-30 17:47:51,503 DEBUG HandlerThread:19599 [meta.py:probe():252] probe done
22
+ 2022-01-30 17:47:51,511 DEBUG SenderThread:19599 [sender.py:send():234] send: files
23
+ 2022-01-30 17:47:51,511 INFO SenderThread:19599 [sender.py:_save_file():939] saving file wandb-metadata.json with policy now
24
+ 2022-01-30 17:47:51,523 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: stop_status
25
+ 2022-01-30 17:47:51,524 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: stop_status
26
+ 2022-01-30 17:47:51,704 DEBUG SenderThread:19599 [sender.py:send():234] send: config
27
+ 2022-01-30 17:47:51,706 DEBUG SenderThread:19599 [sender.py:send():234] send: metric
28
+ 2022-01-30 17:47:51,707 DEBUG SenderThread:19599 [sender.py:send():234] send: metric
29
+ 2022-01-30 17:47:51,707 WARNING SenderThread:19599 [sender.py:send_metric():897] Seen metric with glob (shouldnt happen)
30
+ 2022-01-30 17:47:51,961 INFO Thread-8 :19599 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/wandb-metadata.json
31
+ 2022-01-30 17:47:51,961 INFO Thread-8 :19599 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/requirements.txt
32
+ 2022-01-30 17:47:51,961 INFO Thread-8 :19599 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/output.log
33
+ 2022-01-30 17:47:51,962 INFO Thread-8 :19599 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/wandb-summary.json
34
+ 2022-01-30 17:47:51,962 INFO Thread-8 :19599 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/conda-environment.yaml
35
+ 2022-01-30 17:47:52,037 INFO Thread-11 :19599 [upload_job.py:push():137] Uploaded file /tmp/tmpl83m3xpuwandb/2em52v7y-wandb-metadata.json
36
+ 2022-01-30 17:47:53,959 INFO Thread-8 :19599 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/output.log
37
+ 2022-01-30 17:47:57,751 DEBUG SenderThread:19599 [sender.py:send():234] send: telemetry
38
+ 2022-01-30 17:47:57,752 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
39
+ 2022-01-30 17:47:57,752 DEBUG SenderThread:19599 [sender.py:send():234] send: exit
40
+ 2022-01-30 17:47:57,753 INFO SenderThread:19599 [sender.py:send_exit():366] handling exit code: 1
41
+ 2022-01-30 17:47:57,753 INFO SenderThread:19599 [sender.py:send_exit():368] handling runtime: 6
42
+ 2022-01-30 17:47:57,754 INFO SenderThread:19599 [sender.py:_save_file():939] saving file wandb-summary.json with policy end
43
+ 2022-01-30 17:47:57,754 INFO SenderThread:19599 [sender.py:send_exit():374] send defer
44
+ 2022-01-30 17:47:57,754 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
45
+ 2022-01-30 17:47:57,755 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
46
+ 2022-01-30 17:47:57,755 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 0
47
+ 2022-01-30 17:47:57,756 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
48
+ 2022-01-30 17:47:57,756 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 0
49
+ 2022-01-30 17:47:57,756 INFO SenderThread:19599 [sender.py:transition_state():387] send defer: 1
50
+ 2022-01-30 17:47:57,757 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
51
+ 2022-01-30 17:47:57,757 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 1
52
+ 2022-01-30 17:47:57,792 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
53
+ 2022-01-30 17:47:57,793 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 1
54
+ 2022-01-30 17:47:57,793 INFO SenderThread:19599 [sender.py:transition_state():387] send defer: 2
55
+ 2022-01-30 17:47:57,794 DEBUG SenderThread:19599 [sender.py:send():234] send: stats
56
+ 2022-01-30 17:47:57,795 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
57
+ 2022-01-30 17:47:57,795 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 2
58
+ 2022-01-30 17:47:57,795 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
59
+ 2022-01-30 17:47:57,796 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 2
60
+ 2022-01-30 17:47:57,796 INFO SenderThread:19599 [sender.py:transition_state():387] send defer: 3
61
+ 2022-01-30 17:47:57,796 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
62
+ 2022-01-30 17:47:57,797 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 3
63
+ 2022-01-30 17:47:57,797 DEBUG SenderThread:19599 [sender.py:send():234] send: summary
64
+ 2022-01-30 17:47:57,798 INFO SenderThread:19599 [sender.py:_save_file():939] saving file wandb-summary.json with policy end
65
+ 2022-01-30 17:47:57,798 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
66
+ 2022-01-30 17:47:57,798 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 3
67
+ 2022-01-30 17:47:57,798 INFO SenderThread:19599 [sender.py:transition_state():387] send defer: 4
68
+ 2022-01-30 17:47:57,799 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
69
+ 2022-01-30 17:47:57,799 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 4
70
+ 2022-01-30 17:47:57,799 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
71
+ 2022-01-30 17:47:57,799 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 4
72
+ 2022-01-30 17:47:57,857 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
73
+ 2022-01-30 17:47:57,962 INFO Thread-8 :19599 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/wandb-summary.json
74
+ 2022-01-30 17:47:57,962 INFO Thread-8 :19599 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/output.log
75
+ 2022-01-30 17:47:58,077 INFO SenderThread:19599 [sender.py:transition_state():387] send defer: 5
76
+ 2022-01-30 17:47:58,077 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
77
+ 2022-01-30 17:47:58,078 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
78
+ 2022-01-30 17:47:58,079 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 5
79
+ 2022-01-30 17:47:58,079 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
80
+ 2022-01-30 17:47:58,079 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 5
81
+ 2022-01-30 17:47:58,079 INFO SenderThread:19599 [dir_watcher.py:finish():283] shutting down directory watcher
82
+ 2022-01-30 17:47:58,180 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
83
+ 2022-01-30 17:47:58,962 INFO Thread-8 :19599 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/config.yaml
84
+ 2022-01-30 17:47:58,964 INFO SenderThread:19599 [dir_watcher.py:finish():313] scan: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files
85
+ 2022-01-30 17:47:58,965 INFO SenderThread:19599 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/output.log output.log
86
+ 2022-01-30 17:47:58,965 INFO SenderThread:19599 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/wandb-metadata.json wandb-metadata.json
87
+ 2022-01-30 17:47:58,966 INFO SenderThread:19599 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/config.yaml config.yaml
88
+ 2022-01-30 17:47:58,970 INFO SenderThread:19599 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/conda-environment.yaml conda-environment.yaml
89
+ 2022-01-30 17:47:58,977 INFO SenderThread:19599 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/wandb-summary.json wandb-summary.json
90
+ 2022-01-30 17:47:58,978 INFO SenderThread:19599 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/requirements.txt requirements.txt
91
+ 2022-01-30 17:47:58,985 INFO SenderThread:19599 [sender.py:transition_state():387] send defer: 6
92
+ 2022-01-30 17:47:58,986 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
93
+ 2022-01-30 17:47:58,986 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
94
+ 2022-01-30 17:47:58,987 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 6
95
+ 2022-01-30 17:47:58,988 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
96
+ 2022-01-30 17:47:58,993 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 6
97
+ 2022-01-30 17:47:58,994 INFO SenderThread:19599 [file_pusher.py:finish():177] shutting down file pusher
98
+ 2022-01-30 17:47:59,089 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
99
+ 2022-01-30 17:47:59,089 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
100
+ 2022-01-30 17:47:59,192 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
101
+ 2022-01-30 17:47:59,192 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
102
+ 2022-01-30 17:47:59,295 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
103
+ 2022-01-30 17:47:59,296 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
104
+ 2022-01-30 17:47:59,399 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
105
+ 2022-01-30 17:47:59,399 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
106
+ 2022-01-30 17:47:59,490 INFO Thread-14 :19599 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/wandb-summary.json
107
+ 2022-01-30 17:47:59,495 INFO Thread-15 :19599 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/requirements.txt
108
+ 2022-01-30 17:47:59,502 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
109
+ 2022-01-30 17:47:59,502 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
110
+ 2022-01-30 17:47:59,513 INFO Thread-12 :19599 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/output.log
111
+ 2022-01-30 17:47:59,591 INFO Thread-13 :19599 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/files/config.yaml
112
+ 2022-01-30 17:47:59,604 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
113
+ 2022-01-30 17:47:59,605 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
114
+ 2022-01-30 17:47:59,707 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
115
+ 2022-01-30 17:47:59,707 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
116
+ 2022-01-30 17:47:59,793 INFO Thread-7 :19599 [sender.py:transition_state():387] send defer: 7
117
+ 2022-01-30 17:47:59,794 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
118
+ 2022-01-30 17:47:59,794 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 7
119
+ 2022-01-30 17:47:59,794 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
120
+ 2022-01-30 17:47:59,794 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 7
121
+ 2022-01-30 17:47:59,809 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
122
+ 2022-01-30 17:48:00,234 INFO SenderThread:19599 [sender.py:transition_state():387] send defer: 8
123
+ 2022-01-30 17:48:00,235 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
124
+ 2022-01-30 17:48:00,236 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
125
+ 2022-01-30 17:48:00,236 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 8
126
+ 2022-01-30 17:48:00,237 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
127
+ 2022-01-30 17:48:00,237 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 8
128
+ 2022-01-30 17:48:00,237 INFO SenderThread:19599 [sender.py:transition_state():387] send defer: 9
129
+ 2022-01-30 17:48:00,238 DEBUG SenderThread:19599 [sender.py:send():234] send: final
130
+ 2022-01-30 17:48:00,238 DEBUG SenderThread:19599 [sender.py:send():234] send: footer
131
+ 2022-01-30 17:48:00,239 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: defer
132
+ 2022-01-30 17:48:00,239 INFO HandlerThread:19599 [handler.py:handle_request_defer():147] handle defer: 9
133
+ 2022-01-30 17:48:00,239 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: defer
134
+ 2022-01-30 17:48:00,239 INFO SenderThread:19599 [sender.py:send_request_defer():383] handle sender defer: 9
135
+ 2022-01-30 17:48:00,338 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: poll_exit
136
+ 2022-01-30 17:48:00,338 DEBUG SenderThread:19599 [sender.py:send_request():248] send_request: poll_exit
137
+ 2022-01-30 17:48:00,338 INFO SenderThread:19599 [file_pusher.py:join():182] waiting for file pusher
138
+ 2022-01-30 17:48:00,657 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: get_summary
139
+ 2022-01-30 17:48:00,659 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: sampled_history
140
+ 2022-01-30 17:48:00,660 DEBUG HandlerThread:19599 [handler.py:handle_request():130] handle_request: shutdown
141
+ 2022-01-30 17:48:00,660 INFO HandlerThread:19599 [handler.py:finish():731] shutting down handler
142
+ 2022-01-30 17:48:01,238 INFO WriterThread:19599 [datastore.py:close():281] close: /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/run-kzuygp3o.wandb
143
+ 2022-01-30 17:48:01,655 INFO SenderThread:19599 [sender.py:finish():1070] shutting down sender
144
+ 2022-01-30 17:48:01,655 INFO SenderThread:19599 [file_pusher.py:finish():177] shutting down file pusher
145
+ 2022-01-30 17:48:01,655 INFO SenderThread:19599 [file_pusher.py:join():182] waiting for file pusher
146
+ 2022-01-30 17:48:01,659 INFO MainThread:19599 [internal.py:handle_exit():77] Internal process exited
wandb/run-20220130_174749-kzuygp3o/logs/debug.log ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-01-30 17:47:49,163 INFO MainThread:18676 [wandb_setup.py:_flush():71] setting env: {}
2
+ 2022-01-30 17:47:49,164 INFO MainThread:18676 [wandb_setup.py:_flush():71] setting login settings: {}
3
+ 2022-01-30 17:47:49,164 INFO MainThread:18676 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/logs/debug.log
4
+ 2022-01-30 17:47:49,164 INFO MainThread:18676 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220130_174749-kzuygp3o/logs/debug-internal.log
5
+ 2022-01-30 17:47:49,164 INFO MainThread:18676 [wandb_init.py:init():404] calling init triggers
6
+ 2022-01-30 17:47:49,165 INFO MainThread:18676 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
7
+ config: {}
8
+ 2022-01-30 17:47:49,165 INFO MainThread:18676 [wandb_init.py:init():460] starting backend
9
+ 2022-01-30 17:47:49,165 INFO MainThread:18676 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
10
+ 2022-01-30 17:47:49,246 INFO MainThread:18676 [backend.py:ensure_launched():216] starting backend process...
11
+ 2022-01-30 17:47:49,320 INFO MainThread:18676 [backend.py:ensure_launched():221] started backend process with pid: 19599
12
+ 2022-01-30 17:47:49,323 INFO MainThread:18676 [wandb_init.py:init():469] backend started and connected
13
+ 2022-01-30 17:47:49,332 INFO MainThread:18676 [wandb_init.py:init():533] updated telemetry
14
+ 2022-01-30 17:47:49,493 INFO MainThread:18676 [wandb_init.py:init():563] communicating current version
15
+ 2022-01-30 17:47:50,213 INFO MainThread:18676 [wandb_init.py:init():568] got version response
16
+ 2022-01-30 17:47:50,213 INFO MainThread:18676 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
17
+ 2022-01-30 17:47:50,958 INFO MainThread:18676 [wandb_init.py:init():606] starting run threads in backend
18
+ 2022-01-30 17:47:51,522 INFO MainThread:18676 [wandb_run.py:_console_start():1810] atexit reg
19
+ 2022-01-30 17:47:51,523 INFO MainThread:18676 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
20
+ 2022-01-30 17:47:51,524 INFO MainThread:18676 [wandb_run.py:_redirect():1689] Redirecting console.
21
+ 2022-01-30 17:47:51,530 INFO MainThread:18676 [wandb_run.py:_redirect():1745] Redirects installed.
22
+ 2022-01-30 17:47:51,530 INFO MainThread:18676 [wandb_init.py:init():633] run started, returning control to user process
23
+ 2022-01-30 17:47:51,533 INFO MainThread:18676 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 216, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 216, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.6, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 4.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jan30_17-46-00_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
24
+ 2022-01-30 17:47:51,536 INFO MainThread:18676 [wandb_watch.py:watch():43] Watching
25
+ 2022-01-30 17:47:55,376 INFO MainThread:18676 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 1
26
+ 2022-01-30 17:47:55,378 INFO MainThread:18676 [wandb_run.py:_restore():1752] restore
27
+ 2022-01-30 17:47:57,755 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
28
+ wandb_count: 1
29
+ }
30
+ pusher_stats {
31
+ uploaded_bytes: 2216
32
+ total_bytes: 2216
33
+ }
34
+
35
+ 2022-01-30 17:47:58,079 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
36
+ wandb_count: 1
37
+ }
38
+ pusher_stats {
39
+ uploaded_bytes: 2216
40
+ total_bytes: 2216
41
+ }
42
+
43
+ 2022-01-30 17:47:58,987 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
44
+ wandb_count: 5
45
+ }
46
+ pusher_stats {
47
+ uploaded_bytes: 2216
48
+ total_bytes: 16101
49
+ }
50
+
51
+ 2022-01-30 17:47:59,090 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
52
+ wandb_count: 5
53
+ }
54
+ pusher_stats {
55
+ uploaded_bytes: 2216
56
+ total_bytes: 16101
57
+ }
58
+
59
+ 2022-01-30 17:47:59,194 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
60
+ wandb_count: 5
61
+ }
62
+ pusher_stats {
63
+ uploaded_bytes: 3544
64
+ total_bytes: 16101
65
+ }
66
+
67
+ 2022-01-30 17:47:59,297 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
68
+ wandb_count: 5
69
+ }
70
+ pusher_stats {
71
+ uploaded_bytes: 16101
72
+ total_bytes: 16101
73
+ }
74
+
75
+ 2022-01-30 17:47:59,400 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
76
+ wandb_count: 5
77
+ }
78
+ pusher_stats {
79
+ uploaded_bytes: 16101
80
+ total_bytes: 16101
81
+ }
82
+
83
+ 2022-01-30 17:47:59,503 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
84
+ wandb_count: 5
85
+ }
86
+ pusher_stats {
87
+ uploaded_bytes: 16101
88
+ total_bytes: 16101
89
+ }
90
+
91
+ 2022-01-30 17:47:59,605 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
92
+ wandb_count: 5
93
+ }
94
+ pusher_stats {
95
+ uploaded_bytes: 16101
96
+ total_bytes: 16101
97
+ }
98
+
99
+ 2022-01-30 17:47:59,708 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
100
+ wandb_count: 5
101
+ }
102
+ pusher_stats {
103
+ uploaded_bytes: 16101
104
+ total_bytes: 16101
105
+ }
106
+
107
+ 2022-01-30 17:48:00,236 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
108
+ wandb_count: 5
109
+ }
110
+ pusher_stats {
111
+ uploaded_bytes: 16101
112
+ total_bytes: 16101
113
+ }
114
+
115
+ 2022-01-30 17:48:00,656 INFO MainThread:18676 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
116
+ exit_result {
117
+ }
118
+ file_counts {
119
+ wandb_count: 5
120
+ }
121
+ pusher_stats {
122
+ uploaded_bytes: 16101
123
+ total_bytes: 16101
124
+ }
125
+ local_info {
126
+ }
127
+
128
+ 2022-01-30 17:48:01,823 INFO MainThread:18676 [wandb_run.py:_append_files():2180] logging synced files
wandb/run-20220130_174749-kzuygp3o/run-kzuygp3o.wandb ADDED
Binary file (7.92 kB). View file
 
wandb/run-20220130_175135-mz3dbiz4/files/conda-environment.yaml ADDED
File without changes
wandb/run-20220130_175135-mz3dbiz4/files/config.yaml ADDED
@@ -0,0 +1,654 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: facebook/wav2vec2-xls-r-300m
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.9
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ python_version: 3.8.8
22
+ start_time: 1643565096
23
+ t:
24
+ 1:
25
+ - 1
26
+ - 5
27
+ - 11
28
+ 2:
29
+ - 1
30
+ - 5
31
+ - 11
32
+ 3:
33
+ - 1
34
+ - 7
35
+ - 13
36
+ 4: 3.8.8
37
+ 5: 0.12.9
38
+ 6: 4.17.0.dev0
39
+ 8:
40
+ - 5
41
+ activation_dropout:
42
+ desc: null
43
+ value: 0.05
44
+ adafactor:
45
+ desc: null
46
+ value: false
47
+ adam_beta1:
48
+ desc: null
49
+ value: 0.9
50
+ adam_beta2:
51
+ desc: null
52
+ value: 0.999
53
+ adam_epsilon:
54
+ desc: null
55
+ value: 1.0e-08
56
+ adapter_kernel_size:
57
+ desc: null
58
+ value: 3
59
+ adapter_stride:
60
+ desc: null
61
+ value: 2
62
+ add_adapter:
63
+ desc: null
64
+ value: false
65
+ add_cross_attention:
66
+ desc: null
67
+ value: false
68
+ apply_spec_augment:
69
+ desc: null
70
+ value: true
71
+ architectures:
72
+ desc: null
73
+ value:
74
+ - Wav2Vec2ForPreTraining
75
+ attention_dropout:
76
+ desc: null
77
+ value: 0.0
78
+ bad_words_ids:
79
+ desc: null
80
+ value: null
81
+ bf16:
82
+ desc: null
83
+ value: false
84
+ bf16_full_eval:
85
+ desc: null
86
+ value: false
87
+ bos_token_id:
88
+ desc: null
89
+ value: 1
90
+ chunk_size_feed_forward:
91
+ desc: null
92
+ value: 0
93
+ classifier_proj_size:
94
+ desc: null
95
+ value: 256
96
+ codevector_dim:
97
+ desc: null
98
+ value: 768
99
+ contrastive_logits_temperature:
100
+ desc: null
101
+ value: 0.1
102
+ conv_bias:
103
+ desc: null
104
+ value: true
105
+ conv_dim:
106
+ desc: null
107
+ value:
108
+ - 512
109
+ - 512
110
+ - 512
111
+ - 512
112
+ - 512
113
+ - 512
114
+ - 512
115
+ conv_kernel:
116
+ desc: null
117
+ value:
118
+ - 10
119
+ - 3
120
+ - 3
121
+ - 3
122
+ - 3
123
+ - 2
124
+ - 2
125
+ conv_stride:
126
+ desc: null
127
+ value:
128
+ - 5
129
+ - 2
130
+ - 2
131
+ - 2
132
+ - 2
133
+ - 2
134
+ - 2
135
+ cross_attention_hidden_size:
136
+ desc: null
137
+ value: null
138
+ ctc_loss_reduction:
139
+ desc: null
140
+ value: mean
141
+ ctc_zero_infinity:
142
+ desc: null
143
+ value: false
144
+ dataloader_drop_last:
145
+ desc: null
146
+ value: false
147
+ dataloader_num_workers:
148
+ desc: null
149
+ value: 0
150
+ dataloader_pin_memory:
151
+ desc: null
152
+ value: true
153
+ ddp_bucket_cap_mb:
154
+ desc: null
155
+ value: None
156
+ ddp_find_unused_parameters:
157
+ desc: null
158
+ value: None
159
+ debug:
160
+ desc: null
161
+ value: '[]'
162
+ decoder_start_token_id:
163
+ desc: null
164
+ value: null
165
+ deepspeed:
166
+ desc: null
167
+ value: None
168
+ disable_tqdm:
169
+ desc: null
170
+ value: false
171
+ diversity_loss_weight:
172
+ desc: null
173
+ value: 0.1
174
+ diversity_penalty:
175
+ desc: null
176
+ value: 0.0
177
+ do_eval:
178
+ desc: null
179
+ value: true
180
+ do_predict:
181
+ desc: null
182
+ value: false
183
+ do_sample:
184
+ desc: null
185
+ value: false
186
+ do_stable_layer_norm:
187
+ desc: null
188
+ value: true
189
+ do_train:
190
+ desc: null
191
+ value: true
192
+ early_stopping:
193
+ desc: null
194
+ value: false
195
+ encoder_no_repeat_ngram_size:
196
+ desc: null
197
+ value: 0
198
+ eos_token_id:
199
+ desc: null
200
+ value: 2
201
+ eval_accumulation_steps:
202
+ desc: null
203
+ value: None
204
+ eval_batch_size:
205
+ desc: null
206
+ value: 64
207
+ eval_steps:
208
+ desc: null
209
+ value: 500
210
+ evaluation_strategy:
211
+ desc: null
212
+ value: steps
213
+ feat_extract_activation:
214
+ desc: null
215
+ value: gelu
216
+ feat_extract_dropout:
217
+ desc: null
218
+ value: 0.0
219
+ feat_extract_norm:
220
+ desc: null
221
+ value: layer
222
+ feat_proj_dropout:
223
+ desc: null
224
+ value: 0.0
225
+ feat_quantizer_dropout:
226
+ desc: null
227
+ value: 0.0
228
+ final_dropout:
229
+ desc: null
230
+ value: 0.0
231
+ finetuning_task:
232
+ desc: null
233
+ value: null
234
+ forced_bos_token_id:
235
+ desc: null
236
+ value: null
237
+ forced_eos_token_id:
238
+ desc: null
239
+ value: null
240
+ fp16:
241
+ desc: null
242
+ value: true
243
+ fp16_backend:
244
+ desc: null
245
+ value: auto
246
+ fp16_full_eval:
247
+ desc: null
248
+ value: false
249
+ fp16_opt_level:
250
+ desc: null
251
+ value: O1
252
+ gradient_accumulation_steps:
253
+ desc: null
254
+ value: 1
255
+ gradient_checkpointing:
256
+ desc: null
257
+ value: true
258
+ greater_is_better:
259
+ desc: null
260
+ value: false
261
+ group_by_length:
262
+ desc: null
263
+ value: true
264
+ half_precision_backend:
265
+ desc: null
266
+ value: amp
267
+ hidden_act:
268
+ desc: null
269
+ value: gelu
270
+ hidden_dropout:
271
+ desc: null
272
+ value: 0.0
273
+ hidden_size:
274
+ desc: null
275
+ value: 1024
276
+ hub_model_id:
277
+ desc: null
278
+ value: None
279
+ hub_strategy:
280
+ desc: null
281
+ value: every_save
282
+ hub_token:
283
+ desc: null
284
+ value: <HUB_TOKEN>
285
+ id2label:
286
+ desc: null
287
+ value:
288
+ '0': LABEL_0
289
+ '1': LABEL_1
290
+ ignore_data_skip:
291
+ desc: null
292
+ value: false
293
+ initializer_range:
294
+ desc: null
295
+ value: 0.02
296
+ intermediate_size:
297
+ desc: null
298
+ value: 4096
299
+ is_decoder:
300
+ desc: null
301
+ value: false
302
+ is_encoder_decoder:
303
+ desc: null
304
+ value: false
305
+ label2id:
306
+ desc: null
307
+ value:
308
+ LABEL_0: 0
309
+ LABEL_1: 1
310
+ label_names:
311
+ desc: null
312
+ value: None
313
+ label_smoothing_factor:
314
+ desc: null
315
+ value: 0.0
316
+ layer_norm_eps:
317
+ desc: null
318
+ value: 1.0e-05
319
+ layerdrop:
320
+ desc: null
321
+ value: 0.0
322
+ learning_rate:
323
+ desc: null
324
+ value: 0.0002
325
+ length_column_name:
326
+ desc: null
327
+ value: input_length
328
+ length_penalty:
329
+ desc: null
330
+ value: 1.0
331
+ load_best_model_at_end:
332
+ desc: null
333
+ value: true
334
+ local_rank:
335
+ desc: null
336
+ value: -1
337
+ log_level:
338
+ desc: null
339
+ value: -1
340
+ log_level_replica:
341
+ desc: null
342
+ value: -1
343
+ log_on_each_node:
344
+ desc: null
345
+ value: true
346
+ logging_dir:
347
+ desc: null
348
+ value: ./runs/Jan30_17-50-54_job-1abccd0a-3293-4ffe-8274-9e8f841f653f
349
+ logging_first_step:
350
+ desc: null
351
+ value: false
352
+ logging_nan_inf_filter:
353
+ desc: null
354
+ value: true
355
+ logging_steps:
356
+ desc: null
357
+ value: 100
358
+ logging_strategy:
359
+ desc: null
360
+ value: steps
361
+ lr_scheduler_type:
362
+ desc: null
363
+ value: linear
364
+ mask_feature_length:
365
+ desc: null
366
+ value: 10
367
+ mask_feature_min_masks:
368
+ desc: null
369
+ value: 0
370
+ mask_feature_prob:
371
+ desc: null
372
+ value: 0.25
373
+ mask_time_length:
374
+ desc: null
375
+ value: 10
376
+ mask_time_min_masks:
377
+ desc: null
378
+ value: 2
379
+ mask_time_prob:
380
+ desc: null
381
+ value: 0.6
382
+ max_grad_norm:
383
+ desc: null
384
+ value: 1.0
385
+ max_length:
386
+ desc: null
387
+ value: 20
388
+ max_steps:
389
+ desc: null
390
+ value: -1
391
+ metric_for_best_model:
392
+ desc: null
393
+ value: loss
394
+ min_length:
395
+ desc: null
396
+ value: 0
397
+ model_type:
398
+ desc: null
399
+ value: wav2vec2
400
+ mp_parameters:
401
+ desc: null
402
+ value: ''
403
+ no_cuda:
404
+ desc: null
405
+ value: false
406
+ no_repeat_ngram_size:
407
+ desc: null
408
+ value: 0
409
+ num_adapter_layers:
410
+ desc: null
411
+ value: 3
412
+ num_attention_heads:
413
+ desc: null
414
+ value: 16
415
+ num_beam_groups:
416
+ desc: null
417
+ value: 1
418
+ num_beams:
419
+ desc: null
420
+ value: 1
421
+ num_codevector_groups:
422
+ desc: null
423
+ value: 2
424
+ num_codevectors_per_group:
425
+ desc: null
426
+ value: 320
427
+ num_conv_pos_embedding_groups:
428
+ desc: null
429
+ value: 16
430
+ num_conv_pos_embeddings:
431
+ desc: null
432
+ value: 128
433
+ num_feat_extract_layers:
434
+ desc: null
435
+ value: 7
436
+ num_hidden_layers:
437
+ desc: null
438
+ value: 24
439
+ num_negatives:
440
+ desc: null
441
+ value: 100
442
+ num_return_sequences:
443
+ desc: null
444
+ value: 1
445
+ num_train_epochs:
446
+ desc: null
447
+ value: 4.0
448
+ optim:
449
+ desc: null
450
+ value: adamw_hf
451
+ output_attentions:
452
+ desc: null
453
+ value: false
454
+ output_dir:
455
+ desc: null
456
+ value: ./
457
+ output_hidden_size:
458
+ desc: null
459
+ value: 1024
460
+ output_hidden_states:
461
+ desc: null
462
+ value: false
463
+ output_scores:
464
+ desc: null
465
+ value: false
466
+ overwrite_output_dir:
467
+ desc: null
468
+ value: true
469
+ pad_token_id:
470
+ desc: null
471
+ value: 218
472
+ past_index:
473
+ desc: null
474
+ value: -1
475
+ per_device_eval_batch_size:
476
+ desc: null
477
+ value: 64
478
+ per_device_train_batch_size:
479
+ desc: null
480
+ value: 64
481
+ per_gpu_eval_batch_size:
482
+ desc: null
483
+ value: None
484
+ per_gpu_train_batch_size:
485
+ desc: null
486
+ value: None
487
+ prediction_loss_only:
488
+ desc: null
489
+ value: false
490
+ prefix:
491
+ desc: null
492
+ value: null
493
+ problem_type:
494
+ desc: null
495
+ value: null
496
+ proj_codevector_dim:
497
+ desc: null
498
+ value: 768
499
+ pruned_heads:
500
+ desc: null
501
+ value: {}
502
+ push_to_hub:
503
+ desc: null
504
+ value: true
505
+ push_to_hub_model_id:
506
+ desc: null
507
+ value: None
508
+ push_to_hub_organization:
509
+ desc: null
510
+ value: None
511
+ push_to_hub_token:
512
+ desc: null
513
+ value: <PUSH_TO_HUB_TOKEN>
514
+ remove_invalid_values:
515
+ desc: null
516
+ value: false
517
+ remove_unused_columns:
518
+ desc: null
519
+ value: true
520
+ repetition_penalty:
521
+ desc: null
522
+ value: 1.0
523
+ report_to:
524
+ desc: null
525
+ value: '[''wandb'']'
526
+ resume_from_checkpoint:
527
+ desc: null
528
+ value: None
529
+ return_dict:
530
+ desc: null
531
+ value: true
532
+ return_dict_in_generate:
533
+ desc: null
534
+ value: false
535
+ run_name:
536
+ desc: null
537
+ value: xls-r-300m-fr
538
+ save_on_each_node:
539
+ desc: null
540
+ value: false
541
+ save_steps:
542
+ desc: null
543
+ value: 500
544
+ save_strategy:
545
+ desc: null
546
+ value: steps
547
+ save_total_limit:
548
+ desc: null
549
+ value: 2
550
+ seed:
551
+ desc: null
552
+ value: 42
553
+ sep_token_id:
554
+ desc: null
555
+ value: null
556
+ sharded_ddp:
557
+ desc: null
558
+ value: '[]'
559
+ skip_memory_metrics:
560
+ desc: null
561
+ value: true
562
+ task_specific_params:
563
+ desc: null
564
+ value: null
565
+ tdnn_dilation:
566
+ desc: null
567
+ value:
568
+ - 1
569
+ - 2
570
+ - 3
571
+ - 1
572
+ - 1
573
+ tdnn_dim:
574
+ desc: null
575
+ value:
576
+ - 512
577
+ - 512
578
+ - 512
579
+ - 512
580
+ - 1500
581
+ tdnn_kernel:
582
+ desc: null
583
+ value:
584
+ - 5
585
+ - 3
586
+ - 3
587
+ - 1
588
+ - 1
589
+ temperature:
590
+ desc: null
591
+ value: 1.0
592
+ tf32:
593
+ desc: null
594
+ value: None
595
+ tie_encoder_decoder:
596
+ desc: null
597
+ value: false
598
+ tie_word_embeddings:
599
+ desc: null
600
+ value: true
601
+ tokenizer_class:
602
+ desc: null
603
+ value: null
604
+ top_k:
605
+ desc: null
606
+ value: 50
607
+ top_p:
608
+ desc: null
609
+ value: 1.0
610
+ torch_dtype:
611
+ desc: null
612
+ value: float32
613
+ torchscript:
614
+ desc: null
615
+ value: false
616
+ tpu_metrics_debug:
617
+ desc: null
618
+ value: false
619
+ tpu_num_cores:
620
+ desc: null
621
+ value: None
622
+ train_batch_size:
623
+ desc: null
624
+ value: 64
625
+ transformers_version:
626
+ desc: null
627
+ value: 4.17.0.dev0
628
+ use_bfloat16:
629
+ desc: null
630
+ value: false
631
+ use_legacy_prediction_loop:
632
+ desc: null
633
+ value: false
634
+ use_weighted_layer_sum:
635
+ desc: null
636
+ value: false
637
+ vocab_size:
638
+ desc: null
639
+ value: 219
640
+ warmup_ratio:
641
+ desc: null
642
+ value: 0.0
643
+ warmup_steps:
644
+ desc: null
645
+ value: 2500
646
+ weight_decay:
647
+ desc: null
648
+ value: 0.0
649
+ xpu_backend:
650
+ desc: null
651
+ value: None
652
+ xvector_output_dim:
653
+ desc: null
654
+ value: 512
wandb/run-20220130_175135-mz3dbiz4/files/output.log ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+
4
+
5
+
6
+
7
+
8
+
9
+
10
+
11
+
12
+
13
+
14
+
15
+
16
+
17
+
18
+
19
+
20
+
21
+
22
+
23
+
24
+
25
+
26
+
27
+
28
+
29
+
30
+ 45%|█████████████████████████████████████████████████████████████████████████████▉ | 29/64 [01:44<02:10, 3.72s/it]Traceback (most recent call last):
31
+ File "run_speech_recognition_ctc.py", line 749, in <module>
32
+ main()
33
+ File "run_speech_recognition_ctc.py", line 700, in main
34
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
35
+ File "/opt/conda/lib/python3.8/site-packages/transformers/trainer.py", line 1347, in train
36
+ for step, inputs in enumerate(epoch_iterator):
37
+ File "/opt/conda/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 521, in __next__
38
+ data = self._next_data()
39
+ File "/opt/conda/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 561, in _next_data
40
+ data = self._dataset_fetcher.fetch(index) # may raise StopIteration
41
+ File "/opt/conda/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 52, in fetch
42
+ return self.collate_fn(data)
43
+ File "run_speech_recognition_ctc.py", line 282, in __call__
44
+ batch = self.processor.pad(
45
+ File "/opt/conda/lib/python3.8/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py", line 147, in pad
46
+ return self.current_processor.pad(*args, **kwargs)
47
+ File "/opt/conda/lib/python3.8/site-packages/transformers/feature_extraction_sequence_utils.py", line 217, in pad
48
+ outputs = self._pad(
49
+ File "/opt/conda/lib/python3.8/site-packages/transformers/feature_extraction_sequence_utils.py", line 286, in _pad
50
+ processed_features[self.model_input_names[0]] = np.pad(
51
+ File "<__array_function__ internals>", line 5, in pad
52
+ File "/opt/conda/lib/python3.8/site-packages/numpy/lib/arraypad.py", line 796, in pad
53
+ padded, original_area_slice = _pad_simple(array, pad_width)
54
+ File "/opt/conda/lib/python3.8/site-packages/numpy/lib/arraypad.py", line 124, in _pad_simple
55
+ padded[original_area_slice] = array
56
+ KeyboardInterrupt
wandb/run-20220130_175135-mz3dbiz4/files/requirements.txt ADDED
@@ -0,0 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ aiohttp==3.8.1
2
+ aiosignal==1.2.0
3
+ analytics-python==1.4.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asgiref==3.5.0
9
+ asttokens==2.0.5
10
+ async-timeout==4.0.2
11
+ attrs==21.4.0
12
+ audioread==2.1.9
13
+ backcall==0.2.0
14
+ backoff==1.10.0
15
+ bcrypt==3.2.0
16
+ beautifulsoup4==4.9.3
17
+ black==21.12b0
18
+ bleach==4.1.0
19
+ brotlipy==0.7.0
20
+ certifi==2020.12.5
21
+ cffi==1.14.3
22
+ chardet==3.0.4
23
+ charset-normalizer==2.0.10
24
+ click==8.0.3
25
+ conda-build==3.21.4
26
+ conda-package-handling==1.7.2
27
+ conda==4.9.2
28
+ configparser==5.2.0
29
+ cryptography==3.2.1
30
+ cycler==0.11.0
31
+ datasets==1.18.2.dev0
32
+ debugpy==1.5.1
33
+ decorator==4.4.2
34
+ defusedxml==0.7.1
35
+ dill==0.3.4
36
+ dnspython==2.1.0
37
+ docker-pycreds==0.4.0
38
+ entrypoints==0.3
39
+ executing==0.8.2
40
+ fastapi==0.73.0
41
+ ffmpy==0.3.0
42
+ filelock==3.0.12
43
+ fonttools==4.29.0
44
+ frozenlist==1.3.0
45
+ fsspec==2022.1.0
46
+ gitdb==4.0.9
47
+ gitpython==3.1.26
48
+ glob2==0.7
49
+ gradio==2.7.5.2
50
+ h11==0.13.0
51
+ huggingface-hub==0.4.0
52
+ idna==2.10
53
+ importlib-resources==5.4.0
54
+ ipykernel==6.7.0
55
+ ipython-genutils==0.2.0
56
+ ipython==8.0.1
57
+ ipywidgets==7.6.3
58
+ jedi==0.17.0
59
+ jinja2==2.11.3
60
+ jiwer==2.3.0
61
+ joblib==1.1.0
62
+ json5==0.9.6
63
+ jsonschema==4.4.0
64
+ jupyter-client==7.1.2
65
+ jupyter-core==4.9.1
66
+ jupyterlab-pygments==0.1.2
67
+ jupyterlab-server==1.2.0
68
+ jupyterlab-widgets==1.0.2
69
+ jupyterlab==2.2.9
70
+ kiwisolver==1.3.2
71
+ libarchive-c==2.9
72
+ librosa==0.8.1
73
+ llvmlite==0.38.0
74
+ markdown2==2.4.2
75
+ markupsafe==1.1.1
76
+ matplotlib-inline==0.1.3
77
+ matplotlib==3.5.1
78
+ mistune==0.8.4
79
+ mkl-fft==1.3.0
80
+ mkl-random==1.1.1
81
+ mkl-service==2.3.0
82
+ monotonic==1.6
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nano==0.10.0
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.19.2
94
+ olefile==0.46
95
+ packaging==21.3
96
+ pandas==1.4.0
97
+ pandocfilters==1.5.0
98
+ paramiko==2.9.2
99
+ parso==0.8.1
100
+ pathspec==0.9.0
101
+ pathtools==0.1.2
102
+ pexpect==4.8.0
103
+ pickleshare==0.7.5
104
+ pillow==8.1.2
105
+ pip==21.3.1
106
+ pkginfo==1.7.0
107
+ platformdirs==2.4.1
108
+ pooch==1.6.0
109
+ prometheus-client==0.13.0
110
+ promise==2.3
111
+ prompt-toolkit==3.0.8
112
+ protobuf==3.19.4
113
+ psutil==5.8.0
114
+ ptyprocess==0.7.0
115
+ pure-eval==0.2.2
116
+ pyarrow==6.0.1
117
+ pycosat==0.6.3
118
+ pycparser==2.20
119
+ pycryptodome==3.13.0
120
+ pydantic==1.9.0
121
+ pydub==0.25.1
122
+ pygments==2.8.0
123
+ pynacl==1.5.0
124
+ pyopenssl==19.1.0
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ pysocks==1.7.1
128
+ python-dateutil==2.8.2
129
+ python-etcd==0.4.5
130
+ python-levenshtein==0.12.2
131
+ python-multipart==0.0.5
132
+ pytz==2021.1
133
+ pyyaml==5.4.1
134
+ pyzmq==22.3.0
135
+ regex==2022.1.18
136
+ requests==2.24.0
137
+ resampy==0.2.2
138
+ ruamel-yaml==0.15.87
139
+ sacremoses==0.0.47
140
+ scikit-learn==1.0.2
141
+ scipy==1.7.3
142
+ send2trash==1.8.0
143
+ sentry-sdk==1.5.4
144
+ setuptools==50.3.1.post20201107
145
+ shortuuid==1.0.8
146
+ six==1.15.0
147
+ smmap==5.0.0
148
+ sniffio==1.2.0
149
+ soundfile==0.10.3.post1
150
+ soupsieve==2.2
151
+ stack-data==0.1.4
152
+ starlette==0.17.1
153
+ subprocess32==3.5.4
154
+ termcolor==1.1.0
155
+ terminado==0.13.1
156
+ testpath==0.5.0
157
+ threadpoolctl==3.0.0
158
+ tokenizers==0.11.4
159
+ tomli==1.2.3
160
+ torch==1.10.2
161
+ torchaudio==0.10.2
162
+ torchelastic==0.2.2
163
+ torchtext==0.9.1
164
+ torchvision==0.9.1
165
+ tornado==6.1
166
+ tqdm==4.62.3
167
+ traitlets==5.1.1
168
+ transformers==4.17.0.dev0
169
+ typing-extensions==4.0.1
170
+ urllib3==1.25.11
171
+ uvicorn==0.17.1
172
+ wandb==0.12.9
173
+ wcwidth==0.2.5
174
+ webencodings==0.5.1
175
+ wheel==0.35.1
176
+ widgetsnbextension==3.5.2
177
+ xxhash==2.0.2
178
+ yarl==1.7.2
179
+ yaspin==2.1.0
180
+ zipp==3.7.0
wandb/run-20220130_175135-mz3dbiz4/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.15.0-151-generic-x86_64-with-glibc2.10",
3
+ "python": "3.8.8",
4
+ "heartbeatAt": "2022-01-30T17:51:37.144939",
5
+ "startedAt": "2022-01-30T17:51:35.856275",
6
+ "docker": null,
7
+ "gpu": "Tesla V100S-PCIE-32GB",
8
+ "gpu_count": 1,
9
+ "cpu_count": 60,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=mozilla-foundation/common_voice_8_0",
13
+ "--model_name_or_path=facebook/wav2vec2-xls-r-300m",
14
+ "--dataset_config_name=fr",
15
+ "--tokenizer_name_or_path=./",
16
+ "--output_dir=./",
17
+ "--overwrite_output_dir",
18
+ "--num_train_epochs=4",
19
+ "--per_device_train_batch_size=64",
20
+ "--per_device_eval_batch_size=64",
21
+ "--gradient_accumulation_steps=1",
22
+ "--learning_rate=2e-4",
23
+ "--warmup_steps=2500",
24
+ "--length_column_name=input_length",
25
+ "--evaluation_strategy=steps",
26
+ "--text_column_name=sentence",
27
+ "--save_steps=500",
28
+ "--eval_steps=500",
29
+ "--logging_steps=100",
30
+ "--layerdrop=0.0",
31
+ "--activation_dropout=0.05",
32
+ "--save_total_limit=2",
33
+ "--freeze_feature_encoder",
34
+ "--feat_proj_dropout=0.0",
35
+ "--mask_time_prob=0.6",
36
+ "--mask_time_length=10",
37
+ "--mask_feature_prob=0.25",
38
+ "--mask_feature_length=10",
39
+ "--gradient_checkpointing",
40
+ "--report_to=wandb",
41
+ "--run_name=xls-r-300m-fr",
42
+ "--max_train_samples=1000",
43
+ "--max_eval_samples=60",
44
+ "--max_duration_in_seconds=10",
45
+ "--use_auth_token",
46
+ "--fp16",
47
+ "--group_by_length",
48
+ "--preprocessing_num_workers=64",
49
+ "--do_train",
50
+ "--do_eval",
51
+ "--load_best_model_at_end",
52
+ "--push_to_hub"
53
+ ],
54
+ "state": "running",
55
+ "program": "run_speech_recognition_ctc.py",
56
+ "codePath": "run_speech_recognition_ctc.py",
57
+ "git": {
58
+ "remote": "https://huggingface.co/AlexN/xls-r-300m-fr",
59
+ "commit": "9d07727def2251b0909520344156eb711ff096b9"
60
+ },
61
+ "email": "alexandrenap@hotmail.fr",
62
+ "root": "/workspace/xls-r-300m-fr",
63
+ "host": "job-1abccd0a-3293-4ffe-8274-9e8f841f653f",
64
+ "username": "ovh",
65
+ "executable": "/opt/conda/bin/python"
66
+ }
wandb/run-20220130_175135-mz3dbiz4/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_wandb": {"runtime": 109}}
wandb/run-20220130_175135-mz3dbiz4/logs/debug-internal.log ADDED
@@ -0,0 +1,192 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-01-30 17:51:36,839 INFO MainThread:21293 [internal.py:wandb_internal():87] W&B internal server running at pid: 21293, started at: 2022-01-30 17:51:36.839410
2
+ 2022-01-30 17:51:36,843 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: check_version
3
+ 2022-01-30 17:51:36,844 INFO WriterThread:21293 [datastore.py:open_for_write():77] open: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/run-mz3dbiz4.wandb
4
+ 2022-01-30 17:51:36,847 DEBUG SenderThread:21293 [sender.py:send():234] send: header
5
+ 2022-01-30 17:51:36,848 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: check_version
6
+ 2022-01-30 17:51:36,917 DEBUG SenderThread:21293 [sender.py:send():234] send: run
7
+ 2022-01-30 17:51:37,133 INFO SenderThread:21293 [dir_watcher.py:__init__():169] watching files in: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files
8
+ 2022-01-30 17:51:37,134 INFO SenderThread:21293 [sender.py:_start_run_threads():804] run started: mz3dbiz4 with start time 1643565096
9
+ 2022-01-30 17:51:37,134 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: run_start
10
+ 2022-01-30 17:51:37,134 DEBUG SenderThread:21293 [sender.py:send():234] send: summary
11
+ 2022-01-30 17:51:37,136 INFO SenderThread:21293 [sender.py:_save_file():939] saving file wandb-summary.json with policy end
12
+ 2022-01-30 17:51:37,144 DEBUG HandlerThread:21293 [meta.py:__init__():40] meta init
13
+ 2022-01-30 17:51:37,144 DEBUG HandlerThread:21293 [meta.py:__init__():54] meta init done
14
+ 2022-01-30 17:51:37,144 DEBUG HandlerThread:21293 [meta.py:probe():214] probe
15
+ 2022-01-30 17:51:37,154 DEBUG HandlerThread:21293 [meta.py:_setup_git():204] setup git
16
+ 2022-01-30 17:51:37,190 DEBUG HandlerThread:21293 [meta.py:_setup_git():211] setup git done
17
+ 2022-01-30 17:51:37,191 DEBUG HandlerThread:21293 [meta.py:_save_pip():58] save pip
18
+ 2022-01-30 17:51:37,193 DEBUG HandlerThread:21293 [meta.py:_save_pip():72] save pip done
19
+ 2022-01-30 17:51:37,193 DEBUG HandlerThread:21293 [meta.py:_save_conda():79] save conda
20
+ 2022-01-30 17:51:37,717 DEBUG HandlerThread:21293 [meta.py:_save_conda():89] save conda done
21
+ 2022-01-30 17:51:37,718 DEBUG HandlerThread:21293 [meta.py:probe():252] probe done
22
+ 2022-01-30 17:51:37,726 DEBUG SenderThread:21293 [sender.py:send():234] send: files
23
+ 2022-01-30 17:51:37,726 INFO SenderThread:21293 [sender.py:_save_file():939] saving file wandb-metadata.json with policy now
24
+ 2022-01-30 17:51:37,736 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: stop_status
25
+ 2022-01-30 17:51:37,737 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: stop_status
26
+ 2022-01-30 17:51:37,890 DEBUG SenderThread:21293 [sender.py:send():234] send: config
27
+ 2022-01-30 17:51:37,892 DEBUG SenderThread:21293 [sender.py:send():234] send: metric
28
+ 2022-01-30 17:51:37,893 DEBUG SenderThread:21293 [sender.py:send():234] send: metric
29
+ 2022-01-30 17:51:37,893 WARNING SenderThread:21293 [sender.py:send_metric():897] Seen metric with glob (shouldnt happen)
30
+ 2022-01-30 17:51:38,136 INFO Thread-8 :21293 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/conda-environment.yaml
31
+ 2022-01-30 17:51:38,137 INFO Thread-8 :21293 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/wandb-summary.json
32
+ 2022-01-30 17:51:38,137 INFO Thread-8 :21293 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/requirements.txt
33
+ 2022-01-30 17:51:38,138 INFO Thread-8 :21293 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
34
+ 2022-01-30 17:51:38,138 INFO Thread-8 :21293 [dir_watcher.py:_on_file_created():217] file/dir created: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/wandb-metadata.json
35
+ 2022-01-30 17:51:38,250 INFO Thread-11 :21293 [upload_job.py:push():137] Uploaded file /tmp/tmp7m8cqty2wandb/246jybel-wandb-metadata.json
36
+ 2022-01-30 17:51:40,135 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
37
+ 2022-01-30 17:51:46,138 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
38
+ 2022-01-30 17:51:48,140 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
39
+ 2022-01-30 17:51:50,142 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
40
+ 2022-01-30 17:51:53,043 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: stop_status
41
+ 2022-01-30 17:51:53,043 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: stop_status
42
+ 2022-01-30 17:51:56,145 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
43
+ 2022-01-30 17:52:00,148 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
44
+ 2022-01-30 17:52:02,149 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
45
+ 2022-01-30 17:52:05,410 DEBUG SenderThread:21293 [sender.py:send():234] send: stats
46
+ 2022-01-30 17:52:06,151 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
47
+ 2022-01-30 17:52:08,153 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/config.yaml
48
+ 2022-01-30 17:52:08,539 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: stop_status
49
+ 2022-01-30 17:52:08,540 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: stop_status
50
+ 2022-01-30 17:52:10,154 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
51
+ 2022-01-30 17:52:12,156 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
52
+ 2022-01-30 17:52:16,158 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
53
+ 2022-01-30 17:52:21,162 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
54
+ 2022-01-30 17:52:23,163 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
55
+ 2022-01-30 17:52:23,731 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: stop_status
56
+ 2022-01-30 17:52:23,732 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: stop_status
57
+ 2022-01-30 17:52:29,168 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
58
+ 2022-01-30 17:52:31,169 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
59
+ 2022-01-30 17:52:33,170 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
60
+ 2022-01-30 17:52:35,634 DEBUG SenderThread:21293 [sender.py:send():234] send: stats
61
+ 2022-01-30 17:52:37,174 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
62
+ 2022-01-30 17:52:38,896 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: stop_status
63
+ 2022-01-30 17:52:38,896 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: stop_status
64
+ 2022-01-30 17:52:43,178 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
65
+ 2022-01-30 17:52:45,179 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
66
+ 2022-01-30 17:52:47,180 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
67
+ 2022-01-30 17:52:53,183 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
68
+ 2022-01-30 17:52:54,057 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: stop_status
69
+ 2022-01-30 17:52:54,058 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: stop_status
70
+ 2022-01-30 17:52:57,186 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
71
+ 2022-01-30 17:52:59,187 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
72
+ 2022-01-30 17:53:03,190 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
73
+ 2022-01-30 17:53:05,866 DEBUG SenderThread:21293 [sender.py:send():234] send: stats
74
+ 2022-01-30 17:53:07,192 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
75
+ 2022-01-30 17:53:09,194 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
76
+ 2022-01-30 17:53:09,222 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: stop_status
77
+ 2022-01-30 17:53:09,223 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: stop_status
78
+ 2022-01-30 17:53:15,197 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
79
+ 2022-01-30 17:53:17,198 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
80
+ 2022-01-30 17:53:19,200 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
81
+ 2022-01-30 17:53:23,992 WARNING MainThread:21293 [internal.py:wandb_internal():146] Internal process interrupt: 1
82
+ 2022-01-30 17:53:25,205 INFO Thread-8 :21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
83
+ 2022-01-30 17:53:26,380 DEBUG SenderThread:21293 [sender.py:send():234] send: telemetry
84
+ 2022-01-30 17:53:26,380 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
85
+ 2022-01-30 17:53:26,380 DEBUG SenderThread:21293 [sender.py:send():234] send: exit
86
+ 2022-01-30 17:53:26,381 INFO SenderThread:21293 [sender.py:send_exit():366] handling exit code: 255
87
+ 2022-01-30 17:53:26,381 INFO SenderThread:21293 [sender.py:send_exit():368] handling runtime: 109
88
+ 2022-01-30 17:53:26,381 INFO SenderThread:21293 [sender.py:_save_file():939] saving file wandb-summary.json with policy end
89
+ 2022-01-30 17:53:26,382 INFO SenderThread:21293 [sender.py:send_exit():374] send defer
90
+ 2022-01-30 17:53:26,382 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
91
+ 2022-01-30 17:53:26,383 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
92
+ 2022-01-30 17:53:26,383 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 0
93
+ 2022-01-30 17:53:26,384 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
94
+ 2022-01-30 17:53:26,384 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 0
95
+ 2022-01-30 17:53:26,384 INFO SenderThread:21293 [sender.py:transition_state():387] send defer: 1
96
+ 2022-01-30 17:53:26,384 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
97
+ 2022-01-30 17:53:26,385 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 1
98
+ 2022-01-30 17:53:26,424 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
99
+ 2022-01-30 17:53:26,424 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 1
100
+ 2022-01-30 17:53:26,424 INFO SenderThread:21293 [sender.py:transition_state():387] send defer: 2
101
+ 2022-01-30 17:53:26,425 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
102
+ 2022-01-30 17:53:26,425 DEBUG SenderThread:21293 [sender.py:send():234] send: stats
103
+ 2022-01-30 17:53:26,426 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 2
104
+ 2022-01-30 17:53:26,427 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
105
+ 2022-01-30 17:53:26,427 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 2
106
+ 2022-01-30 17:53:26,427 INFO SenderThread:21293 [sender.py:transition_state():387] send defer: 3
107
+ 2022-01-30 17:53:26,427 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
108
+ 2022-01-30 17:53:26,428 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 3
109
+ 2022-01-30 17:53:26,428 DEBUG SenderThread:21293 [sender.py:send():234] send: summary
110
+ 2022-01-30 17:53:26,429 INFO SenderThread:21293 [sender.py:_save_file():939] saving file wandb-summary.json with policy end
111
+ 2022-01-30 17:53:26,429 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
112
+ 2022-01-30 17:53:26,429 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 3
113
+ 2022-01-30 17:53:26,430 INFO SenderThread:21293 [sender.py:transition_state():387] send defer: 4
114
+ 2022-01-30 17:53:26,430 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
115
+ 2022-01-30 17:53:26,430 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 4
116
+ 2022-01-30 17:53:26,431 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
117
+ 2022-01-30 17:53:26,431 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 4
118
+ 2022-01-30 17:53:26,485 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
119
+ 2022-01-30 17:53:26,670 INFO SenderThread:21293 [sender.py:transition_state():387] send defer: 5
120
+ 2022-01-30 17:53:26,670 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
121
+ 2022-01-30 17:53:26,670 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
122
+ 2022-01-30 17:53:26,671 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 5
123
+ 2022-01-30 17:53:26,671 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
124
+ 2022-01-30 17:53:26,671 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 5
125
+ 2022-01-30 17:53:26,671 INFO SenderThread:21293 [dir_watcher.py:finish():283] shutting down directory watcher
126
+ 2022-01-30 17:53:26,772 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
127
+ 2022-01-30 17:53:27,207 INFO SenderThread:21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/config.yaml
128
+ 2022-01-30 17:53:27,207 INFO SenderThread:21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
129
+ 2022-01-30 17:53:27,208 INFO SenderThread:21293 [dir_watcher.py:_on_file_modified():230] file/dir modified: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/wandb-summary.json
130
+ 2022-01-30 17:53:27,208 INFO SenderThread:21293 [dir_watcher.py:finish():313] scan: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files
131
+ 2022-01-30 17:53:27,209 INFO SenderThread:21293 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log output.log
132
+ 2022-01-30 17:53:27,209 INFO SenderThread:21293 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/wandb-metadata.json wandb-metadata.json
133
+ 2022-01-30 17:53:27,209 INFO SenderThread:21293 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/config.yaml config.yaml
134
+ 2022-01-30 17:53:27,210 INFO SenderThread:21293 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/conda-environment.yaml conda-environment.yaml
135
+ 2022-01-30 17:53:27,222 INFO SenderThread:21293 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/wandb-summary.json wandb-summary.json
136
+ 2022-01-30 17:53:27,223 INFO SenderThread:21293 [dir_watcher.py:finish():327] scan save: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/requirements.txt requirements.txt
137
+ 2022-01-30 17:53:27,224 INFO SenderThread:21293 [sender.py:transition_state():387] send defer: 6
138
+ 2022-01-30 17:53:27,224 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
139
+ 2022-01-30 17:53:27,233 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
140
+ 2022-01-30 17:53:27,233 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 6
141
+ 2022-01-30 17:53:27,241 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
142
+ 2022-01-30 17:53:27,242 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 6
143
+ 2022-01-30 17:53:27,242 INFO SenderThread:21293 [file_pusher.py:finish():177] shutting down file pusher
144
+ 2022-01-30 17:53:27,329 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
145
+ 2022-01-30 17:53:27,329 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
146
+ 2022-01-30 17:53:27,436 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
147
+ 2022-01-30 17:53:27,440 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
148
+ 2022-01-30 17:53:27,543 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
149
+ 2022-01-30 17:53:27,543 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
150
+ 2022-01-30 17:53:27,646 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
151
+ 2022-01-30 17:53:27,646 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
152
+ 2022-01-30 17:53:27,738 INFO Thread-12 :21293 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/output.log
153
+ 2022-01-30 17:53:27,746 INFO Thread-14 :21293 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/wandb-summary.json
154
+ 2022-01-30 17:53:27,748 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
155
+ 2022-01-30 17:53:27,748 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
156
+ 2022-01-30 17:53:27,761 INFO Thread-13 :21293 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/config.yaml
157
+ 2022-01-30 17:53:27,775 INFO Thread-15 :21293 [upload_job.py:push():137] Uploaded file /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/files/requirements.txt
158
+ 2022-01-30 17:53:27,851 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
159
+ 2022-01-30 17:53:27,851 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
160
+ 2022-01-30 17:53:27,954 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
161
+ 2022-01-30 17:53:27,954 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
162
+ 2022-01-30 17:53:27,976 INFO Thread-7 :21293 [sender.py:transition_state():387] send defer: 7
163
+ 2022-01-30 17:53:27,977 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
164
+ 2022-01-30 17:53:27,977 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 7
165
+ 2022-01-30 17:53:27,977 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
166
+ 2022-01-30 17:53:27,977 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 7
167
+ 2022-01-30 17:53:28,057 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
168
+ 2022-01-30 17:53:28,973 INFO SenderThread:21293 [sender.py:transition_state():387] send defer: 8
169
+ 2022-01-30 17:53:28,974 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
170
+ 2022-01-30 17:53:28,974 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
171
+ 2022-01-30 17:53:28,975 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 8
172
+ 2022-01-30 17:53:28,975 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
173
+ 2022-01-30 17:53:28,975 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 8
174
+ 2022-01-30 17:53:28,975 INFO SenderThread:21293 [sender.py:transition_state():387] send defer: 9
175
+ 2022-01-30 17:53:28,976 DEBUG SenderThread:21293 [sender.py:send():234] send: final
176
+ 2022-01-30 17:53:28,977 DEBUG SenderThread:21293 [sender.py:send():234] send: footer
177
+ 2022-01-30 17:53:28,977 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: defer
178
+ 2022-01-30 17:53:28,977 INFO HandlerThread:21293 [handler.py:handle_request_defer():147] handle defer: 9
179
+ 2022-01-30 17:53:28,977 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: defer
180
+ 2022-01-30 17:53:28,977 INFO SenderThread:21293 [sender.py:send_request_defer():383] handle sender defer: 9
181
+ 2022-01-30 17:53:29,077 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: poll_exit
182
+ 2022-01-30 17:53:29,077 DEBUG SenderThread:21293 [sender.py:send_request():248] send_request: poll_exit
183
+ 2022-01-30 17:53:29,078 INFO SenderThread:21293 [file_pusher.py:join():182] waiting for file pusher
184
+ 2022-01-30 17:53:29,426 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: get_summary
185
+ 2022-01-30 17:53:29,427 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: sampled_history
186
+ 2022-01-30 17:53:29,429 DEBUG HandlerThread:21293 [handler.py:handle_request():130] handle_request: shutdown
187
+ 2022-01-30 17:53:29,429 INFO HandlerThread:21293 [handler.py:finish():731] shutting down handler
188
+ 2022-01-30 17:53:29,977 INFO WriterThread:21293 [datastore.py:close():281] close: /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/run-mz3dbiz4.wandb
189
+ 2022-01-30 17:53:30,424 INFO SenderThread:21293 [sender.py:finish():1070] shutting down sender
190
+ 2022-01-30 17:53:30,425 INFO SenderThread:21293 [file_pusher.py:finish():177] shutting down file pusher
191
+ 2022-01-30 17:53:30,425 INFO SenderThread:21293 [file_pusher.py:join():182] waiting for file pusher
192
+ 2022-01-30 17:53:30,428 INFO MainThread:21293 [internal.py:handle_exit():77] Internal process exited
wandb/run-20220130_175135-mz3dbiz4/logs/debug.log ADDED
@@ -0,0 +1,128 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2022-01-30 17:51:35,860 INFO MainThread:20808 [wandb_setup.py:_flush():71] setting env: {}
2
+ 2022-01-30 17:51:35,860 INFO MainThread:20808 [wandb_setup.py:_flush():71] setting login settings: {}
3
+ 2022-01-30 17:51:35,861 INFO MainThread:20808 [wandb_init.py:_log_setup():371] Logging user logs to /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/logs/debug.log
4
+ 2022-01-30 17:51:35,861 INFO MainThread:20808 [wandb_init.py:_log_setup():372] Logging internal logs to /workspace/xls-r-300m-fr/wandb/run-20220130_175135-mz3dbiz4/logs/debug-internal.log
5
+ 2022-01-30 17:51:35,861 INFO MainThread:20808 [wandb_init.py:init():404] calling init triggers
6
+ 2022-01-30 17:51:35,861 INFO MainThread:20808 [wandb_init.py:init():409] wandb.init called with sweep_config: {}
7
+ config: {}
8
+ 2022-01-30 17:51:35,861 INFO MainThread:20808 [wandb_init.py:init():460] starting backend
9
+ 2022-01-30 17:51:35,862 INFO MainThread:20808 [backend.py:_multiprocessing_setup():99] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
10
+ 2022-01-30 17:51:35,951 INFO MainThread:20808 [backend.py:ensure_launched():216] starting backend process...
11
+ 2022-01-30 17:51:36,027 INFO MainThread:20808 [backend.py:ensure_launched():221] started backend process with pid: 21293
12
+ 2022-01-30 17:51:36,030 INFO MainThread:20808 [wandb_init.py:init():469] backend started and connected
13
+ 2022-01-30 17:51:36,039 INFO MainThread:20808 [wandb_init.py:init():533] updated telemetry
14
+ 2022-01-30 17:51:36,196 INFO MainThread:20808 [wandb_init.py:init():563] communicating current version
15
+ 2022-01-30 17:51:36,915 INFO MainThread:20808 [wandb_init.py:init():568] got version response
16
+ 2022-01-30 17:51:36,915 INFO MainThread:20808 [wandb_init.py:init():578] communicating run to backend with 30 second timeout
17
+ 2022-01-30 17:51:37,132 INFO MainThread:20808 [wandb_init.py:init():606] starting run threads in backend
18
+ 2022-01-30 17:51:37,730 INFO MainThread:20808 [wandb_run.py:_console_start():1810] atexit reg
19
+ 2022-01-30 17:51:37,731 INFO MainThread:20808 [wandb_run.py:_redirect():1684] redirect: SettingsConsole.REDIRECT
20
+ 2022-01-30 17:51:37,732 INFO MainThread:20808 [wandb_run.py:_redirect():1689] Redirecting console.
21
+ 2022-01-30 17:51:37,736 INFO MainThread:20808 [wandb_run.py:_redirect():1745] Redirects installed.
22
+ 2022-01-30 17:51:37,736 INFO MainThread:20808 [wandb_init.py:init():633] run started, returning control to user process
23
+ 2022-01-30 17:51:37,737 INFO MainThread:20808 [wandb_run.py:_config_callback():956] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 218, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-xls-r-300m', 'transformers_version': '4.17.0.dev0', 'feat_extract_dropout': 0.0, 'model_type': 'wav2vec2', 'num_feat_extract_layers': 7, 'hidden_size': 1024, 'feat_extract_norm': 'layer', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': True, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 24, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'num_attention_heads': 16, 'hidden_dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.05, 'feat_proj_dropout': 0.0, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 219, 'do_stable_layer_norm': True, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.6, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.25, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 768, 'proj_codevector_dim': 768, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'mean', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 1024, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 64, 'per_gpu_train_batch_size': 'None', 'per_gpu_eval_batch_size': 'None', 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 'None', 'learning_rate': 0.0002, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 4.0, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'warmup_ratio': 0.0, 'warmup_steps': 2500, 'log_level': -1, 'log_level_replica': -1, 'log_on_each_node': True, 'logging_dir': './runs/Jan30_17-50-54_job-1abccd0a-3293-4ffe-8274-9e8f841f653f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 100, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_on_each_node': False, 'no_cuda': False, 'seed': 42, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'amp', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': 'None', 'local_rank': -1, 'xpu_backend': 'None', 'tpu_num_cores': 'None', 'tpu_metrics_debug': False, 'debug': '[]', 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'xls-r-300m-fr', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': 'None', 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'sharded_ddp': '[]', 'deepspeed': 'None', 'label_smoothing_factor': 0.0, 'optim': 'adamw_hf', 'adafactor': False, 'group_by_length': True, 'length_column_name': 'input_length', 'report_to': "['wandb']", 'ddp_find_unused_parameters': 'None', 'ddp_bucket_cap_mb': 'None', 'dataloader_pin_memory': True, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': 'None', 'hub_model_id': 'None', 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'gradient_checkpointing': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': 'None', 'push_to_hub_organization': 'None', 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', '_n_gpu': 1, 'mp_parameters': '', 'train_batch_size': 64, 'eval_batch_size': 64}
24
+ 2022-01-30 17:51:37,741 INFO MainThread:20808 [wandb_watch.py:watch():43] Watching
25
+ 2022-01-30 17:53:23,994 INFO MainThread:20808 [wandb_run.py:_atexit_cleanup():1780] got exitcode: 255
26
+ 2022-01-30 17:53:23,996 INFO MainThread:20808 [wandb_run.py:_restore():1752] restore
27
+ 2022-01-30 17:53:26,384 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
28
+ wandb_count: 1
29
+ }
30
+ pusher_stats {
31
+ uploaded_bytes: 2214
32
+ total_bytes: 2214
33
+ }
34
+
35
+ 2022-01-30 17:53:26,671 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
36
+ wandb_count: 1
37
+ }
38
+ pusher_stats {
39
+ uploaded_bytes: 2214
40
+ total_bytes: 2214
41
+ }
42
+
43
+ 2022-01-30 17:53:27,227 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
44
+ wandb_count: 3
45
+ }
46
+ pusher_stats {
47
+ uploaded_bytes: 2214
48
+ total_bytes: 13771
49
+ }
50
+
51
+ 2022-01-30 17:53:27,331 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
52
+ wandb_count: 5
53
+ }
54
+ pusher_stats {
55
+ uploaded_bytes: 2214
56
+ total_bytes: 16932
57
+ }
58
+
59
+ 2022-01-30 17:53:27,442 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
60
+ wandb_count: 5
61
+ }
62
+ pusher_stats {
63
+ uploaded_bytes: 2214
64
+ total_bytes: 16932
65
+ }
66
+
67
+ 2022-01-30 17:53:27,544 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
68
+ wandb_count: 5
69
+ }
70
+ pusher_stats {
71
+ uploaded_bytes: 16932
72
+ total_bytes: 16932
73
+ }
74
+
75
+ 2022-01-30 17:53:27,647 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
76
+ wandb_count: 5
77
+ }
78
+ pusher_stats {
79
+ uploaded_bytes: 16932
80
+ total_bytes: 16932
81
+ }
82
+
83
+ 2022-01-30 17:53:27,749 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
84
+ wandb_count: 5
85
+ }
86
+ pusher_stats {
87
+ uploaded_bytes: 16932
88
+ total_bytes: 16932
89
+ }
90
+
91
+ 2022-01-30 17:53:27,852 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
92
+ wandb_count: 5
93
+ }
94
+ pusher_stats {
95
+ uploaded_bytes: 16932
96
+ total_bytes: 16932
97
+ }
98
+
99
+ 2022-01-30 17:53:27,955 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
100
+ wandb_count: 5
101
+ }
102
+ pusher_stats {
103
+ uploaded_bytes: 16932
104
+ total_bytes: 16932
105
+ }
106
+
107
+ 2022-01-30 17:53:28,975 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: file_counts {
108
+ wandb_count: 5
109
+ }
110
+ pusher_stats {
111
+ uploaded_bytes: 16932
112
+ total_bytes: 16932
113
+ }
114
+
115
+ 2022-01-30 17:53:29,424 INFO MainThread:20808 [wandb_run.py:_wait_for_finish():1912] got exit ret: done: true
116
+ exit_result {
117
+ }
118
+ file_counts {
119
+ wandb_count: 5
120
+ }
121
+ pusher_stats {
122
+ uploaded_bytes: 16932
123
+ total_bytes: 16932
124
+ }
125
+ local_info {
126
+ }
127
+
128
+ 2022-01-30 17:53:30,573 INFO MainThread:20808 [wandb_run.py:_append_files():2180] logging synced files
wandb/run-20220130_175135-mz3dbiz4/run-mz3dbiz4.wandb ADDED
Binary file (18.8 kB). View file
 
wandb/run-20220130_175417-3n44neww/files/conda-environment.yaml ADDED
File without changes
wandb/run-20220130_175417-3n44neww/files/config.yaml ADDED
@@ -0,0 +1,686 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _n_gpu:
4
+ desc: null
5
+ value: 1
6
+ _name_or_path:
7
+ desc: null
8
+ value: facebook/wav2vec2-xls-r-300m
9
+ _wandb:
10
+ desc: null
11
+ value:
12
+ cli_version: 0.12.9
13
+ framework: huggingface
14
+ huggingface_version: 4.17.0.dev0
15
+ is_jupyter_run: false
16
+ is_kaggle_kernel: false
17
+ m:
18
+ - 1: train/global_step
19
+ 6:
20
+ - 3
21
+ - 1: train/loss
22
+ 5: 1
23
+ 6:
24
+ - 1
25
+ - 1: train/learning_rate
26
+ 5: 1
27
+ 6:
28
+ - 1
29
+ - 1: train/epoch
30
+ 5: 1
31
+ 6:
32
+ - 1
33
+ - 1: eval/loss
34
+ 5: 1
35
+ 6:
36
+ - 1
37
+ - 1: eval/wer
38
+ 5: 1
39
+ 6:
40
+ - 1
41
+ - 1: eval/runtime
42
+ 5: 1
43
+ 6:
44
+ - 1
45
+ - 1: eval/samples_per_second
46
+ 5: 1
47
+ 6:
48
+ - 1
49
+ - 1: eval/steps_per_second
50
+ 5: 1
51
+ 6:
52
+ - 1
53
+ python_version: 3.8.8
54
+ start_time: 1643565257
55
+ t:
56
+ 1:
57
+ - 1
58
+ - 5
59
+ - 11
60
+ 2:
61
+ - 1
62
+ - 5
63
+ - 11
64
+ 3:
65
+ - 1
66
+ - 7
67
+ - 13
68
+ 4: 3.8.8
69
+ 5: 0.12.9
70
+ 6: 4.17.0.dev0
71
+ 8:
72
+ - 5
73
+ activation_dropout:
74
+ desc: null
75
+ value: 0.05
76
+ adafactor:
77
+ desc: null
78
+ value: false
79
+ adam_beta1:
80
+ desc: null
81
+ value: 0.9
82
+ adam_beta2:
83
+ desc: null
84
+ value: 0.999
85
+ adam_epsilon:
86
+ desc: null
87
+ value: 1.0e-08
88
+ adapter_kernel_size:
89
+ desc: null
90
+ value: 3
91
+ adapter_stride:
92
+ desc: null
93
+ value: 2
94
+ add_adapter:
95
+ desc: null
96
+ value: false
97
+ add_cross_attention:
98
+ desc: null
99
+ value: false
100
+ apply_spec_augment:
101
+ desc: null
102
+ value: true
103
+ architectures:
104
+ desc: null
105
+ value:
106
+ - Wav2Vec2ForPreTraining
107
+ attention_dropout:
108
+ desc: null
109
+ value: 0.0
110
+ bad_words_ids:
111
+ desc: null
112
+ value: null
113
+ bf16:
114
+ desc: null
115
+ value: false
116
+ bf16_full_eval:
117
+ desc: null
118
+ value: false
119
+ bos_token_id:
120
+ desc: null
121
+ value: 1
122
+ chunk_size_feed_forward:
123
+ desc: null
124
+ value: 0
125
+ classifier_proj_size:
126
+ desc: null
127
+ value: 256
128
+ codevector_dim:
129
+ desc: null
130
+ value: 768
131
+ contrastive_logits_temperature:
132
+ desc: null
133
+ value: 0.1
134
+ conv_bias:
135
+ desc: null
136
+ value: true
137
+ conv_dim:
138
+ desc: null
139
+ value:
140
+ - 512
141
+ - 512
142
+ - 512
143
+ - 512
144
+ - 512
145
+ - 512
146
+ - 512
147
+ conv_kernel:
148
+ desc: null
149
+ value:
150
+ - 10
151
+ - 3
152
+ - 3
153
+ - 3
154
+ - 3
155
+ - 2
156
+ - 2
157
+ conv_stride:
158
+ desc: null
159
+ value:
160
+ - 5
161
+ - 2
162
+ - 2
163
+ - 2
164
+ - 2
165
+ - 2
166
+ - 2
167
+ cross_attention_hidden_size:
168
+ desc: null
169
+ value: null
170
+ ctc_loss_reduction:
171
+ desc: null
172
+ value: mean
173
+ ctc_zero_infinity:
174
+ desc: null
175
+ value: false
176
+ dataloader_drop_last:
177
+ desc: null
178
+ value: false
179
+ dataloader_num_workers:
180
+ desc: null
181
+ value: 0
182
+ dataloader_pin_memory:
183
+ desc: null
184
+ value: true
185
+ ddp_bucket_cap_mb:
186
+ desc: null
187
+ value: None
188
+ ddp_find_unused_parameters:
189
+ desc: null
190
+ value: None
191
+ debug:
192
+ desc: null
193
+ value: '[]'
194
+ decoder_start_token_id:
195
+ desc: null
196
+ value: null
197
+ deepspeed:
198
+ desc: null
199
+ value: None
200
+ disable_tqdm:
201
+ desc: null
202
+ value: false
203
+ diversity_loss_weight:
204
+ desc: null
205
+ value: 0.1
206
+ diversity_penalty:
207
+ desc: null
208
+ value: 0.0
209
+ do_eval:
210
+ desc: null
211
+ value: true
212
+ do_predict:
213
+ desc: null
214
+ value: false
215
+ do_sample:
216
+ desc: null
217
+ value: false
218
+ do_stable_layer_norm:
219
+ desc: null
220
+ value: true
221
+ do_train:
222
+ desc: null
223
+ value: true
224
+ early_stopping:
225
+ desc: null
226
+ value: false
227
+ encoder_no_repeat_ngram_size:
228
+ desc: null
229
+ value: 0
230
+ eos_token_id:
231
+ desc: null
232
+ value: 2
233
+ eval_accumulation_steps:
234
+ desc: null
235
+ value: None
236
+ eval_batch_size:
237
+ desc: null
238
+ value: 64
239
+ eval_steps:
240
+ desc: null
241
+ value: 5
242
+ evaluation_strategy:
243
+ desc: null
244
+ value: steps
245
+ feat_extract_activation:
246
+ desc: null
247
+ value: gelu
248
+ feat_extract_dropout:
249
+ desc: null
250
+ value: 0.0
251
+ feat_extract_norm:
252
+ desc: null
253
+ value: layer
254
+ feat_proj_dropout:
255
+ desc: null
256
+ value: 0.0
257
+ feat_quantizer_dropout:
258
+ desc: null
259
+ value: 0.0
260
+ final_dropout:
261
+ desc: null
262
+ value: 0.0
263
+ finetuning_task:
264
+ desc: null
265
+ value: null
266
+ forced_bos_token_id:
267
+ desc: null
268
+ value: null
269
+ forced_eos_token_id:
270
+ desc: null
271
+ value: null
272
+ fp16:
273
+ desc: null
274
+ value: true
275
+ fp16_backend:
276
+ desc: null
277
+ value: auto
278
+ fp16_full_eval:
279
+ desc: null
280
+ value: false
281
+ fp16_opt_level:
282
+ desc: null
283
+ value: O1
284
+ gradient_accumulation_steps:
285
+ desc: null
286
+ value: 1
287
+ gradient_checkpointing:
288
+ desc: null
289
+ value: true
290
+ greater_is_better:
291
+ desc: null
292
+ value: false
293
+ group_by_length:
294
+ desc: null
295
+ value: true
296
+ half_precision_backend:
297
+ desc: null
298
+ value: amp
299
+ hidden_act:
300
+ desc: null
301
+ value: gelu
302
+ hidden_dropout:
303
+ desc: null
304
+ value: 0.0
305
+ hidden_size:
306
+ desc: null
307
+ value: 1024
308
+ hub_model_id:
309
+ desc: null
310
+ value: None
311
+ hub_strategy:
312
+ desc: null
313
+ value: every_save
314
+ hub_token:
315
+ desc: null
316
+ value: <HUB_TOKEN>
317
+ id2label:
318
+ desc: null
319
+ value:
320
+ '0': LABEL_0
321
+ '1': LABEL_1
322
+ ignore_data_skip:
323
+ desc: null
324
+ value: false
325
+ initializer_range:
326
+ desc: null
327
+ value: 0.02
328
+ intermediate_size:
329
+ desc: null
330
+ value: 4096
331
+ is_decoder:
332
+ desc: null
333
+ value: false
334
+ is_encoder_decoder:
335
+ desc: null
336
+ value: false
337
+ label2id:
338
+ desc: null
339
+ value:
340
+ LABEL_0: 0
341
+ LABEL_1: 1
342
+ label_names:
343
+ desc: null
344
+ value: None
345
+ label_smoothing_factor:
346
+ desc: null
347
+ value: 0.0
348
+ layer_norm_eps:
349
+ desc: null
350
+ value: 1.0e-05
351
+ layerdrop:
352
+ desc: null
353
+ value: 0.0
354
+ learning_rate:
355
+ desc: null
356
+ value: 0.0002
357
+ length_column_name:
358
+ desc: null
359
+ value: input_length
360
+ length_penalty:
361
+ desc: null
362
+ value: 1.0
363
+ load_best_model_at_end:
364
+ desc: null
365
+ value: true
366
+ local_rank:
367
+ desc: null
368
+ value: -1
369
+ log_level:
370
+ desc: null
371
+ value: -1
372
+ log_level_replica:
373
+ desc: null
374
+ value: -1
375
+ log_on_each_node:
376
+ desc: null
377
+ value: true
378
+ logging_dir:
379
+ desc: null
380
+ value: ./runs/Jan30_17-53-36_job-1abccd0a-3293-4ffe-8274-9e8f841f653f
381
+ logging_first_step:
382
+ desc: null
383
+ value: false
384
+ logging_nan_inf_filter:
385
+ desc: null
386
+ value: true
387
+ logging_steps:
388
+ desc: null
389
+ value: 5
390
+ logging_strategy:
391
+ desc: null
392
+ value: steps
393
+ lr_scheduler_type:
394
+ desc: null
395
+ value: linear
396
+ mask_feature_length:
397
+ desc: null
398
+ value: 10
399
+ mask_feature_min_masks:
400
+ desc: null
401
+ value: 0
402
+ mask_feature_prob:
403
+ desc: null
404
+ value: 0.25
405
+ mask_time_length:
406
+ desc: null
407
+ value: 10
408
+ mask_time_min_masks:
409
+ desc: null
410
+ value: 2
411
+ mask_time_prob:
412
+ desc: null
413
+ value: 0.6
414
+ max_grad_norm:
415
+ desc: null
416
+ value: 1.0
417
+ max_length:
418
+ desc: null
419
+ value: 20
420
+ max_steps:
421
+ desc: null
422
+ value: -1
423
+ metric_for_best_model:
424
+ desc: null
425
+ value: loss
426
+ min_length:
427
+ desc: null
428
+ value: 0
429
+ model_type:
430
+ desc: null
431
+ value: wav2vec2
432
+ mp_parameters:
433
+ desc: null
434
+ value: ''
435
+ no_cuda:
436
+ desc: null
437
+ value: false
438
+ no_repeat_ngram_size:
439
+ desc: null
440
+ value: 0
441
+ num_adapter_layers:
442
+ desc: null
443
+ value: 3
444
+ num_attention_heads:
445
+ desc: null
446
+ value: 16
447
+ num_beam_groups:
448
+ desc: null
449
+ value: 1
450
+ num_beams:
451
+ desc: null
452
+ value: 1
453
+ num_codevector_groups:
454
+ desc: null
455
+ value: 2
456
+ num_codevectors_per_group:
457
+ desc: null
458
+ value: 320
459
+ num_conv_pos_embedding_groups:
460
+ desc: null
461
+ value: 16
462
+ num_conv_pos_embeddings:
463
+ desc: null
464
+ value: 128
465
+ num_feat_extract_layers:
466
+ desc: null
467
+ value: 7
468
+ num_hidden_layers:
469
+ desc: null
470
+ value: 24
471
+ num_negatives:
472
+ desc: null
473
+ value: 100
474
+ num_return_sequences:
475
+ desc: null
476
+ value: 1
477
+ num_train_epochs:
478
+ desc: null
479
+ value: 4.0
480
+ optim:
481
+ desc: null
482
+ value: adamw_hf
483
+ output_attentions:
484
+ desc: null
485
+ value: false
486
+ output_dir:
487
+ desc: null
488
+ value: ./
489
+ output_hidden_size:
490
+ desc: null
491
+ value: 1024
492
+ output_hidden_states:
493
+ desc: null
494
+ value: false
495
+ output_scores:
496
+ desc: null
497
+ value: false
498
+ overwrite_output_dir:
499
+ desc: null
500
+ value: true
501
+ pad_token_id:
502
+ desc: null
503
+ value: 218
504
+ past_index:
505
+ desc: null
506
+ value: -1
507
+ per_device_eval_batch_size:
508
+ desc: null
509
+ value: 64
510
+ per_device_train_batch_size:
511
+ desc: null
512
+ value: 64
513
+ per_gpu_eval_batch_size:
514
+ desc: null
515
+ value: None
516
+ per_gpu_train_batch_size:
517
+ desc: null
518
+ value: None
519
+ prediction_loss_only:
520
+ desc: null
521
+ value: false
522
+ prefix:
523
+ desc: null
524
+ value: null
525
+ problem_type:
526
+ desc: null
527
+ value: null
528
+ proj_codevector_dim:
529
+ desc: null
530
+ value: 768
531
+ pruned_heads:
532
+ desc: null
533
+ value: {}
534
+ push_to_hub:
535
+ desc: null
536
+ value: true
537
+ push_to_hub_model_id:
538
+ desc: null
539
+ value: None
540
+ push_to_hub_organization:
541
+ desc: null
542
+ value: None
543
+ push_to_hub_token:
544
+ desc: null
545
+ value: <PUSH_TO_HUB_TOKEN>
546
+ remove_invalid_values:
547
+ desc: null
548
+ value: false
549
+ remove_unused_columns:
550
+ desc: null
551
+ value: true
552
+ repetition_penalty:
553
+ desc: null
554
+ value: 1.0
555
+ report_to:
556
+ desc: null
557
+ value: '[''wandb'']'
558
+ resume_from_checkpoint:
559
+ desc: null
560
+ value: None
561
+ return_dict:
562
+ desc: null
563
+ value: true
564
+ return_dict_in_generate:
565
+ desc: null
566
+ value: false
567
+ run_name:
568
+ desc: null
569
+ value: xls-r-300m-fr
570
+ save_on_each_node:
571
+ desc: null
572
+ value: false
573
+ save_steps:
574
+ desc: null
575
+ value: 500
576
+ save_strategy:
577
+ desc: null
578
+ value: steps
579
+ save_total_limit:
580
+ desc: null
581
+ value: 20
582
+ seed:
583
+ desc: null
584
+ value: 42
585
+ sep_token_id:
586
+ desc: null
587
+ value: null
588
+ sharded_ddp:
589
+ desc: null
590
+ value: '[]'
591
+ skip_memory_metrics:
592
+ desc: null
593
+ value: true
594
+ task_specific_params:
595
+ desc: null
596
+ value: null
597
+ tdnn_dilation:
598
+ desc: null
599
+ value:
600
+ - 1
601
+ - 2
602
+ - 3
603
+ - 1
604
+ - 1
605
+ tdnn_dim:
606
+ desc: null
607
+ value:
608
+ - 512
609
+ - 512
610
+ - 512
611
+ - 512
612
+ - 1500
613
+ tdnn_kernel:
614
+ desc: null
615
+ value:
616
+ - 5
617
+ - 3
618
+ - 3
619
+ - 1
620
+ - 1
621
+ temperature:
622
+ desc: null
623
+ value: 1.0
624
+ tf32:
625
+ desc: null
626
+ value: None
627
+ tie_encoder_decoder:
628
+ desc: null
629
+ value: false
630
+ tie_word_embeddings:
631
+ desc: null
632
+ value: true
633
+ tokenizer_class:
634
+ desc: null
635
+ value: null
636
+ top_k:
637
+ desc: null
638
+ value: 50
639
+ top_p:
640
+ desc: null
641
+ value: 1.0
642
+ torch_dtype:
643
+ desc: null
644
+ value: float32
645
+ torchscript:
646
+ desc: null
647
+ value: false
648
+ tpu_metrics_debug:
649
+ desc: null
650
+ value: false
651
+ tpu_num_cores:
652
+ desc: null
653
+ value: None
654
+ train_batch_size:
655
+ desc: null
656
+ value: 64
657
+ transformers_version:
658
+ desc: null
659
+ value: 4.17.0.dev0
660
+ use_bfloat16:
661
+ desc: null
662
+ value: false
663
+ use_legacy_prediction_loop:
664
+ desc: null
665
+ value: false
666
+ use_weighted_layer_sum:
667
+ desc: null
668
+ value: false
669
+ vocab_size:
670
+ desc: null
671
+ value: 219
672
+ warmup_ratio:
673
+ desc: null
674
+ value: 0.0
675
+ warmup_steps:
676
+ desc: null
677
+ value: 2500
678
+ weight_decay:
679
+ desc: null
680
+ value: 0.0
681
+ xpu_backend:
682
+ desc: null
683
+ value: None
684
+ xvector_output_dim:
685
+ desc: null
686
+ value: 512
wandb/run-20220130_175417-3n44neww/files/output.log ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+
4
+
5
+
6
+ 6%|██████████▊ | 4/64 [00:16<04:04, 4.08s/it]
7
+ 8%|█████████████▌ | 5/64 [00:19<03:39, 3.73s/it]The following columns in the evaluation set don't have a corresponding argument in `Wav2Vec2ForCTC.forward` and have been ignored: input_length.
8
+ ***** Running Evaluation *****
9
+ Num examples = 60
10
+ Batch size = 64
11
+ ['~զ‐զ‐a‐め‐զ‐զ‐めզ‐め‐զ‐め‐めզめ‐めզ‐զ‐զ‐զめզ‐め‐զ‐զ‐զ—զ‐զːめզ‐զめ‐զ‐զ‐ːめզ‐զ‐զ‐め~青βまиi=', '~զːզ—զaզa*զaզaզːզːզ—զːզaզ—զ—ːզaզː—զ—զ=~β~β=ま=まиaզa', '~‐զ‐զːզめզめ‐զːզaզ—զ‐զːզːզ—զaզ—‐զ‐զ—զ‐զ‐զ‐զ‐զ‐զ‐—‐—iиまи', 'ψ‐ː‐զ‐—‐զ‐զ‐—զa—զめզめզ‐զ‐զ‐զ‐զめզ‐զ‐զ‐զめզ—զ‐զめ—զːզめզ—めːզめզ—զːaզaզ‐զːզ‐めզ‐զめզめ‐զめ‐め‐め‐め‐め‐=ᆨiま=ま', 'ψaզ‐զ‐զ‐a‐զːզ‐զ‐զːզ—զ‐զ‐զ‐զ—aզ‐զːaːaզめaզ‐զːզːզ‐զめզ—ː—զ甌‐めզめզ‐*զ‐a‐aզ‐զ‐զ‐a~β~β~E=', 'ψːզːզ‐զ‐զめ—զ‐め‐ːզ‐զ‐զ*զːզ‐զ‐զaզ*զ—ːզ‐զaզ*զaːզːզːզ‐զːզめզ‐զ‐めզːզ‐զめզ‐զ=iᆨβ~βᆨβᆨβᆨβまβまβまβ=β=', 'β‐զめզ‐զ—զ—զ—զ‐զ‐զ—ːզ‐զめզ‐զ—զːզːզ‐զ‐զめ‐զ‐—զ—զめզめզ‐զめ‐めզめ‐զ‐め‐զめզ‐զ‐զ‐めզめզ‐զ—զ‐զめ‐զ‐aiиiиま', '~‐զ‐զ‐զ‐զ‐զ‐a‐զ‐めզめզ‐めզめːզ‐զ‐めզ‐զめզ‐զ‐զːզ‐զめզ‐զ—ːզ‐զ‐զ—զ‐զ‐զめզ‐め‐め‐զ‐a‐aզ‐a‐a‐=β~β', '~զめaиaզ—めզめզめզめզめզめզめզめ—めզ—aզめզめզːaզめ—めզめզめզめզめզめ—めզめզːaզめ—զめզめզめ—ːզめզめզめզめːaզめզめզ—a—զめզめզめզめզめզめզめզめզめզめզめզめզめզ—めզめզめզめaᆨиβ', '~‐ːզめզ—զ‐—զめ‐め‐め‐и‐զ‐め—ːզ‐զ‐զ‐զ‐զːզaզa—զめզ‐զ—զめզめզa‐զ‐զめզ‐めզːզめզめզ‐ːめզ‐զ‐զめզめ—め‐զめːզ‐—めզめզ‐զ‐զːզ‐めզめզ‐զ‐զ‐զ‐めզめ‐め‐զ‐զ‐めզaめ*めզめ‐めզめզ‐=iEiቀまቀま甌ま甌β', '~‐զ‐զめզ‐զめզめզ‐めː—aզ*ːզ‐զ‐զめզ‐めզ‐めզめզː—ːզːaզめզ‐զめզめːզ‐めզめ—զ‐զ‐զaզ*զ‐զ‐め‐めզ‐զ‐զめզː‐զ—ːզ‐զ‐զめզːզ‐զ‐զ‐զ‐զ‐զ‐զːめզめզめզめզ=զめ‐めզめ‐め‐զめզめ‐զ‐めզめզ‐զ‐=i~i~i~', 'ψզ‐a—զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐め‐զ‐զめ‐զ‐め‐*‐զ‐զ‐զ‐զ—ːզ‐զ‐զ—զめզめզ‐զめ‐め‐զめզめզːզめզːaզːめզめզ‐զ‐—‐め‐め‐—զめ‐め‐զ‐め‐զ‐—‐զ‐ː‐զ—զːզ‐զ‐զ‐=iиiиi甌i甌β', '~‐и‐զ‐a‐め‐*զ‐*‐めզ甌զめզ‐*զ‐めզ‐զ‐զ*めզ‐զ*‐甌‐զ‐զ—զ*զ‐զ‐զ‐զ*զ‐զ*զ‐զめզめզめզめզ‐զめզめզ‐めզめզ‐զ‐զ*զめզめզ—‐զ‐զ‐զ‐—զ‐զめզめզ*‐զめզめզ‐*‐զ‐めզ‐զ‐*‐め‐*‐め‐め‐め‐め‐め‐—‐ː‐զ—‐甌‐め‐め‐aiτi', '~‐—‐—‐զ‐զ‐զ‐ː‐めզːզめզめզ—զ‐զ‐զ—զ—ː—զめզめզ‐զːզːզ‐զːզめ‐զ‐զːզ‐զ‐զːめզːզめզ—զːզːզめ‐めզ‐զːめːզめզめ—զめզ‐զ‐զ‐զ‐զ—զːզːզ‐ː—‐զ‐զ‐զ‐aզ‐aiβ~まaまiβ', '~‐ːզめզめզ‐զa*զ—aզːզːզ—զaːaːզめզ—զa—aզːaզ—aːզːզːզ—զːզːզめզ‐զa青β~青まaи=', '~a—めaզ‐aզ‐զ‐զめզ‐a‐զ‐զaզaːզːզ‐զːզ‐めզːզ‐զaզaզa‐զːզ—զ‐զ‐զ*զめզ—զめզめզaզ‐ːզaːaզ‐a=~и', '~=ːaզ‐զ‐զ‐—‐զ‐a‐a‐զ‐զ‐—զめaզaզ‐զ—զaզ—զ‐զː—զ—զめ‐—‐զ*զ—‐զaզː—զ‐զ‐զ‐զ—aզː—զ‐զ‐զː—զ‐զ‐=iβ~βまи=', '~զaզめզ‐զ‐զ‐զ‐զ‐զ—զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐aզaզaզaզ—aզ‐զː—զ—զ‐զ‐զ‐զː—զa—զ‐զ—զ—զːaզ—զめզaզː—զːզː—զ‐զ‐զːզめզ‐զa=β~βቀ=', 'ψː—զめզ—զ*զa‐զ*‐զ‐զːզaզ—aզaզめզめզめզめ—զaզa—զめ‐զ*զ‐զ*զめզaզ‐զ‐զめզaզːզ~яиβиβия', 'ψː—aզめզ‐զː‐aզːaզ—aː—զ—զːզːզめզaզa—ːզ*—aզ—ːզめ‐զ‐aզ‐զ‐a‐め‐զːզめզ‐զ‐ːզːզ‐զ‐զ‐զaզː‐—զ‐զ‐ːզ‐զ‐aզめզ‐զ‐զaզ‐զ‐ː—ːզめզaզ‐զ≥›βEиaи=ቀ', '~զ‐め‐めզ‐զaզめզ*զ—զめզːaːaզ—զめզ—ːaզːզaզaզ—զːզ‐զ—ːզːզaզめզめզめa‐զめզ‐զめ~青~βββ', 'ψզ—‐め‐զ‐め‐զ‐め‐զːզaզ—զ‐զːզめզ‐զ—զ‐め‐զめ‐めզ*—ːզ‐め—զːզː—զ‐զ—զ—め—զːզめ‐め‐զ‐զめ*—aզːզːզ‐զːզː—զめզめːզ—զːզ‐—‐ːզめզ—զ—զ—ːզːզ—զː‐め‐—զめզめ‐զ‐a‐め‐め‐զ‐めզ‐զ‐զ‐≥青~', '~aզaːզa—a—aզめզめզ—aզaզːզːզa—ːaːaզ*—aːaզaզːզaզ—aːզː—ːզ~=β~β~β~β', 'β‐—‐զ‐զ‐զ‐զ‐զ‐զめզ‐զ‐զ‐զ‐զめ‐զ—զaզめ‐զ—a—めզ‐めզ‐զめզaզ‐զ‐զ‐めզ‐զ‐めզ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐aᆨiиiиまi', '~զː—զaզ—a—aզaզːզa—aːaզ—ːզ—զ*—զ—ːaզaզaզːaːaզaː—aզ=~β~β~', '~զ‐զ—ːզ—զめզ‐զaզ—զːaզːզ—զːզ—զː*—զːզ—զaզaզːզaզ—զ‐զːզ‐զa—ː*ːaզ—զ—զːզaːaզ‐զ—զめզːզ‐զ=‐զa=β=~=β=β', '~め‐զめиめզめ‐めզめզめզめaめզ‐զめa—զa—զめզめզلaզめa—զめa—զめaզめզめզめզ—զaզ—めզ—զめզめզ—զめզ—զ—զiи~β', '~զ‐զ‐զめ‐զめ‐—‐զ‐զ‐զ—զ‐զ‐զ—զめզめզ—զ—զa‐զ‐զ‐զ‐զ‐զめզ‐զ—զめզめː—զめ—զːզめզ—զ—զːզ‐զ‐զaզ—ːզ‐զːaզめզaզaզ‐զ‐զめզ‐զ‐զːզ‐զ‐ːզ‐aզ‐զ‐զa=β~β=β', '~ː—‐զ‐զ‐a‐զ‐a‐զ‐զ‐ːզaզ*զaզː—զ—ːaզːaզ‐զ*aզ—ː—aզːզːզːզめզ‐զaiaia', '~—‐զ‐զ‐զ‐զ‐զめ‐զːզ—ː—զ—aզːզaզ‐զ—զめզ—զa‐めզ‐զ—ːզ—ːզめ*めզ‐զめզめզ—զːզ‐զaզめզめզ‐զめզ‐զ—զːզ‐զ—զ‐զ‐զめa青β~β', '~զ‐զ‐զ‐զ‐め‐զaզaզ‐めզ*—=զめզ—զめզ‐զ‐զ—ːզ‐զめզめզ‐զ—զめզめզめզめզ‐զ—զːզːզ厳զ=‐զめ—めզめզめզめզめզːզめːզ‐զ—めːզ‐զ‐զめզ—զめ‐զめզ‐めզめ‐め‐զ‐—ψiまиi', 'ψaːզ—զ—զ—զめզ‐զ—զ—զ—‐զ‐—զ—ːզ‐զ‐զaզ—aզːզːզ—զːaզːզ—զaզ—զ—զːզ—め—զ—զ‐զ‐զ‐զ—զ—զ—զ‐めզ—iまiまя=i=', '~‐めզːզ‐めզ‐զ‐զ‐զ‐め‐զめզめզ‐めզːզ‐զ‐めզ‐め—aզ‐զめզ‐*զ‐զ‐զ‐*‐=‐aめ=զ‐զ‐=青βEiβ', '~զ‐զ‐զ‐զaզ‐զ‐զ‐զaզa—զ—զːզaզ—զaզ*զ—զ*զ—զaզめզ‐զ‐զ‐զː—զːզ=', 'ψզめզ‐զ‐զ‐զ‐զ‐զ‐զめզ‐զめզːզːզ—զめզ*զː—զめզめզ—զ‐զめզめզ*‐め‐զ‐=i~i~a', '~‐զ‐զ‐զ‐め‐զ‐զ‐զ‐զ‐めզめ‐զ‐զ‐զめզめզめ‐զ‐զめզめզ‐զめզ—զめզ‐զめ‐զ‐զめզːめ‐զ‐զめզ‐զ‐զ‐զ‐զめզめզ‐めզ‐զ‐զめզめ‐めզ‐զ‐զ‐զ—‐զめ‐め‐め‐め‐め‐~i', '~‐զ‐め‐զ‐զ‐զ‐めզ‐զːզ‐զめզ‐զ‐զːa‐—զ—զめզ*զ‐զːզめզめզ‐զ‐めզ‐ː‐զːզ‐め‐զめ‐զめ‐=›i›~›i~ま~', '~‐զ‐զ‐め‐め‐զ‐զ‐զ‐めզ‐զめզめզ‐զ‐め‐զ‐զ‐զめզめ‐զ‐զ‐զ—զ‐զaզ‐զ‐զ‐զ‐めզ—‐զ‐զ‐զめզ‐զ‐զめզめզ‐め‐զ‐զ—め—‐զ‐զめզː‐զ‐զめ‐め‐aiβiβiβiβiβまa', '~‐めզ‐զ‐め‐զ‐զ‐զめզめ‐զ‐զ‐զ—զ‐め‐զ‐զ‐զ‐զ‐զめզ‐զ‐զ*め‐զめզめզaզめզaզ‐զa‐զ‐ː‐զ‐զ‐զ‐=β~β=β~', '~ːa—ːめզ‐め‐め‐め‐—‐め‐զaめ—զ‐め‐め‐め‐め‐զ‐めːզ‐զːզ—aզ‐զː*—զaզːզːզ—aː—զ‐զ=яøβ=β=β', '~aզめզ—ː‐め‐a‐め‐め‐め‐め‐め‐め‐め‐め‐—‐զ‐զː—aːզめզめզめզ—aզめզ‐զ—ːզ‐զ‐զ‐զːզめզ—aզめ—զːaː‐զ—զ‐զめզめզ‐め‐め‐ːaめːaզめզ—զːզ‐=iβiᆨβя.иa.яβ', 'ψ‐զ‐զ‐め‐զ‐—զːզ‐զ‐զ‐զ*aզ‐め—めզaːզ‐զ‐զ‐զ‐めզː‐զ‐—զ‐զ‐զ‐զ‐զ‐ːզ‐զ‐զめ—ːզ‐զ‐զ‐զめզ—aզ‐ː‐զː—ːզ‐զːզ‐զː‐զ‐ːզ‐—զ‐め‐=ββ=β', '=զめզめզː‐զ—զめզめզːզːզːaզめa—զ‐զ—لaզ—զ—aզːզː—ːզめaːզ—ːզːզ*զːaզめզ‐զめզ—ːզめ—‐め‐a—ːզ—զめː‐め‐a‐զa~β~β', '~‐զ‐զ‐めզ‐め‐め‐a‐զめզ‐զ‐զ‐զ‐զ—զ—զめզːզ‐զ‐զːզ—ː—ːզめզ—զːզ—ːզ‐զ‐զ—զ‐զ‐զめː—զめզ‐զ—զ‐զ‐զ‐զめզːզːզ��զ‐զ‐め—ː—զ‐զ‐զめզ—ːզ‐զːզː—զː—զめզ‐aβ~まaまяEβ', '~զaզめզːզめզːզ‐զめ—զaզめզ‐զːզめզ—զ‐զ‐զ—զːaզːզːզめզめզːզːզ—զ‐め‐զめa青β~β~β', '~‐զ‐զめզ*զ‐զ‐զ‐めզめզa‐めզ‐めզ‐ː‐զ‐զ‐ːզめզ‐めզ‐զ‐զ‐ːզ‐զめ‐めզ‐զ‐զ‐զめզめզめզめզ‐զ‐զ‐զめզ‐զめզ‐զ‐め‐զ‐め‐ل‐め‐զ‐զ‐=i', '~‐ː‐զ‐a‐ːզ‐զめ—զめզめ‐զ‐զ‐զ—զ‐զ‐զːզ‐զ—զ‐զ‐զːզ‐զ‐զaզ—զ‐զaզ—զ‐զ‐—ːզ‐զ‐զ‐զ‐a青~=', '~‐զめ‐め‐め‐զ‐զ‐めզ‐զ‐զめ‐զめ‐め‐զめ‐めզ‐զめզ‐զ‐めզめ‐զ‐զ‐զめզめ‐զ‐めզ‐զめ‐զ‐զ‐զ‐めզ‐めզめ‐めզ‐めզめ‐զめզめզめ‐めզ‐め‐զ‐めզめ‐զ‐զ‐զ‐զ‐めզめ‐め‐め‐めզ‐め‐զめզ‐զ‐զ‐զ‐め‐զ‐め‐զ‐め‐—β~βиi', '~*ː—=զめզ*զめզaզ—զめզ—զ*զaզ—զːզめզːめːզめզめզめ*զaːaզめզ*զːզめզめզめզːaզめզめզめզめզめզaզ=i~', 'ψզ—ːզ—զ‐զ‐め‐զ‐զめզ*զ‐զ‐զ‐զ‐めզ‐զ‐aզ‐զ‐զ‐զ—ːզːզːզː—զ—զ—զːզめզː—զːզ‐զ‐≥i~β~яまяя=я=яβ', '~ː—めզめզめ‐めզめզめզめզめզ—a—aզめզaզ*զ*զaզaզ—ː—զaզ—a—aめզめզaː*զめզめ—ːa*aզ—զ—զ—զめզめːզめaめզaᆨяi~яaяa~a~', 'ψ‐—զ—‐ːզ—զめզめզ—ːզ‐զ‐めզ‐զ‐զ—a—a—զ‐զ—ːզ—զめզ‐めզめ=i=まa', '~զː‐զ‐a‐զ‐զ‐զ‐զ‐զめզ‐ːզ‐զ‐զ‐զ‐զːզː—aզːaզːզ—ːզ‐զːa—զ‐զ—զaզ‐զːզ‐զ=β~β', 'βː‐զ=‐զめզめզめզ‐めզめզめ—զめզめզめ—զː—զめզめ‐զ‐めզめզめզめզめզ‐めզ—զめզめaめ‐め—զaめ—զめ—めզめզ‐めզめզめ—ː‐զめ—めզめ—զめզ—զ—զめզめ—զめ—ψiβiβ', '~ːզaːめaզ‐զ‐զ‐め‐め‐め‐め‐a‐զːզめ‐めզ‐զːaզめ—aզː*aզめզめզ—ːզaզめզː—զːզːզ—զ—զ—զめզ‐զ‐ː—ːզめ‐զ‐aiβiEи=иaиa', 'ψːզめ‐ːa—զ—զ‐զ‐զ‐aզ‐—めզ‐զ‐զめːզːզめզːզめːզːզ‐զ—ː—զːզ—զːզːզ—զːզːզaզːզめզ‐զaめզːզ‐め‐զ‐a青βя=β', '~—ːզ‐զ‐զ—めզːզːզ—զaզa—めզaզaզ=β~a=', '~ːզ‐めզ‐a‐—զ‐զ—ːզ‐զ—aզa—զめզ—զaզ—զ‐զ—aːզめզ—զめզめզ—զめզ*—զ—aːզ‐զめ*զ—զːaː‐զめզ‐めզめզ‐めa‐զa~', '~‐զ‐զ‐զ‐զ‐զa‐*զ‐զ‐զ‐めa—զ‐զ‐զː—զめզa—զめզ‐めզめզめզ‐めզːզめզ‐զめզ‐めզː—զːaզːզ‐զ‐ːզːզめզめզ—զめ—‐զ‐զ—øiиiиま', '~զめզ—めզ‐զ‐զː‐զaզ‐զ‐զ‐զめզ—‐զːզ‐զめզa—զaːզaզːզː—զ‐զa—զːaզ‐զめːզ‐զ—զː‐զ‐զ—ːզːզ=βቀ']
12
+ {'eval_loss': 20.994997024536133, 'eval_wer': 1.0, 'eval_runtime': 2.8273, 'eval_samples_per_second': 21.222, 'eval_steps_per_second': 0.354, 'epoch': 0.31}
13
+
14
+
15
+
16
+
17
+ 14%|████████████████████████▎ | 9/64 [00:34<03:10, 3.46s/it]
18
+ 16%|██████████████████████████▉ | 10/64 [00:40<03:34, 3.98s/it]The following columns in the evaluation set don't have a corresponding argument in `Wav2Vec2ForCTC.forward` and have been ignored: input_length.
19
+ ***** Running Evaluation *****
20
+ Num examples = 60
21
+ Batch size = 64
22
+ ['~զ‐զ‐a‐め‐զ‐զ‐めզ‐めզ‐զ‐め‐めզめ‐めզ‐զ‐զ‐զめզ‐め‐զ‐զ‐զ—զ‐զːめզ‐զめ‐զ‐զ‐ːめզめ‐զ‐զ‐め~青βまиi=', '~զːզ—զaզa*զaզaզːզːզ—զaզ—զ—ːզaզː—զ—զ=~β~β=ま=まиaզa', '~‐զ‐զːզめզめ‐զːզaզ—զ‐զːզːզ—զaզ—‐զ‐զ—զ‐զ‐զ‐զ‐զ‐զ‐—‐—iиまи', 'ψ‐ː‐զ‐—‐զ‐め‐զ‐め—զa—զめզめզ‐զ‐զ‐զめզ‐զ‐զ‐զめզ—զ‐զめ—զ—զめզ—めːզめզ—զːaզaզ‐զ‐めզ‐զめզめզめ‐զめ‐め‐め‐め‐め‐=ᆨiま=ま', 'ψaզ‐զ‐զ‐a‐զːզ‐զ‐զːզ—զ‐զ‐զ‐զ—aզ‐めզːaːaզめaզ‐զːզːզ‐զめզ—ː—զ甌‐めզめզ‐*զ‐զ‐aզ‐զ‐զ‐a~β~β~E~E=', 'ψːզːզ‐զ‐զめ—զ‐め‐ːզ‐զ‐զ*զːզ‐զ‐զaզ*զ—ːզ‐զaզ*զaːզːզːզːզ‐զːզめզ‐զ‐めզːզ‐զめզ‐զ=iᆨ~βᆨβᆨβまβ=', 'β‐զめզめզ‐զ—զ—զ—զ‐զ‐զ—ːզ‐զめզ‐զ—զːզːզ‐զ‐զめ‐զ‐—զ—զめ‐զめզ‐զめ‐めզめ‐զ‐め‐め‐զめզ‐զ‐զːզ‐めզめզ‐զ—զ‐զめ‐զ‐aiиiиま', '~‐զ‐զ‐զ‐զ‐a‐զ‐めզめզ‐զめːզ‐զ‐めզ‐զめզ‐զ‐զːզ‐զめզ‐զ—ːզ‐զ‐զ—զ‐զ‐զめզ‐め‐め‐զ‐a‐aզ‐a‐a‐=β~β', '~զめaиaզ—めզめզめզめզめզめզめզめզめ—めզ—ːaզめզめզːaզめ—めզめզめզめզめզ—め—めզめզːaզめ—զめզめզめ—ːզめզめզめզめզめːaզめզ—a—զめզめզめզめզめզめզめզめզめզめզめզ—めզめզめզめaᆨиβ', '~‐ːզめզ—զ‐ːզめ‐め‐め‐и‐զ‐め—ːզ‐զ‐զ‐զ‐զːզaզa—զめզ‐զ‐զ—զめզめզa‐զ‐զめզ‐めզ—զめզめզ‐ːめզ‐զ‐զめզめ—め‐զめːզ‐—めզめզ‐զ‐զːզ‐めզめզ‐զ‐զ‐զ‐めզめ‐め‐զ‐զ‐めզaめ*めզめ‐めզめզ‐=iEiቀまቀま甌β', '~‐զ‐զめզ‐զめզめզ‐めː—aզ*ːզ‐զ‐զめզ‐めզ‐めզめզː—ːզːaզめզめզめզめզめːզ‐めզめ—զ‐զ‐զaզ*զ‐զ‐め‐զ‐զ‐զめզː‐զ—ːզ‐զ‐զめզːզ‐զ‐զ‐զ‐զ‐զ‐զːめզめզめզめզ=զめ‐めզめ‐め‐զめզめ‐զ‐めզめզ‐զ‐=i~i~i~', 'ψզ‐a—զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐め‐զ‐զ‐զめ‐զ‐め‐*‐զ‐զ‐զ‐զ—ːզ‐զ‐զ—զめզめզ‐զめ‐め‐զめզめզːզめզːaզːめզめզ‐զ‐—‐め‐め‐—զめ‐め‐զ‐զ‐め‐め‐զ‐—‐զ‐ː‐զ—զːզ‐զ‐զ‐=iиiиi甌i甌β', '~‐и‐զ‐a‐め‐*զ‐*‐めզ甌զめզ‐*զ‐զ‐զ‐զ*めզ‐զ*‐甌‐զ‐զ—զ*զ‐զ‐զ‐զ*զ‐զ*զ‐զめզめզめզめզ‐զめզめզ‐めզめզ‐զ‐զ*զめզめզ—‐զ‐զ‐զ‐—զ‐զめզめզ*‐զめզ‐զ‐*‐զ‐めզ‐զ‐*‐め‐*‐め‐め‐め‐め‐め‐—‐ː‐զ—‐甌‐め‐め‐aiτi', '~‐—‐—‐զ‐զ‐զ‐ː‐めզːզめզめզ—զ‐զ‐զ—զ—ː—զめզめզ‐զːզːզ‐զːզめ‐զːզ‐զ‐զːめզːզめզ—զːզːզめ‐めզ‐զːめːզめզめ—զめզ‐զ‐զ‐զ‐զ—զːզːզ‐ː—‐զ‐զ‐aզ‐aiβ~まaまиiβ', '~‐ːզめզめզ‐զa*զ—aզːզːզ—զaːaːզめզ—զa—aզːaզ—aːզːզːզ—զːզːզ‐զa青β~青まaи=', '~a—めaզ‐aզ‐զ‐զめզ‐a‐զ‐զaզaːզːզ‐զːզ‐めզːզ‐զaզaզa‐զːզ—զ‐զ‐զ*զ‐զ—զめզめզaզ‐ːզaːaզ‐a=~и', '~=ːaզ‐զ‐զ‐—‐զ‐a‐a‐զ‐զ‐—զaզaզ‐զ—զaզ—զ‐զː—զ—զめ‐—‐զ*զ—‐զaզː—զ‐զ‐զ‐զ—aզː—զ‐զ‐զː—զ‐զ‐=iβ~βまи=и=', '~զaզめզ‐զ‐զ‐զ‐զ‐զ—զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐aզaզaզaզ—aզ‐զː—զ—զ‐զ‐զː—զa—զ‐զ—զ—զːaզ—զめզaզː—զːզː—զ‐զ‐զːզめզ‐զa=β~βቀ=', 'ψː—զめզ—զ*զaզ‐զ*‐զ‐զːզaզ—aզaզめզめզめզめ—զaզa—զめ‐զ*զ‐զ*զめզaզ‐զ‐զめզaզːզ~яиβия', 'ψːզ—aզめզ‐զː‐aզːaզ—aː—զ—զːզːզめզaզa—ːզ*—aզ—ːզめ‐զ‐aզ‐զ‐a‐め‐զːզめզ‐զ‐ːզːզ‐զ‐զ‐զaզː‐—զ‐զ‐ːզ‐զ‐aզめզ‐զ‐զaզ‐զ‐ː—ːզめզaզ‐զ≥›βEиaи=ቀ', '~զ‐め‐めզ‐զaզめզ*զ—զめզːaːaզ—զめզ—ːaզːզaզaզ—զːզ‐զ—ːզːզaզめզめզめa‐զめզ‐զめ~青~ββ', 'ψզ—‐め‐զ‐め‐զ‐め‐զːզaզ—զ‐զːզめզ‐զ—զ‐զ‐զめ‐めզ*—ːզ‐め—զːզː—զ‐զ—զ—め—զːզめ‐め‐զ‐զめ*—aզːզːզ‐զːզː—զめːզ—զːզ‐—‐ːզめզ—զ—զ—ːզːզ—զ—զ‐め‐—զめզめ‐զ‐a‐め‐め‐զ‐めզ‐զ‐զ‐≥青~', '~aզaːզa—a—aզめզめզ—aզaզːզːզa—ːaːaզ*—aːaզaզːզaզ—aːզː—ːզ~=β~β~β~β~β', 'β‐—‐զ‐զ‐զ‐զ‐զ‐զ‐զめզ‐զ‐զ‐զ‐զめ‐զ—զaզめ‐զ—a—めզ‐めզ‐զめզめզaզ‐զ‐զ‐めզ‐զ‐めզ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐aᆨiиiиまi', '~զːaզaզ—a—aզaզːզa—aːaզ—ːզ—զ*—զ—ːaզaզaզːaːaզaː—aզ=~β~β~', '~զ‐զ—ːզ—զめզ‐զaզ—զːaզːզ—զːզ—զː*—զːզ—զaզaզːզaզ—զ‐զːզ‐զa—ː*ːaզ—զ—զːզaːaզ‐զ—զめզːզ‐զ=‐զa=β=~=β=β', '~め‐զめиめզめ‐めզめզめզめaめզ‐զめa—զa—զめզめզلaզめa—զめa—զめaզめզめզめզ—զaզ—めզ—զめզめզ—զめզ—զ—զiи~β', '~զ‐զ‐զめ‐զめ‐—‐զ‐զ‐զ—զ‐զ‐զ—զめզめզ—զ—զa‐զ‐զ‐զ‐զ‐զめզ‐զ—զめզめː—զめ—զːզめզ—զ—զːզ‐զ‐զːaզ—ːզ‐զːaզめզaզaզ‐զ‐զめզ‐զ‐զːզ‐զ‐ːզ‐aզ‐զ‐զ‐զa=β~β=β', '~ː—‐զ‐զ‐a‐զ‐զ‐a‐զ‐զ‐ːզaզ*զaզː—զ—aզːaզ‐զ*aզ—ː—aզːզːզːզめզ‐զaiaia', '~—զ‐զ‐զ‐զ‐զめ‐զːզ—ː—զ—aզaզ—զめզ—զa‐めզ‐զ—ːզ—ːզめ*めզ‐զめզめզ—զːզ‐զaզめզめզ‐զめզ‐զ—զːզ‐զ—զ‐զ‐զめa青β~β', '~զ‐զ‐զ‐զ‐め‐զaզaզ‐めզ*—=զめզ—զめզ‐զ‐զ—ːզ‐զめզめզ‐զ—զめզめզめզめզ‐զ—զːզːզ厳զ=‐զめ—めզめզめզめզめզːզめːզ‐զ—めːզ‐զ‐զめզ—զめ‐զめզ‐めզめ‐め‐զ‐—ψiまиi', 'ψaːզ—զ—զ—զめզ‐զ—զ—‐զ‐—զ—զːզ‐զ‐զaզ—aզːզːզ—զːaզːզ—զaզ—զ—զːզ—め—զ—զ‐զ‐զ‐զ—զ—զ—զ‐めզ—iまiま=i=', '~‐めզːզ‐めզ‐զ‐զ‐զ‐め‐զめզめզ‐めզːզ‐զ‐めզ‐め—aզ‐զめզ‐*զ‐զ‐զ‐*‐=‐aめ=զ‐զ‐=青βEiβ', '~զ‐զ‐զ‐զaզ‐զ‐զ‐զa—aզa—զ—զːզaզ—զaզ*զ—զ*զ—զaզめզ‐զ‐զめզ‐զː—զːզ=', 'ψզめզ‐զ‐զ‐զ‐զ‐զ‐զめզ‐զめզːզːզ—զめզ*զː—զめզめզ—զ‐զめզめզ*‐め‐զ‐=i~i~a', '~‐զ‐զ‐զ‐め‐զ‐զ‐զ‐զ‐めզめ‐զ‐զ‐զめզめզめ‐զ‐զめզ‐めզ‐զめզ—զめզ‐զめ‐զ‐զめզːめ‐զ‐զめզ‐զ‐զ‐զめզめզ‐めզ‐զ‐զめզめ‐めզ‐զ‐զ‐զ—‐զめ‐め‐め‐め‐め‐~i', '~‐զ‐め‐զ‐զ‐զ‐めզ‐զːզ‐զめզ‐զ‐զːa‐—զ—զめզ*զ‐զːզめզめզ‐զ‐めզ‐ː‐զːզ‐め‐զめ‐զめ‐=›i›~›i~ま~', '~‐զ‐զ‐め‐め‐զ‐զ‐զ‐めզ‐զめզめզ‐զ‐め‐զ‐զ‐զめզめ‐զ‐զ‐զ‐զ—զ‐զaզ‐զ‐զ‐զ‐めզ—‐զ‐զ‐զめզ‐զ‐զめզめզ‐め‐զ‐զ—め—‐զ‐զめզː‐զ‐զめ‐め‐aiβiβiβiβiβまa', '~‐めզ‐զ‐め‐զ‐զ‐զめզめ‐զ‐զ‐զ—զ‐զ‐զ‐զ‐զ‐զめզ‐զ‐զ*め‐զめզめզaզめզaզ‐զa‐զ‐ː‐զ‐զ‐=β~β=β~', '~ːa—ːめ‐め‐め‐め‐—‐め‐զaめ—զ‐め‐め‐め‐め‐զ‐めːզ‐զːզ—aզ‐զː*—զaզːզːզ—aː—զ‐զ=яøβ=β=β', '~aզめ—զ—ː‐a‐め‐め‐め‐め‐め‐め‐め‐め‐—‐զ‐զː—aːզめզめզめզめզ—aզめզ‐զ—ːզ‐զ‐զ‐զːզめզ—aզめ—զːaː‐զ—զ‐զめզめզ‐め‐め‐ːaめːaզめզ—զːզ‐=iβiᆨβя.иa.яβ', 'ψ‐զ‐զ‐め‐զ‐—զːզ‐զ‐զ‐զ—aզ‐め—めզaːզ‐զ‐զ‐զ‐—զː‐զ‐—զ‐զ‐զめզ‐զ‐զ‐ːզ‐զ—ːզ‐զ‐զ‐զめզ—aզ‐ː‐զめզː—ːզ‐զːզ‐զː‐ːզ‐—զ‐めa‐=ββ=β', '=զめզː‐զ—զめզめզːզːզːaզめː—զ‐զ—لaզ—զ—aզːզː—ːզめaːզ—ːզːզːզめːaզめզ‐զめզ—ːզめ—‐め‐a—ːզ—զめː‐め‐զ‐a‐զa~β~β', '~‐զ‐զ‐めզ‐め‐め‐a‐զめզ‐զ‐զ‐զ‐զ—զ—զめզːզ‐զ‐զːզ—ː—ːզめզ—զːզ—ːզ‐զ‐զ—զ‐զ‐զめː—զめզ‐զ—զ‐զ‐զ‐զめզːզːզ‐զ‐զ‐めզː—զ‐զ‐զめզ—ːզ‐զːզː—զː—զめզ‐a青β~まaまEяEβ', '~զaզめզːզめզːզ‐զめ—aզめզ‐զːզめզ—զ‐զ‐զ—զːaզːզːզめզめզːզːզ—զ‐めզ‐զめa青β~β~β', '~‐զ‐զめզ*զ‐զ‐զ‐めզめզa‐めզ‐めզ‐ː‐զ‐զ‐ːզめզ‐めզ‐զ‐զ‐ːզ‐զめ‐めզめզ‐զ‐զめզめզめզめզ‐զ‐զ‐զめզ‐զめզ‐զ‐め‐զ‐め‐ل‐め‐զ‐զ‐=i', '~‐ː‐զ‐a‐ːզ‐զめ—զめզめ‐զ‐զ‐めզ—զ‐զ‐զːզ‐զ—զ‐զ‐զːզ‐զ‐զaզ—զ‐զaզ—զ‐զ‐—ːզ‐զ‐զ‐զ‐a青~=', '~‐զめ‐め‐め‐զ‐զ‐めզ‐զ‐զめ‐զめ‐め‐զめզ‐めզ‐զめզ‐զ‐めզめ‐զ‐զ‐զめզめ‐զ‐めզ‐զめ‐զ‐զ‐զ‐めզ‐めզめ‐めզ‐めզめ‐զめզめզ‐զめզめ‐めզ‐め‐զ‐めզめ‐զ‐զ‐զ‐զ‐めզめ‐め‐め‐めզ‐め‐զめզ‐զ‐զ‐զ‐め‐զ‐め‐զ‐め‐—β~βиi', '~*ː—=զめզ*զめզaզ—զめզ—զ*զaզ—զːզめզːめːզめզめզめզめ*զaːaզめզ*զːզめզめզめզːaզめզめզめզめզめզaզ=i~', 'ψզ—ːզ—զ‐զ‐め‐զ‐զめզ*զ‐զ‐զ‐զ‐めզ‐զ‐aզ‐զ‐զ‐զ—ːզːզːզːզ—զ—զːզめզː—զːզ‐≥i~β~まяя=я=я=яβ', '~ː—めզめզめզめզめզめզめզ—a—aզめզaզ*զ*զaզaզ—ː—զaզa—aめզめզaː*զめզめ—ːa*aզ—զ—զ—զめզめːզめaめզaᆨяi~яaまa~a~', 'ψ‐—զ—‐ːզ—զめզめզ—ːզ‐զ‐めզめզ‐զ‐զ—a—a—զ‐զ—ːզ—զめզ‐めզめ=i=まa', '~զː‐զ‐a‐զ‐զ‐զ‐զ‐զめզ‐ːզ‐զ‐զ‐զ‐զːզː—aզːaզːզ—ːaզ‐զːa—զ‐զ—զaզ‐զ‐զ=β~β', 'βː‐զ=‐զめզめզめզ‐めզめզめզめզめ—զːզめզめ‐զ‐めզめզめզめզめզ‐めզ—զめզめaめ‐め—զaめ—զめ—めզめզ‐めզめզめ—ː‐զめ—めզめ—զめզ—զ—զめզめ—զめ—ψiβiβ', '~ːզaːめaզ‐զ‐զ‐め‐め‐め‐a‐զːզめ‐めզ‐զːaզめ—aզː*aզめզめզ—ːզaզめզː—զːզːզ—զ—զ—զめզ‐զ‐ː—ːզめ‐զ‐aiβiEиaиa', 'ψːզめ‐ːa—զ‐զ‐զ‐aզ‐—めզ‐զ‐զめːզːզめզːզめːզːզ‐զ—ː—զːզ—զːզːզ—զːզːզaզːզめզ‐զaめզːզ‐め‐զ‐զ‐a青βя=β', '~—ːզ‐զ‐զ—めզːզːզ—զaզa—զaզaզ=β~a=', '~ːզ‐զ‐めզ‐a‐—զ‐զ—ːզ‐զ—aզa—զめզ—զaզ—զ‐զ—aːզめզ—զめզめզ—զめզ*—զ—aːզ‐զめ*զ—զːaː‐զめզ‐めզめզ‐めa‐զa~', '~‐զ‐զ‐զ‐զ‐զa‐*զ‐զ‐զ‐めa—զ‐զ‐զː—զめզa—զめզ‐めզめզめզ‐めզːզめզ‐զめզ‐զː—զːaզːզ‐զ‐ːզːզめզめզ—զめ—‐զ‐զ—øiиiиま', '~զめզ—めզ‐զ‐զː‐զaզ‐զ‐զめզ—‐զːզ‐զめզa—զaզaզːզː—զ‐զa—զːaզ‐զめːզ‐զ—զ‐զ‐զ—ːզːզ=βቀ']
23
+ {'eval_loss': 20.991931915283203, 'eval_wer': 1.0, 'eval_runtime': 2.7507, 'eval_samples_per_second': 21.812, 'eval_steps_per_second': 0.364, 'epoch': 0.62}
24
+
25
+
26
+
27
+
28
+ 22%|█████████████████████████████████████▋ | 14/64 [00:56<03:19, 3.99s/it]
29
+ 23%|████████████████████████████████████████▎ | 15/64 [00:59<02:48, 3.43s/it]The following columns in the evaluation set don't have a corresponding argument in `Wav2Vec2ForCTC.forward` and have been ignored: input_length.
30
+ ***** Running Evaluation *****
31
+ Num examples = 60
32
+ Batch size = 64
33
+
34
+
35
+
36
+
37
+ 30%|███████████████████████████████████████████████████ | 19/64 [01:16<02:46, 3.69s/it]
38
+ ['~զ‐զ‐a‐め‐զ‐զ‐めզ‐め‐զ‐め‐めզめ‐めզ‐զ‐զ‐զめզ‐め‐զ‐զ‐զ—զ‐զːめզ‐զめ‐զ‐զ‐ːめզ‐զ‐զ‐め~青βまиi=', '~զːզ—զaզa*զaզaզːզːզ—զaզ—զ—ːզaզː—զ—զ=~β~β=ま=まиaզa', '~‐զ‐զːզめզめ‐զːզ‐զaզ—զ‐めզːզːզ—զaզ—‐զ‐զ—զ‐զ‐զ‐զ‐զ‐զ‐—‐—iиまи', 'ψ‐ː‐զ‐—‐զ‐め‐զ‐—զa—զめզめզ‐զ‐զ‐զ‐զめզ‐զ‐զ‐զ‐զめզ—զ‐զめ—զ—զめզ—めːզめզ—զːaզaզ‐զːզ‐め‐զめզめզめ‐զめ‐め‐め‐め‐め‐=ᆨiま=ま', 'ψaզ‐զ‐զ‐a‐զːզ‐զ‐զːզ—զ‐զ‐զ‐զ—aզ‐զːaːaզめaզ‐զːզːզ‐զめզ—ː—զ甌‐めզめզ‐*զ‐a‐aզ‐զ‐զ‐a~β~β~E~E=', 'ψːզːզ‐զ‐զめ—զ‐ːզ‐զ‐զ*զːզ‐զ‐զaզ*զ—ːզ‐զaզ*զaːզːզːզːզ‐զːզめզ‐զ‐զːզ‐զめզ‐զ=iᆨβ~βᆨβᆨβ=', 'β‐զめզめզ‐զ—զ—զ—զ‐զ‐զ—ːզ‐զめզ‐զ—զːզːզ‐զ‐զめ‐զ‐—զ—զめզめզ‐զめ‐めզめ‐զ‐め‐め‐զめզ‐զ‐զ‐めզめզ‐զ—զ‐զめ‐զ‐aiиiиまиま', '~‐զ‐զ‐զ‐զ‐a‐զ‐め‐զめզ‐めզめːզ‐զ‐めզ‐զめզ‐զ‐զːզ‐զめզ‐զ—ːզ‐զ‐զ—զ‐զ‐զ‐զめզ‐め‐め‐զ‐a‐aզ‐a‐a‐=β~β', '~զめaиaզめզめզめզめզめզめզめզめզめ—めզ—aզめզめզːaզめ—めզめզめզめզめզ—め—めզめզːaզめ—զめզめզめ—ːզめզめզめզめզめaզめզめզ—a—զめզめզめզめզめզめզめզめզめզめզめզめզめզめզ—めզめզめզめaᆨиβ', '~‐ːզめզ—զ‐ːզめ‐め‐め‐и‐զ‐め—ːզ‐զ‐զ‐զ‐զːզaզa—զめզ‐զ—զめզめզa‐զ‐զめզ‐めզ—զめզ‐զめզ‐ːめզ‐զ‐զめզめզめ—め‐զめːզ‐—めզめզ‐զ‐զːզ‐めզめզ‐զ‐զ‐զ‐զ‐めզめ‐め‐զ‐զ‐զめզaめ*めզめ‐めզめզ‐=iEiቀま甌ま甌β', '~‐զ‐զめզ‐զめզめզ‐めː—aզ*ːզ‐զ‐զめզ‐めզ‐めզめզː—ːզːaզめզ‐զめզめզめːզ‐めզめ—զ‐զ‐զaզ*զ‐զ‐め‐めզ‐զ‐զめզː‐զ—ːզ‐զ‐զめզːզ‐զ‐զめ‐զ‐զ‐զ‐զːめզめզめզめզ=զめ��めզめ‐め‐զめզめ‐զ‐めզめզ‐զ‐=i~i~i~', 'ψզ‐a—զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐め‐զ‐զ‐め‐զめ‐զ‐め‐*‐զ‐զ‐զ‐զ—ːզ‐զ‐զ—զめզめզ‐զめ‐め‐զめզめզːզめզːaզːめզめզ‐զ‐—‐め‐め‐—զめ‐め‐զ‐め‐め‐զ‐—‐ː‐զ—զːզ‐զ‐զ‐=iиiиi甌i甌β', '~‐и‐զ‐a‐め‐*զ‐*‐めզ甌զめզ‐*զ‐զ‐զ‐զ‐*めզ‐զ*‐甌‐զ‐զ—զ*զ‐զ‐զ‐զ*զ‐զ*զ‐զめզめզめզめզ‐զめզめզ‐めզめզ‐զ‐զ*զめզめզ—‐զ‐զ‐զ‐—զ‐զめզめզ*‐զめզめզ‐*‐զ‐めզ‐զ‐*‐め‐*‐め‐め‐め‐め‐め‐—‐ː‐զ—‐*‐め‐め‐aiτi', '~‐—‐—‐զ‐զ‐զ‐ː‐めզːզめզめզ—զ‐զ‐զ—զ—ː—զめզめզ‐զめզːզːզ‐めզːզめ‐զ‐զːզ‐զ‐զːめզːめզめզ—զːզːզめ‐めզ‐զːめːզめզめ—զめզ‐զ‐զ‐զ‐զ—զːզːզ‐զ‐ː—‐զ‐զ‐a‐զ‐aiβ~βまaまиiβ', '~‐ːզめզめզ‐զaզ—aզːզːզ—զaːaːզめզ—զa—aզaːaզ—aːզːզːզ—զːզːզ‐զa青β~青まaи=', '~a—めaզ‐aզ‐զ‐զめզ‐a‐զ‐զaզaːզːզ‐զːզ‐めզːզ‐զaզ‐զaզa‐զːզ—զ‐զ‐զ*զ‐զ—զめզめզaզ‐ːզaːaզ‐a=~~и', '~=ːaզ‐զ‐զ‐զ‐a‐a‐զ‐զ‐—զめaզaզ‐զ—զaզ—զ‐զː—զ—զめ‐—‐զ*ːզː‐զaզː—զめզ‐զ‐զ‐զ—aզː—զ‐զ‐զ‐զː—զ‐զ‐=iβ~βまи=и=', '~զaզめզ‐զ‐զ‐զ‐զ‐զ—զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐aզaզaզaզ—aզ‐զː—զ‐զ—զ‐զ‐զ‐զː—զa—զ‐զ—զ—զ‐զːaզ—զめզaզː—զːզː—զ‐զ‐զːզめզ‐զa=β~βቀ=', 'ψː—զめզ—զ*զaզ‐զ*‐զ‐զːզaզ—aզaզめզめզめզめ—զaզa—զめ‐զ*զ‐զ*զめզaզ‐զ‐զめզaզːզ~яиβия', 'ψːզ—aզめզ‐զː‐aզaզ—aː—զ—զːզːզめզaզa—ːզ*—aզ—ːզめ‐զ‐aզ‐զ‐ː‐め‐զːզめզ‐զ‐ːզːզ‐զ‐զ‐զaզː‐—զ‐զ‐ːզ‐զ‐aզめզ‐զ‐զaզ‐զ‐ː—ːզめզaզ‐զ≥›βEиaи=ቀ', '~զ‐め‐めզ‐զaզめզ*զ—զめզːaːaզ—զめզ—ːaզːզaզaզ—զːզ‐զ—ːզːզaզめզめզめa‐զめզ‐զめ~青~ββ', 'ψզ—‐め‐զ‐め‐զ‐め‐զːզaզ—զ‐զːզめզ‐զ—զ‐め‐զめ‐めզ*—ːզ‐め—զːզː—զ‐զ—զ—め—զːզめ‐め‐զ‐զめ*—aզːզːզ‐զːզː—զめզめːզ—զːզ‐—‐ːզめզ—զ—զ—ːզːզ—զ‐め‐—զめզめ‐զ‐a‐め‐め‐զ‐め‐զ‐զ‐≥青~', '~aզaːզa—a—aզめզめզ—aզaզːզːզa—ːaːaզ*—aːaզaզːզaզ—aːզː—ːզ~=β~β~β~β~β', 'β‐—‐զ‐զ‐զ‐զ‐զ‐զ‐めզ‐զ‐զ‐զ‐զめ‐զ—զaզめ‐զ—a—めզ‐めզ‐զめզめզaզ‐զ‐զ‐めզ‐զ‐めզ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐aᆨiиiиまi', '~զː—զaզ—a—aզaզːզa—aːaզ—ːզ—զ*—զ—ːaզaզaզːaːaզaː—aզ=~β~β~', '~զ‐զ—ːզ—զめզ‐զaզ—զːaզːզ—զːզ—զː*—զːզ—զaզaզːզaզ—զ‐զːզ‐զa—զː*ːaզ—զ—զːզaːaզ‐զ—զめզːզ‐զ=‐զa=β=~=β=β', '~め‐զめиめզめ‐めզめզめզめaめզ‐զめa—զa—զめզめզلaզめaզめa—զめaզめզめզめզ—զaզ—めզ—զめզめզ—զめզ—զ—զiи~β', '~զ‐զ‐զめ‐զめ‐—‐զ‐զ‐զ‐զ—a—զ‐զ‐զ—զめզめզ—զ—զa‐զ‐զ‐զ‐զ‐զめզ‐զ—զめզめː—զめ—զːզめզ—զ—զːզ‐զ‐զːaզ—ːզ‐զːaզ‐զめզaզaզ‐զ‐զめզ‐զ‐զːզ‐զ‐ːզ‐aզ‐զ‐զ‐զa=β~β=β', '~ː—‐զ‐զ‐a‐զ‐զ‐a‐զ‐զ‐ːզaզ*զaզː—զ—ːaզːaզ‐զ*aզ—ː—aզːզːզːզめզ‐զaiaia', '~—‐զ‐զ‐զ‐զ‐զめ‐զːզ—ː—զ—aզːզaզ—զめզ—զa‐めզ‐զ—ːզ—ːզめ*めզ‐զめզめզ—զːզ‐զaզめզめզ‐զめզ‐զ—զːզ‐զ—զ‐զ‐զめa青β~β', '~զ‐զ‐զ‐զ‐め‐զaզaզ‐めզ*—=զめզ—զめզ‐զ‐զ—ːզ‐զめզめզ‐զ—զめզめզめզめզ‐զ—զːզːզ厳զ=‐զめ—めզめզːզめզめզめզːզめːզ‐զ—めːզ‐զ‐զめ‐զ—զめ‐զめզ‐めզめ‐め‐զ‐—ψiまиiиi', 'ψaːզ—զ—զ—զめզ‐զ—զ—‐զ‐—զ—զːզ‐զ‐զaզ—aզːզːզ—զːaզːզ—զaզ—զ—զːզ—め—զ—զ‐զ‐զ‐զ—զ—զ—զ‐めզ—iまiま=я=i=', '~‐めզːզ‐めզ‐զ‐զ‐զ‐め‐զめզめզ‐めզːզ‐զ‐めզ‐め—aզ‐զめզ‐*զ‐զ‐զ‐*‐=‐aめ=զ‐զ‐=青βEiβ', '~զ‐զ‐զ‐aզ‐զ‐զ‐զaզa—զ—զːզaզ—զaզ*զ—զ*զ—զaզめզ‐զ‐զ‐զː—զːզ=', 'ψզめզ‐զ‐զ‐զ‐զ‐զ‐զめզ‐զめզːզːզ—զめզ*—ː—զ��զめզ—զ‐զめզめզ*‐め‐զ‐=i~i~a', '~‐զ‐զ‐զ‐め‐զ‐զ‐զ‐զ‐めզめ‐զ‐զ‐զめզめզめ‐զ‐զめզ‐めզ‐զめզ—զめզ‐զめ‐զ‐զめզːめ‐զ‐զめզ‐զ‐զ‐զ‐զめզめզ‐めզ‐զ‐զめզめ‐めզ‐զ‐զ‐զ—‐զめ‐め‐め‐め‐め‐~i', '~‐զ‐め‐զ‐զ‐զ‐めզ‐զːզ‐զめզ‐զ‐զːa‐—զ—զめզ*զ‐զːզめզめզ‐զ‐めզ‐ː‐զːզ‐զめ‐զめ‐=›i›~›i~ま~', '~‐զ‐զ‐め‐め‐զ‐զ‐զ‐めզ‐զめզめզ‐զ‐め‐զ‐զ‐զめզめ‐զ‐զ‐զ—զ‐զaզめզ‐զ‐զ‐զ‐めզ—‐զ‐զ‐զめզ‐զ‐զめզめզ‐め‐զ‐զ—め—‐զ‐զめզː‐զ‐զめ‐め‐aiβiβiβiβiβまa', '~‐めզ‐զ‐め‐զ‐զ‐զめզめ‐めզ‐զ‐զ—զ‐զ‐զ‐զ‐զ‐զめզ‐զ‐զ*め‐զめզめզaզめզaզ‐զaզ‐զ‐ː‐զ‐զ‐=β~β=β~', '~ːa—ːめզ‐め‐め‐め‐—‐め‐զaめ—զ‐め‐め‐め‐め‐զめːզ‐զː—aզ‐զː*—զaզːզːզ—aː—զ‐զ=яøβ=β=β', '~aզめ—զ—ː‐め‐a‐め‐め‐め‐め‐め‐め‐め‐め‐—‐զ‐զː—aːզめզめզめզ—aզめզ‐զ—ːզ‐զ‐զ‐զːզめզ—aզめ—զːaː‐զ—զ‐զめզめզ‐め‐め‐ːaめːaզめզ—զːզ‐=iβiᆨβя.иa.яβ', 'ψ‐զ‐զ‐め‐զ‐—զːզ‐զ‐զ‐զ*aզ‐め—めզaːզ‐զ‐զ‐զ‐めզː‐զ‐—ːզ‐զ‐めզ‐զ‐ːզ‐զ‐զめ—ːզ‐զ‐զ‐զめզːaզ‐ː‐զː—ːզ‐զ‐զːզ‐զ‐ː‐ːզ‐—զ‐a‐=ββ=β', '=զめզめզː‐զ—զめզめզːզːզːaզめa—զ‐զ—لaզ—զ—aզめːզː—ːզめaːզ—ːզːզ*զːaզめզ‐զめզ—ːզめ—‐め‐a—ːզ—զめː‐め‐զ‐a‐զa~β~β', '~‐զ‐զ‐め‐め‐め‐a‐զ‐զめզ‐զ‐զ‐զ‐զ—զ—զめզːզ‐զ‐զːզːզ—ː—ːզめզ—զːզ—ːզ‐զ‐զ—զ‐զ‐զめː—զめզ‐զ—զ‐զ‐զめզːզːզ‐զ‐զ‐めզː—զ‐զ‐զめզ—ːզ‐զːզː—զː—զめզ‐aβ~まaまEяEβ', '~զaզめզːզめզːզ‐զめ—զaզめզ‐զːզめզːզ‐զ‐զaզːaզ‐զめːզːզめզめզːզめզːզ—զ‐め‐զめa青β~β~β', '~‐զ‐զめզ*զ‐զ‐զ‐めզめզa‐めզ‐めզ‐ː‐զ‐զ‐ːզめզ‐めզ‐զ‐զ‐զ‐ːզ‐զめ‐めզめզ‐զ‐զめզめզめզめզ‐զ‐զ‐զːզめզ‐զめզ‐զ‐め‐ل‐め‐զ‐զ‐=i', '~‐ː‐զ‐a‐ːզ‐զめ—զめզめ‐զ‐զ‐զ—զ‐զ‐զːզ‐զ—զ‐զ‐զːզ‐զ‐զaզ—զ‐զaզ—զ‐զ‐—ːզ‐զ‐զ‐զ‐a青~=', '~‐զめ‐め‐め‐զ‐զ‐めզ‐զ‐զめ‐զめ‐め‐զめ‐めզ‐զめզ‐めզめ‐զ‐զ‐զめզめ‐զ‐めզ‐զめ‐զ‐զ‐զ‐めզ‐めզめզめզ‐めզめ‐զめզめզ‐զめ‐めզ‐め‐զ‐めզめ‐զ‐զ‐զ‐զ‐めզめ‐め‐め‐めզ‐め‐զめզ‐զ‐զ‐զ‐め‐め‐զ‐め‐—β~βиi', '~ː—=զめզ*զめզaզ—զめզ—զ*զaզ—զːզめզːめːզめզめզめզめ*զaːaզめզ*զːզめզめզめզːaզめզめզめզめզめզaզ=i~', 'ψզ—ːզaզ‐զ‐め‐զ‐զめզ*զ‐զ‐զ‐զ‐めզ‐զ‐aզ‐զ‐զ‐զ—ːզːզːaːզːզ—զ—զːզめզː—զːզ‐զ‐≥i~β~β~яまяя=я=яβ', '~ː—めզめզめզめզめզめզめզめզ—a—aզめզaզ*զ*զaզaզ—ː—զaզa—aめզめզaː*զめզめ—ːa*aզ—զ—զ—զめզめզめːզめaめզaᆨяi~яaまa~a~', 'ψ‐ː—զ—‐ːզ—զめզめզ‐զ—ːզ‐զ‐めզめզ‐զ‐զ—a—a—զ‐զ—ːզ—զめզ‐めզめ=i=まa', '~զː‐զ‐a‐զ‐զ‐զ‐զ‐զめզ‐aզ‐զ‐զ‐զːզː—aզːaզːզ—ːaզ‐զːa—զ‐զ‐զ—զaզ‐զ‐զa=β~яβ', 'βː‐զ=‐զめզめզめզ‐めզめզめզめզめզめ—զː—զめզめ‐զ‐めզめզめզめզめզ‐めզ—զめզめaめ‐め—զaめ—զめ—めզめզ‐めզめզめ—ː‐զめ—めզめ—զめզ—զ—զめզめ—զめ—ψiβiβ', '~ːզaːめaզ‐զ‐զ‐め‐め‐め‐a‐զːզめ‐めզ‐զːaզめ—aզː*aզめզめզ—ːզaզめզː—զːզːզ—զ—զ—զめզ‐զ‐ː—ːզ‐め‐զ‐aiEи=и=иa=a', 'ψːզめ‐ːa—զ‐զ‐aզ‐—めզ‐զ‐զめːզːզめզːզめːզːզ‐զ—ː—զːզ—զːզːզ—զːզːզaզːզめզ‐զaめզめզːզ‐め‐զ‐a青βя=β', '~—ːզ‐զ‐զ—めզːզːզ—զaզa—めզaզaզ=β~a=', '~ːզ‐めզ‐a‐—զ‐զ—ːզ‐զ—aզa—զめզ—զaզ—զ‐զ—aːզめզ—զめզめզ—զめզ*—զ‐զ—aːզ‐զめ*զ—զːaː‐զめզ‐めզめզ‐めa‐զa~', '~‐զ‐զ‐զ‐զ‐զa‐*զ‐զ‐զ‐めa—զ‐զ‐զː—զめզa—զめզ‐めզめզめզ‐めզːզめզ‐զめզ‐զː—զːaզːզ‐զ‐ːզːզめզめզ—զめ—‐զ‐զ—øiиiиiиま', '~զめզ—��զ‐զ‐զː‐զaզ‐զ‐զめզ—‐զːզ‐զめզa—զaզaզːզː—զ‐զa—զːaզ‐զめːզ‐զːզː‐զ‐զ—ːզːզ=βቀ']
39
+ {'eval_loss': 20.983184814453125, 'eval_wer': 1.0, 'eval_runtime': 2.7688, 'eval_samples_per_second': 21.67, 'eval_steps_per_second': 0.361, 'epoch': 0.94}
40
+ 31%|█████████████████████████████████████████████████████▊ | 20/64 [01:22<03:05, 4.21s/it]The following columns in the evaluation set don't have a corresponding argument in `Wav2Vec2ForCTC.forward` and have been ignored: input_length.
41
+ ***** Running Evaluation *****
42
+ Num examples = 60
43
+ Batch size = 64
44
+ 31%|█████████████████████████████████████████████████████▊ | 20/64 [01:24<03:05, 4.21s/it]
45
+ ['~զめզ‐զ‐a‐め‐զ‐զ‐めզ‐め‐զ‐め‐めզめ‐めզ‐զ‐զ‐զめզ‐め‐զ‐զ‐զ—զ‐զːめզ‐զめ‐զ‐զ‐ːめզ‐զ‐զ‐め~青βまиi=', '~զːզ—զaզa*զaզaզːզːզ—զːզaզ—զ—ːզaզː—զ—զ=~β~β=ま=まиaզa', '~‐զ‐զaզːզめզめ‐զaːզ‐զaզ—զ‐զːզːզ—զaզ—‐զ‐զaզ‐զ‐զ‐զ‐զ‐—‐—iиまиまиまи', 'ψ‐ː‐—‐զ‐—զaզめզめզ‐զ‐զ‐զ‐զめզ‐զ‐զ‐զ‐զめզ—զ‐զめ—զːզ‐めզ—めːզ‐めզ—զːaզaզ‐զːզ‐め‐զめզめզめ‐զ‐め‐め‐め‐め‐め‐め‐=ᆨiまi=ま', 'ψaզ‐զ‐զ‐a‐զːզ‐զ‐զːզ—զ‐զ‐զ‐զ—aզ‐めզːaːaզめaզ‐զːզ—ːզ‐զめզ—ː—զa‐めզめզ‐*զ‐a‐aզ‐զ‐զ‐a~β~β~E=E=', 'ψːզːզ‐զ‐զ‐զめ—զ‐め‐ːզ‐զ‐զ*զːզ‐զ‐զaզ*զ—ːզ‐զaզ*զaːզːզːզ‐զːզめզ‐զ‐զːզ‐զめզ‐զ=iᆨβ~βᆨβᆨβ=', 'β‐զめզめզ‐զ—զa—զ—զ‐զ‐զ‐զ—ːզ‐զめզ‐զ—զːզːզ‐զ‐զめ‐զ‐—զ—զめ‐զめզ‐զめ‐めզめ‐զ‐め‐զめզ‐զ‐զːզ‐めզめզ‐զ—զ‐զめ‐զ‐զ‐aiиiиまиまиまиま', '~‐զ‐զ‐զ‐a‐զ‐め‐զめզめզ‐めզめːզ‐զ‐めզ‐զめզ‐զ‐զːզ‐զめզめզ‐զ—ːզ‐զ‐զ—զ‐զ‐զ‐զめզ‐め‐め‐զ‐a‐aզ‐a‐a‐=β~β', '~զ‐զиaզめզめզめզめզめ‐めզめզめզめ—めզ—ːaզめզめզːaզめ—ːզめզめզめզめզ—めզめզːaզめ—զめզめզめ—ːզめզめզめզめːaզめզ—a—զめզめզめզめզめզめզめզめզめզめզめզめզめզめզ—めզめզめզめaᆨиβ', '~‐ːզめզ—զ‐aզめ‐め‐め‐и‐զ‐め—ːզ‐զ‐զ‐զ‐զːզaզa—զめզ‐զ—զめզめզa‐զ‐զめզ‐めզːզめզ‐զめզ‐ːめզ‐զめ‐զめզめզ‐զめ—め‐զめːզ‐—めզめզ‐զ‐めզːզ‐めզめզ‐զ‐զ‐զ‐զ‐めզめ‐զ‐զ‐めaめ*めզめ‐めզめզ‐=iEiቀま甌β', '~‐զ‐զめզ‐զめզめզ‐めː—aզ*ːզ‐զ‐զめզ‐めզ‐めզめզː—զːaզめզ‐զめզめːզ‐めզめ—զ‐զ‐զaզ*զ‐զ‐զ‐め‐めզ‐զ‐զめզː‐զめզ—ːզ‐զ‐զめզːզ‐զ‐զめ‐զ‐զ‐զ‐զːめզめզめզめ=զ‐զめ‐めզめ‐め‐զめզめ‐զ‐めզめ‐զ‐=i~i~i~', 'ψզ‐a—զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐め‐զ‐め‐զ‐め‐*‐զ‐զ‐զ—ːզ‐զ‐զ‐զ—զめզめզ‐զめ‐め‐զめզめզːզめզːaզːめզめզ‐զ‐—‐め‐め‐—զ‐զ‐め‐զ‐め‐め‐զ‐—‐ː‐զ—զːզ‐զ‐զ‐=iиiиi甌i甌β', '~‐и‐զ‐a‐め‐*‐զ‐*‐めզ甌զめզ‐*զ‐զ‐զ‐զ‐*めզ‐զ*‐甌‐զ‐զ—զ*զ‐զ‐զ‐զ*զ‐զ*զ‐զめզめզめզめզ‐զめզめզ‐めզめզ‐զ‐զ*‐զめզめ‐—‐զ‐զ‐զ‐—զ‐զめզめզ*‐զめզ‐զ‐*‐զ‐めզ‐զ‐*‐め‐*‐め‐め‐め‐め‐め‐—‐ː‐զ—‐*‐め‐め‐aiτi', '~‐—‐—‐զ‐զ‐զ‐ː‐めզːզめզめզ—զ‐զ‐զ—զ—ː—զめզめզ‐զめզːզːզ‐զːզめ‐զːզ‐զ‐զːめզːめզめզ—զaզːզめ‐めզ‐զːめːզめզめ—զめզ‐զ‐զ‐զ‐զ—զːզːզ‐զ‐ː—‐զ‐զ‐a‐զ‐aiβ~まaまiβ', '~‐ːզめզめզめզ‐զ‐զa*զ—aզ‐զːզːզ—զaːaːզめզ—զa—aզːaզ—aːզːզːզ—զːզːզめզ‐զa青β~青まaиa=', '~a—めa‐զ‐aզ‐զ‐զめզ‐a‐զ‐զaզaːզːզ‐զːզ‐めզːզ‐զaզ‐զaզa‐զːզ—զ‐զ‐զめզaզ‐զ—զめզめզa‐ːզaːaզ‐a=~~и', '~=ːaզ‐զ‐զ‐զ‐a‐a‐զ‐զ‐—զめaզaզ‐զ—զaզ—զ—զ‐զː—զ—զめ‐—‐զ*ːզː‐զaզː—զめզ‐զ‐զ‐զaզːզ‐զ‐զː—զ‐զ‐=iβ~βまиaииии=', '~զaզめզ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐a‐aզ‐զaզaː—aզ‐զː—զ‐զ—զ‐զ‐զ‐զめզː—զːզa—զ‐զ—զ—զ‐զːaզ—զめզaզː—զːզː—զ‐զːզ‐զːզめզ‐զa=β~βቀ=', 'ψː—զめզ—զ*զa‐զ*‐զ‐զːզaզ—aզaզめզめզめզめ—զaզa—զめզ‐զめ‐զ*զ‐զ*զめզaզ‐զ‐զめզaզːզ~яиβиβия', 'ψːզ—aզめզ‐զめզː‐aզːaզ—aː—զ—զːզːզめզaզa—ːզ*—aզ*—ːզめ‐զ‐aզ‐զ‐a‐め‐զːզめզ‐զ‐ːզːզ‐զ‐զ‐զaզː‐—զ‐զ‐ːզ‐զ‐aզめզ‐զ‐զaզ‐զ‐ː—ːզめզaզ‐զ≥›βEиaи=ቀ', '~զ‐め‐めզ‐զaզめզ*զa—զめզːaːaզ—զめզ—ːaզːզaզaզ—զːզ‐զ—ːզːզaզめզめզめa‐զめզ‐զめ~青~~β', 'ψզ—‐զ‐め‐զ‐め‐զːզaզ‐զ—aզ‐զːզːզめզ‐զ—զ‐め‐զめ‐め‐զ‐*ːզ‐め—‐զːզː—զ‐զ—զ—め—めզːզめ‐め‐զめ*—aզːզːզ‐զːզːզめզめːզ—ːզ‐—‐ːզめզ—զ—զ—ːզːզ—զ‐め‐զめզめ‐a‐め‐め‐զ‐զ‐≥青~', '~aզaːզa—a—aզめզめզ—aզaզːզːզa—ːaːaզ*aզaզːզaզ—aːզː—ːզ~=β~β~β~β~β', 'β‐—‐զ‐զ‐զ‐զ‐զ‐զ‐めզ‐զ‐զ‐զめ‐զ—զaզめ‐զaめզ‐めզ‐զめզaզ‐զ‐զ‐めզ‐զ‐զ‐めզ‐めզ‐զ‐զ‐զ‐զ‐զ‐զ‐զ‐aᆨiиiиまi', '~զːaզaːզ—a—aզaզːզa—aːaզ—ːզ—զ*aզa—ːaզaզaզːaːaզaː—aզ=~ββ~β~', '~զ‐զ—ːզ—զめզ‐զaզ—զaզːզːզːզ—զː*—զːզ—զːaզaզːզaզ—զ‐զːզ‐զa—ː*ːaզ—զ—զːզaːaզ‐զ—զめզːզ‐զ=‐զa=β=~=β=β', '~め‐զめиめզめ‐め‐めզめ‐զめaめզ‐զめa—զa—զめզ‐めզめلaզめaզめa—զめaզめզめզめզ—զaզめզ—զめզめզ—զめզ—զめ—զiи~β', '~զ‐զ‐զめ‐զめ‐—‐զ‐զ‐զ‐զ—a—զ‐զ‐զ—զめզめզ—զ—զa‐զ‐զ‐զ‐զ‐զめզ‐զ—զめզめː—զめ—զːզめզ—զ—զːզ‐զ‐զːaզ—ːզ‐զːaզ‐զめզaզaզ‐զ‐զめզ‐զ‐զːզ‐զ‐ːզ‐aզ‐զ‐զ‐զa=β~ββ=β', '~ː—‐զ‐զ‐a‐զ‐a‐զ‐զ‐ːզaզ*զaզː—զ—ːaզːaզ‐めզ*aզ—ː—aզːզːզːզめզ‐զaiaia', '~—‐զ‐զ‐զ‐զ‐զ‐め‐զめզːզ—զ—aզːզaզ‐զ—զめզ—զa‐めզ‐զ—ːզ—ːզ‐զめ*めզ‐զめզめզ—զːզ‐զaզめզめզ‐զめզ‐զ—զːզ‐զ—զ‐զ‐զめa青β~β', '~զ‐զ‐զ‐զ‐め‐զaզaզ‐めզ*—=զめզ—զめզ‐զ‐զ—ːզ‐զ‐զ‐զめզめզ‐զ—զめզめզめզめզ‐զ—զːզːզ厳զ=‐զめ—めզめզːզめզめզːզめːզ‐զ—めːզ‐զ‐զめ‐զ—զめ‐զめզ‐めզめ‐め‐զ‐—ψiまиi', 'ψaːզ—զ—զ—զめզ‐զ—զ—‐զ‐—զ—զːզ‐զ‐զaզ—aզːզːaզ—զːaզːզːզ—զaզ—զ—զːզ—め—զ—զ‐զ‐զ‐զ—զ—զ—զ‐めզ—i=iま=i=', '~‐めզːզ‐めզ‐զ‐զ‐զ‐め‐զめզめ‐めզːզ‐զ‐め‐զ‐め—aզめ‐զめզ‐*զ‐զ‐*‐=‐aめ=զ‐զ‐=青βEiβ', '~զ‐զ‐զ‐aզ‐զ‐զ‐զaզa—զ—զːզaզ—զaզ*զ—զ*զ—զaզめզめզ‐զ‐զ‐զめզ‐զː—զːզ=', 'ψզめզ‐զ‐զ‐զ‐զ‐զ‐զ‐զめզ‐զめզːզːզ—զめզ*—ː—զめզめզ—զ‐զめզめզ*‐め‐め‐=i~i~a', '~‐զ‐զ‐め‐զ‐զ‐զ‐զ‐めզめ‐զ‐զ‐զめզめ‐め‐զ‐զめզめ‐めզ‐զめզ—զめզ‐զめ‐զ‐զめզːめ‐めզ‐զめզ‐զ‐զ‐զ‐զめզめզ‐めզ‐զ‐զめզめ‐めզ‐զ‐զ‐զ—‐զめ‐め‐め‐め‐め‐~i', '~‐զ‐め‐զ‐զ‐զ‐めզ‐զːզ‐զ‐զːզめզ‐զ‐զːa‐—զ—զめզ*զ‐զːզめզめզ‐զ‐めզ‐ː‐զːզ‐զめ‐め‐=›i›~›i›i~ま~', '~‐զ‐զ‐め‐め‐զ‐զ‐զ‐めզ‐զめզめզ‐զ‐め‐զ‐զ‐զめզめ‐զ‐զ‐զ—զ‐զaզめզ‐զ‐զ‐զ‐めզ—‐զ‐զ‐զめզ‐զ‐զめզめզ‐め‐զ‐զ‐զ—め—‐զ‐զめզ‐զː‐զ‐զ‐զめ‐め‐aiβiβiβiβiβまa', '~‐めզ‐զ‐め‐զ‐զ‐զめզめ‐めզ‐զ‐զ—զ‐զ‐զ‐զ‐զ‐զ‐զめզ‐զ‐զ*め‐զ‐めզめզaզめզaզ‐զa‐զ‐ː‐զ‐զ‐=β~β=~', '~ːa—ːめ‐め‐զ‐め‐—‐め‐զaめ—զ‐め‐め‐め‐め‐զ‐めːզ‐զ—aզ‐զːa—զaզːզːզ—aː—զ‐զ=яøβ=β=β', '~aզめզː‐め‐a‐め‐め‐め‐め‐め‐め‐め‐め‐—‐զ‐զː—aːզめզめզめզめզ—aզめ‐զ—ːզ‐զ‐զ‐զːզめզ—aզめ—զːaː‐զ—զ‐զめզめզ‐め‐め‐ːaめːaզめզ—զːզ‐=iβiᆨβяa.яβ', 'ψ‐զ‐զ‐զ‐���զːզ‐զ‐զ‐զ*aզːզ‐め—めզaːզ‐զ‐զ‐զ‐めզː‐զ‐—զ‐զːզ‐めզ‐զ‐ːめ‐զ‐զめ—ːզ‐զ‐զ‐զめզːaզ‐ː‐զめզː—ːզ‐զ‐զːզ‐զ‐ː‐ːզ‐—զ‐め‐=β=β', '=զめզめզː‐զ—զめզめզːզːզːaզめa—զ‐զ—لaզ—զ—aզめːզː—ːզめaːզ—ːզːզ*զめːaզめզ‐զめզ—ːզめː‐め‐a—ːզ—զめː‐め‐զ‐a‐զa~β~β', '~‐զ‐զ‐め‐め‐め‐a‐զ‐զめզ‐զ‐զ‐զ‐զ—զ—զめզːզ‐զ‐զːզːզ—ː—ːզめզ—զːզ—ːզ‐զ‐զ—զ‐զ‐զめː—զめզ‐զ—զ‐զ‐զ‐զめզːզːզ‐զ‐զ‐めզː—զ‐զ‐զめզ—ːզ‐զːզː—զː—զめզ‐aβ~まaまEяEβ', '~զaզめզːզめզːզ‐զめ—aզめ‐めզ‐զːզːめզːզ‐զ‐զ‐զ‐զ‐զaզːaզ‐զːզːզめզめզːզめզめːզ—զ‐め‐զめa青β~β~β', '~‐զ‐զめզ*զ‐զ‐զ‐めզめզa‐めզ‐めզ‐ː‐զ‐զ‐ːզめզ‐めզ‐զ‐զ‐զ‐ːզ‐զめ‐めզめզ‐զ‐զめզめզめզめզ‐զ‐զ‐զめզ‐զめզ‐զ‐め‐め‐ل‐め‐զ‐զ‐=i', '~‐ː‐զ‐a‐ːզ‐զめaզめզめ‐զ‐զ‐զ—զ‐զ‐զːզ‐զ—զ‐զ‐զːզ‐զ‐զaզ—զ‐զaզ‐զ—զ‐զ‐—ːզ‐զ‐զ‐զ‐զ‐a青~==', '~‐զめ‐め‐め‐զ‐めզ‐զ‐զめ‐め‐զめ‐め‐めզ‐めզ‐զめզ‐めզめ‐զ‐զ‐զめզめ‐զ‐めզ‐զめ‐զ‐զ‐զ‐めզ‐めզめզめզ‐めզめ‐զめզめզ‐զめ‐めզ‐め‐զ‐めզめ‐զ‐զ‐զ‐զ‐めզめ‐め‐め‐めզ‐め‐զめզ‐զ‐զ‐զ‐め‐զ‐め‐զ‐め‐—β~βиi', '~*զ—=զめզ*զめզaզ—զめզめզ—զ*զaզ—զːզめզːめːզめզめզːզめզめ*զaːaզめզ*զːզめզめզめզːaզめզめզめզめզめզaզ=i~', 'ψզ—ːզaզ‐զ‐め‐զ‐զめզa‐զ‐զ‐զ‐զ‐めզ‐զ‐aզ‐զ‐զ‐զ—ːզːզːaːզː—ːզ—զ—զːզめզː—զːզ‐զ‐զ‐≥i~β~яβ', '~ː—めզめզめզめզめզめզめզ—aզめզaզ*զ*զaզaզ—ː—զaզaめզめզaː*զめզめ—ːa*aզ—զ—զ—զめզめːզめaめզaᆨяi~яaяa~a~', 'ψ‐—զ—‐ːզ—զめզめզ‐զ—ːզ‐զ‐めզめզ‐զ‐զ—a—a—զ‐զ—ːզ—զめզ‐めզめ=i=まa', '~զː‐a‐զ‐զ‐զːզ‐զ‐զめզ‐aզ‐զ‐զ‐զめզ‐զːզːզː—aզːaզːզ—ːaզ‐զ‐զːa—զ‐զ‐զ‐զ—զaզ‐զ‐զa=β~β', 'βː‐զ=‐զめզめզめզ‐めզめզめզめզめ—զːզめզめ‐զ‐めզめզめզめզ‐めզ—զめզめaめ‐め—զaめ—զめ—めզめզ‐めզめզめ—ː‐զめ—めզめ—զめզ—զ—զめզめ—め—ψiβiβ', '~ːめaːめa‐զ‐め‐め‐め‐a‐զːզめ‐めզ‐զːaզめ—aզː*aզめզめզめզ—ːզaզめզː—զːզːզ—զ—զ‐—զめզ‐զ‐ː—ːզ‐め‐զ‐aiE=и=a', 'ψːզめ‐ːa—զ‐զ‐aզ‐—めզ‐զ‐զめː‐զːզめզːզめːզaզaզ‐զ—ː—զːզ—զːզːզ—զːզːզaզːզめզ‐զ‐զaめզめզːզ‐զ‐め‐զ‐a青βя=β', '~—ːզ‐զ‐զ—めզːզːզ—զaզa—めզaզaզ=β~a=', '~ːզ‐めզ‐a‐—զ‐զ—aզ‐զ—aզa—զめ—զ—զaզ—զ‐զ—aːզめզ—զめզめզ—զめզ*—զ‐め—aːզ‐զめ*զ—զːaː‐զめզ‐めզめզ‐めa‐զa~', '~‐զ‐զ‐զ‐զ‐զ‐a‐*զ‐զ‐զ‐めa—զ‐զ‐զː—զめզːa—զめզ‐めզめզめզ‐めզːզめզ‐զめզ‐զː—զめզːaզːզ‐զ‐ːզːզ‐զめզめզ—զめ—‐զ‐զ—øiиiиiиま', '~զめզ—めզ‐զ‐զː‐զaզ‐զ‐զめզ—‐զːզ‐զめզa—զaզaզːզː—զ‐զːզa—զːaզ‐զめːզ‐զːզː‐զ‐զ—ːզːզ=βቀ']
46
+
47
+
48
+ 34%|███████████████████████████████████████████████████████████▏ | 22/64 [01:29<02:44, 3.91s/it]Traceback (most recent call last):
49
+ File "run_speech_recognition_ctc.py", line 749, in <module>
50
+ main()
51
+ File "run_speech_recognition_ctc.py", line 700, in main
52
+ train_result = trainer.train(resume_from_checkpoint=checkpoint)
53
+ File "/opt/conda/lib/python3.8/site-packages/transformers/trainer.py", line 1347, in train
54
+ for step, inputs in enumerate(epoch_iterator):
55
+ File "/opt/conda/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 521, in __next__
56
+ data = self._next_data()
57
+ File "/opt/conda/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 561, in _next_data
58
+ data = self._dataset_fetcher.fetch(index) # may raise StopIteration
59
+ File "/opt/conda/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 52, in fetch
60
+ return self.collate_fn(data)
61
+ File "run_speech_recognition_ctc.py", line 282, in __call__
62
+ batch = self.processor.pad(
63
+ File "/opt/conda/lib/python3.8/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py", line 147, in pad
64
+ return self.current_processor.pad(*args, **kwargs)
65
+ File "/opt/conda/lib/python3.8/site-packages/transformers/feature_extraction_sequence_utils.py", line 229, in pad
66
+ value = value.astype(np.float32)
67
+ KeyboardInterrupt
wandb/run-20220130_175417-3n44neww/files/requirements.txt ADDED
@@ -0,0 +1,180 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ aiohttp==3.8.1
2
+ aiosignal==1.2.0
3
+ analytics-python==1.4.0
4
+ anyio==3.5.0
5
+ appdirs==1.4.4
6
+ argon2-cffi-bindings==21.2.0
7
+ argon2-cffi==21.3.0
8
+ asgiref==3.5.0
9
+ asttokens==2.0.5
10
+ async-timeout==4.0.2
11
+ attrs==21.4.0
12
+ audioread==2.1.9
13
+ backcall==0.2.0
14
+ backoff==1.10.0
15
+ bcrypt==3.2.0
16
+ beautifulsoup4==4.9.3
17
+ black==21.12b0
18
+ bleach==4.1.0
19
+ brotlipy==0.7.0
20
+ certifi==2020.12.5
21
+ cffi==1.14.3
22
+ chardet==3.0.4
23
+ charset-normalizer==2.0.10
24
+ click==8.0.3
25
+ conda-build==3.21.4
26
+ conda-package-handling==1.7.2
27
+ conda==4.9.2
28
+ configparser==5.2.0
29
+ cryptography==3.2.1
30
+ cycler==0.11.0
31
+ datasets==1.18.2.dev0
32
+ debugpy==1.5.1
33
+ decorator==4.4.2
34
+ defusedxml==0.7.1
35
+ dill==0.3.4
36
+ dnspython==2.1.0
37
+ docker-pycreds==0.4.0
38
+ entrypoints==0.3
39
+ executing==0.8.2
40
+ fastapi==0.73.0
41
+ ffmpy==0.3.0
42
+ filelock==3.0.12
43
+ fonttools==4.29.0
44
+ frozenlist==1.3.0
45
+ fsspec==2022.1.0
46
+ gitdb==4.0.9
47
+ gitpython==3.1.26
48
+ glob2==0.7
49
+ gradio==2.7.5.2
50
+ h11==0.13.0
51
+ huggingface-hub==0.4.0
52
+ idna==2.10
53
+ importlib-resources==5.4.0
54
+ ipykernel==6.7.0
55
+ ipython-genutils==0.2.0
56
+ ipython==8.0.1
57
+ ipywidgets==7.6.3
58
+ jedi==0.17.0
59
+ jinja2==2.11.3
60
+ jiwer==2.3.0
61
+ joblib==1.1.0
62
+ json5==0.9.6
63
+ jsonschema==4.4.0
64
+ jupyter-client==7.1.2
65
+ jupyter-core==4.9.1
66
+ jupyterlab-pygments==0.1.2
67
+ jupyterlab-server==1.2.0
68
+ jupyterlab-widgets==1.0.2
69
+ jupyterlab==2.2.9
70
+ kiwisolver==1.3.2
71
+ libarchive-c==2.9
72
+ librosa==0.8.1
73
+ llvmlite==0.38.0
74
+ markdown2==2.4.2
75
+ markupsafe==1.1.1
76
+ matplotlib-inline==0.1.3
77
+ matplotlib==3.5.1
78
+ mistune==0.8.4
79
+ mkl-fft==1.3.0
80
+ mkl-random==1.1.1
81
+ mkl-service==2.3.0
82
+ monotonic==1.6
83
+ multidict==6.0.2
84
+ multiprocess==0.70.12.2
85
+ mypy-extensions==0.4.3
86
+ nano==0.10.0
87
+ nbclient==0.5.10
88
+ nbconvert==6.4.1
89
+ nbformat==5.1.3
90
+ nest-asyncio==1.5.4
91
+ notebook==6.4.8
92
+ numba==0.55.1
93
+ numpy==1.19.2
94
+ olefile==0.46
95
+ packaging==21.3
96
+ pandas==1.4.0
97
+ pandocfilters==1.5.0
98
+ paramiko==2.9.2
99
+ parso==0.8.1
100
+ pathspec==0.9.0
101
+ pathtools==0.1.2
102
+ pexpect==4.8.0
103
+ pickleshare==0.7.5
104
+ pillow==8.1.2
105
+ pip==21.3.1
106
+ pkginfo==1.7.0
107
+ platformdirs==2.4.1
108
+ pooch==1.6.0
109
+ prometheus-client==0.13.0
110
+ promise==2.3
111
+ prompt-toolkit==3.0.8
112
+ protobuf==3.19.4
113
+ psutil==5.8.0
114
+ ptyprocess==0.7.0
115
+ pure-eval==0.2.2
116
+ pyarrow==6.0.1
117
+ pycosat==0.6.3
118
+ pycparser==2.20
119
+ pycryptodome==3.13.0
120
+ pydantic==1.9.0
121
+ pydub==0.25.1
122
+ pygments==2.8.0
123
+ pynacl==1.5.0
124
+ pyopenssl==19.1.0
125
+ pyparsing==3.0.7
126
+ pyrsistent==0.18.1
127
+ pysocks==1.7.1
128
+ python-dateutil==2.8.2
129
+ python-etcd==0.4.5
130
+ python-levenshtein==0.12.2
131
+ python-multipart==0.0.5
132
+ pytz==2021.1
133
+ pyyaml==5.4.1
134
+ pyzmq==22.3.0
135
+ regex==2022.1.18
136
+ requests==2.24.0
137
+ resampy==0.2.2
138
+ ruamel-yaml==0.15.87
139
+ sacremoses==0.0.47
140
+ scikit-learn==1.0.2
141
+ scipy==1.7.3
142
+ send2trash==1.8.0
143
+ sentry-sdk==1.5.4
144
+ setuptools==50.3.1.post20201107
145
+ shortuuid==1.0.8
146
+ six==1.15.0
147
+ smmap==5.0.0
148
+ sniffio==1.2.0
149
+ soundfile==0.10.3.post1
150
+ soupsieve==2.2
151
+ stack-data==0.1.4
152
+ starlette==0.17.1
153
+ subprocess32==3.5.4
154
+ termcolor==1.1.0
155
+ terminado==0.13.1
156
+ testpath==0.5.0
157
+ threadpoolctl==3.0.0
158
+ tokenizers==0.11.4
159
+ tomli==1.2.3
160
+ torch==1.10.2
161
+ torchaudio==0.10.2
162
+ torchelastic==0.2.2
163
+ torchtext==0.9.1
164
+ torchvision==0.9.1
165
+ tornado==6.1
166
+ tqdm==4.62.3
167
+ traitlets==5.1.1
168
+ transformers==4.17.0.dev0
169
+ typing-extensions==4.0.1
170
+ urllib3==1.25.11
171
+ uvicorn==0.17.1
172
+ wandb==0.12.9
173
+ wcwidth==0.2.5
174
+ webencodings==0.5.1
175
+ wheel==0.35.1
176
+ widgetsnbextension==3.5.2
177
+ xxhash==2.0.2
178
+ yarl==1.7.2
179
+ yaspin==2.1.0
180
+ zipp==3.7.0
wandb/run-20220130_175417-3n44neww/files/wandb-metadata.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-4.15.0-151-generic-x86_64-with-glibc2.10",
3
+ "python": "3.8.8",
4
+ "heartbeatAt": "2022-01-30T17:54:18.803448",
5
+ "startedAt": "2022-01-30T17:54:17.515767",
6
+ "docker": null,
7
+ "gpu": "Tesla V100S-PCIE-32GB",
8
+ "gpu_count": 1,
9
+ "cpu_count": 60,
10
+ "cuda": null,
11
+ "args": [
12
+ "--dataset_name=mozilla-foundation/common_voice_8_0",
13
+ "--model_name_or_path=facebook/wav2vec2-xls-r-300m",
14
+ "--dataset_config_name=fr",
15
+ "--tokenizer_name_or_path=./",
16
+ "--output_dir=./",
17
+ "--overwrite_output_dir",
18
+ "--num_train_epochs=4",
19
+ "--per_device_train_batch_size=64",
20
+ "--per_device_eval_batch_size=64",
21
+ "--gradient_accumulation_steps=1",
22
+ "--learning_rate=2e-4",
23
+ "--warmup_steps=2500",
24
+ "--length_column_name=input_length",
25
+ "--evaluation_strategy=steps",
26
+ "--text_column_name=sentence",
27
+ "--save_steps=500",
28
+ "--eval_steps=5",
29
+ "--logging_steps=5",
30
+ "--layerdrop=0.0",
31
+ "--activation_dropout=0.05",
32
+ "--save_total_limit=20",
33
+ "--freeze_feature_encoder",
34
+ "--feat_proj_dropout=0.0",
35
+ "--mask_time_prob=0.6",
36
+ "--mask_time_length=10",
37
+ "--mask_feature_prob=0.25",
38
+ "--mask_feature_length=10",
39
+ "--gradient_checkpointing",
40
+ "--report_to=wandb",
41
+ "--run_name=xls-r-300m-fr",
42
+ "--max_train_samples=1000",
43
+ "--max_eval_samples=60",
44
+ "--max_duration_in_seconds=10",
45
+ "--use_auth_token",
46
+ "--fp16",
47
+ "--group_by_length",
48
+ "--preprocessing_num_workers=64",
49
+ "--do_train",
50
+ "--do_eval",
51
+ "--load_best_model_at_end",
52
+ "--push_to_hub"
53
+ ],
54
+ "state": "running",
55
+ "program": "run_speech_recognition_ctc.py",
56
+ "codePath": "run_speech_recognition_ctc.py",
57
+ "git": {
58
+ "remote": "https://huggingface.co/AlexN/xls-r-300m-fr",
59
+ "commit": "9d07727def2251b0909520344156eb711ff096b9"
60
+ },
61
+ "email": "alexandrenap@hotmail.fr",
62
+ "root": "/workspace/xls-r-300m-fr",
63
+ "host": "job-1abccd0a-3293-4ffe-8274-9e8f841f653f",
64
+ "username": "ovh",
65
+ "executable": "/opt/conda/bin/python"
66
+ }
wandb/run-20220130_175417-3n44neww/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 21.5754, "train/learning_rate": 1.44e-06, "train/epoch": 1.25, "train/global_step": 20, "_runtime": 87, "_timestamp": 1643565344, "_step": 7, "eval/loss": 20.960968017578125, "eval/wer": 1.0, "eval/runtime": 2.6805, "eval/samples_per_second": 22.384, "eval/steps_per_second": 0.373, "_wandb": {"runtime": 95}}