Silemo commited on
Commit
d1cdd1b
·
1 Parent(s): b7e9aca

Training in progress, step 600, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12140b3527e8c387ab119cfaeb385ddf7ee08efb2029bcede8b49ffa881bc50e
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a9b7dc1434e889152ad78d4aa947ff37ca29820dea6192a46160f6ff22987a4
3
  size 966995080
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69d8260db719def1ed71c3ac4aef34515c81b7554c540c6bf09b8c6ef1774fbf
3
  size 1925064044
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99c47b2a3d9857753a01e989689bde14de9d5b794a10ad91a02ad70cbda789d7
3
  size 1925064044
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:309d78c0cbac356af547d46b1e70776079ec1d34f37e987fe5340ad766fae6b9
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:497714003ffbe7f7e7d7fe7d7b3ff01b83d9fe61fe7267edfc2a6269df2eb093
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09ff4109b85e31dcf014a99ae517a55071a006425bd4bf896b18ef8c2d279509
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:614396f462a6923cfc575e3f6427508fc269c766c509ab62ea3cb2154170e4f1
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 70.1955074875208,
3
  "best_model_checkpoint": "./whisper-it/checkpoint-300",
4
- "epoch": 0.7633587786259542,
5
  "eval_steps": 100,
6
- "global_step": 400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -139,6 +139,72 @@
139
  "eval_steps_per_second": 0.102,
140
  "eval_wer": 101.62922906267333,
141
  "step": 400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
142
  }
143
  ],
144
  "logging_steps": 25,
@@ -146,7 +212,7 @@
146
  "num_input_tokens_seen": 0,
147
  "num_train_epochs": 8,
148
  "save_steps": 100,
149
- "total_flos": 3.693893124096e+18,
150
  "trial_name": null,
151
  "trial_params": null
152
  }
 
1
  {
2
  "best_metric": 70.1955074875208,
3
  "best_model_checkpoint": "./whisper-it/checkpoint-300",
4
+ "epoch": 1.1450381679389312,
5
  "eval_steps": 100,
6
+ "global_step": 600,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
139
  "eval_steps_per_second": 0.102,
140
  "eval_wer": 101.62922906267333,
141
  "step": 400
142
+ },
143
+ {
144
+ "epoch": 0.81,
145
+ "learning_rate": 8.380000000000001e-06,
146
+ "loss": 0.2514,
147
+ "step": 425
148
+ },
149
+ {
150
+ "epoch": 0.86,
151
+ "learning_rate": 8.880000000000001e-06,
152
+ "loss": 0.2826,
153
+ "step": 450
154
+ },
155
+ {
156
+ "epoch": 0.91,
157
+ "learning_rate": 9.38e-06,
158
+ "loss": 0.2381,
159
+ "step": 475
160
+ },
161
+ {
162
+ "epoch": 0.95,
163
+ "learning_rate": 9.88e-06,
164
+ "loss": 0.2491,
165
+ "step": 500
166
+ },
167
+ {
168
+ "epoch": 0.95,
169
+ "eval_loss": 0.3922654092311859,
170
+ "eval_runtime": 1735.7941,
171
+ "eval_samples_per_second": 0.864,
172
+ "eval_steps_per_second": 0.108,
173
+ "eval_wer": 112.23655019412091,
174
+ "step": 500
175
+ },
176
+ {
177
+ "epoch": 1.0,
178
+ "learning_rate": 9.945714285714286e-06,
179
+ "loss": 0.2621,
180
+ "step": 525
181
+ },
182
+ {
183
+ "epoch": 1.05,
184
+ "learning_rate": 9.874285714285715e-06,
185
+ "loss": 0.1729,
186
+ "step": 550
187
+ },
188
+ {
189
+ "epoch": 1.1,
190
+ "learning_rate": 9.802857142857144e-06,
191
+ "loss": 0.1725,
192
+ "step": 575
193
+ },
194
+ {
195
+ "epoch": 1.15,
196
+ "learning_rate": 9.731428571428573e-06,
197
+ "loss": 0.17,
198
+ "step": 600
199
+ },
200
+ {
201
+ "epoch": 1.15,
202
+ "eval_loss": 0.3860141336917877,
203
+ "eval_runtime": 1739.7902,
204
+ "eval_samples_per_second": 0.862,
205
+ "eval_steps_per_second": 0.108,
206
+ "eval_wer": 107.05074875207987,
207
+ "step": 600
208
  }
209
  ],
210
  "logging_steps": 25,
 
212
  "num_input_tokens_seen": 0,
213
  "num_train_epochs": 8,
214
  "save_steps": 100,
215
+ "total_flos": 5.53881958834176e+18,
216
  "trial_name": null,
217
  "trial_params": null
218
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6afd188b1bb5040b1e1647512623a4ea330124b3938572fb5350dd1ea4ab41d
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8639ff321d8732686fcd47480f025401063d95f80a6017764993068cc24ca794
3
  size 4792