PereLluis13 commited on
Commit
0c17f2b
1 Parent(s): f4231fb

update epoch 6

Browse files
README.md ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ca
4
+ license: apache-2.0
5
+ tags:
6
+ - automatic-speech-recognition
7
+ - mozilla-foundation/common_voice_8_0
8
+ - collectivat/tv3_parla
9
+ - projecte-aina/parlament_parla
10
+ - generated_from_trainer
11
+ model-index:
12
+ - name: wav2vec2-xls-r-300m-ca
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # wav2vec2-xls-r-300m-ca
20
+
21
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - CA dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 0.2758
24
+ - Wer: 0.1792
25
+
26
+ ## Model description
27
+
28
+ More information needed
29
+
30
+ ## Intended uses & limitations
31
+
32
+ More information needed
33
+
34
+ ## Training and evaluation data
35
+
36
+ More information needed
37
+
38
+ ## Training procedure
39
+
40
+ ### Training hyperparameters
41
+
42
+ The following hyperparameters were used during training:
43
+ - learning_rate: 7.5e-05
44
+ - train_batch_size: 32
45
+ - eval_batch_size: 32
46
+ - seed: 42
47
+ - gradient_accumulation_steps: 4
48
+ - total_train_batch_size: 128
49
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
+ - lr_scheduler_type: linear
51
+ - lr_scheduler_warmup_steps: 2000
52
+ - num_epochs: 6.0
53
+ - mixed_precision_training: Native AMP
54
+
55
+ ### Training results
56
+
57
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
58
+ |:-------------:|:-----:|:-----:|:---------------:|:------:|
59
+ | 6.2099 | 0.09 | 500 | 3.4125 | 1.0 |
60
+ | 2.9961 | 0.18 | 1000 | 2.9224 | 1.0 |
61
+ | 2.2147 | 0.26 | 1500 | 0.6521 | 0.5568 |
62
+ | 1.3017 | 0.35 | 2000 | 0.3153 | 0.2761 |
63
+ | 1.1196 | 0.44 | 2500 | 0.2444 | 0.2367 |
64
+ | 1.0712 | 0.53 | 3000 | 0.2324 | 0.2132 |
65
+ | 1.052 | 0.62 | 3500 | 0.2173 | 0.2032 |
66
+ | 1.2813 | 2.13 | 4000 | 0.3326 | 0.2099 |
67
+ | 1.2365 | 2.4 | 4500 | 0.3224 | 0.2003 |
68
+ | 1.2193 | 2.66 | 5000 | 0.3198 | 0.1957 |
69
+ | 1.2072 | 2.93 | 5500 | 0.3063 | 0.1933 |
70
+ | 1.213 | 3.2 | 6000 | 0.3051 | 0.1980 |
71
+ | 1.2074 | 3.46 | 6500 | 0.3012 | 0.1879 |
72
+ | 1.1918 | 3.73 | 7000 | 0.2947 | 0.1829 |
73
+ | 1.1893 | 4.0 | 7500 | 0.2895 | 0.1807 |
74
+ | 1.1751 | 4.26 | 8000 | 0.2878 | 0.1776 |
75
+ | 1.1628 | 4.53 | 8500 | 0.2835 | 0.1731 |
76
+ | 1.1577 | 4.79 | 9000 | 0.2816 | 0.1761 |
77
+ | 1.1448 | 5.06 | 9500 | 0.2757 | 0.1740 |
78
+ | 1.1407 | 5.33 | 10000 | 0.2768 | 0.1798 |
79
+ | 1.1401 | 5.59 | 10500 | 0.2780 | 0.1816 |
80
+ | 1.1333 | 5.86 | 11000 | 0.2748 | 0.1750 |
81
+
82
+
83
+ ### Framework versions
84
+
85
+ - Transformers 4.16.0.dev0
86
+ - Pytorch 1.10.1+cu102
87
+ - Datasets 1.18.1
88
+ - Tokenizers 0.11.0
all_results.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "epoch": 3.0,
3
- "eval_loss": 0.3033297657966614,
4
- "eval_runtime": 395.7148,
5
  "eval_samples": 4297,
6
- "eval_samples_per_second": 10.859,
7
- "eval_steps_per_second": 0.341,
8
- "eval_wer": 0.19506011201512932,
9
- "train_loss": 0.46717894900821066,
10
- "train_runtime": 42700.8803,
11
  "train_samples": 240334,
12
- "train_samples_per_second": 16.885,
13
- "train_steps_per_second": 0.132
14
  }
1
  {
2
+ "epoch": 6.0,
3
+ "eval_loss": 0.27584555745124817,
4
+ "eval_runtime": 371.4251,
5
  "eval_samples": 4297,
6
+ "eval_samples_per_second": 11.569,
7
+ "eval_steps_per_second": 0.363,
8
+ "eval_wer": 0.17918264285046293,
9
+ "train_loss": 0.5975231319014925,
10
+ "train_runtime": 88923.2167,
11
  "train_samples": 240334,
12
+ "train_samples_per_second": 16.216,
13
+ "train_steps_per_second": 0.127
14
  }
eval_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 3.0,
3
- "eval_loss": 0.3033297657966614,
4
- "eval_runtime": 395.7148,
5
  "eval_samples": 4297,
6
- "eval_samples_per_second": 10.859,
7
- "eval_steps_per_second": 0.341,
8
- "eval_wer": 0.19506011201512932
9
  }
1
  {
2
+ "epoch": 6.0,
3
+ "eval_loss": 0.27584555745124817,
4
+ "eval_runtime": 371.4251,
5
  "eval_samples": 4297,
6
+ "eval_samples_per_second": 11.569,
7
+ "eval_steps_per_second": 0.363,
8
+ "eval_wer": 0.17918264285046293
9
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69b8c3e5a25929fe5518b2db2fc0b10d4c584f3b4511f4b9ec2bc19eb7f1d4a4
3
  size 1262112241
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f89ec5de9dd87da4f2a4c2382f07f15a131a2b2c9de36a7ae27f30c07f7e2540
3
  size 1262112241
runs/Jan29_11-55-18_job-336a688f-553a-4e6e-83b3-ad5d10274b51/events.out.tfevents.1643460077.job-336a688f-553a-4e6e-83b3-ad5d10274b51.1539029.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7b03f34331375aed60c773aca2e4e4c67953313085ef60c3bd9c0aba81ffc68
3
- size 5220
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1247dfe6bd0d58cb74bf4ef78970d1f112256fcf3df6fe455135e580b51f0fe3
3
+ size 10324
runs/Jan29_11-55-18_job-336a688f-553a-4e6e-83b3-ad5d10274b51/events.out.tfevents.1643549374.job-336a688f-553a-4e6e-83b3-ad5d10274b51.1539029.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ea16ea09e4540e71e147fec1294192500786f27cab5e5cc71cbb8e3a338092a
3
+ size 358
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 3.0,
3
- "train_loss": 0.46717894900821066,
4
- "train_runtime": 42700.8803,
5
  "train_samples": 240334,
6
- "train_samples_per_second": 16.885,
7
- "train_steps_per_second": 0.132
8
  }
1
  {
2
+ "epoch": 6.0,
3
+ "train_loss": 0.5975231319014925,
4
+ "train_runtime": 88923.2167,
5
  "train_samples": 240334,
6
+ "train_samples_per_second": 16.216,
7
+ "train_steps_per_second": 0.127
8
  }
trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 2.9996005858074826,
5
- "global_step": 5631,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -173,18 +173,183 @@
173
  "step": 5500
174
  },
175
  {
176
- "epoch": 3.0,
177
- "step": 5631,
178
- "total_flos": 1.4549289807228204e+20,
179
- "train_loss": 0.46717894900821066,
180
- "train_runtime": 42700.8803,
181
- "train_samples_per_second": 16.885,
182
- "train_steps_per_second": 0.132
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
183
  }
184
  ],
185
- "max_steps": 5631,
186
- "num_train_epochs": 3,
187
- "total_flos": 1.4549289807228204e+20,
188
  "trial_name": null,
189
  "trial_params": null
190
  }
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 5.999600585807483,
5
+ "global_step": 11262,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
173
  "step": 5500
174
  },
175
  {
176
+ "epoch": 3.2,
177
+ "learning_rate": 4.263388037141005e-05,
178
+ "loss": 1.213,
179
+ "step": 6000
180
+ },
181
+ {
182
+ "epoch": 3.2,
183
+ "eval_loss": 0.3051116168498993,
184
+ "eval_runtime": 371.5477,
185
+ "eval_samples_per_second": 11.565,
186
+ "eval_steps_per_second": 0.363,
187
+ "eval_wer": 0.19797998690732255,
188
+ "step": 6000
189
+ },
190
+ {
191
+ "epoch": 3.46,
192
+ "learning_rate": 3.8585078816670256e-05,
193
+ "loss": 1.2074,
194
+ "step": 6500
195
+ },
196
+ {
197
+ "epoch": 3.46,
198
+ "eval_loss": 0.3011966645717621,
199
+ "eval_runtime": 372.0587,
200
+ "eval_samples_per_second": 11.549,
201
+ "eval_steps_per_second": 0.363,
202
+ "eval_wer": 0.18792148549934018,
203
+ "step": 6500
204
+ },
205
+ {
206
+ "epoch": 3.73,
207
+ "learning_rate": 3.4536277261930465e-05,
208
+ "loss": 1.1918,
209
+ "step": 7000
210
+ },
211
+ {
212
+ "epoch": 3.73,
213
+ "eval_loss": 0.29471954703330994,
214
+ "eval_runtime": 372.8767,
215
+ "eval_samples_per_second": 11.524,
216
+ "eval_steps_per_second": 0.362,
217
+ "eval_wer": 0.18287145276764655,
218
+ "step": 7000
219
+ },
220
+ {
221
+ "epoch": 4.0,
222
+ "learning_rate": 3.0487475707190668e-05,
223
+ "loss": 1.1893,
224
+ "step": 7500
225
+ },
226
+ {
227
+ "epoch": 4.0,
228
+ "eval_loss": 0.28953900933265686,
229
+ "eval_runtime": 372.4558,
230
+ "eval_samples_per_second": 11.537,
231
+ "eval_steps_per_second": 0.362,
232
+ "eval_wer": 0.1807205129004437,
233
+ "step": 7500
234
+ },
235
+ {
236
+ "epoch": 4.26,
237
+ "learning_rate": 2.643867415245087e-05,
238
+ "loss": 1.1751,
239
+ "step": 8000
240
+ },
241
+ {
242
+ "epoch": 4.26,
243
+ "eval_loss": 0.2877567410469055,
244
+ "eval_runtime": 371.9244,
245
+ "eval_samples_per_second": 11.553,
246
+ "eval_steps_per_second": 0.363,
247
+ "eval_wer": 0.17759281773122604,
248
+ "step": 8000
249
+ },
250
+ {
251
+ "epoch": 4.53,
252
+ "learning_rate": 2.2389872597711073e-05,
253
+ "loss": 1.1628,
254
+ "step": 8500
255
+ },
256
+ {
257
+ "epoch": 4.53,
258
+ "eval_loss": 0.28352275490760803,
259
+ "eval_runtime": 373.4388,
260
+ "eval_samples_per_second": 11.507,
261
+ "eval_steps_per_second": 0.362,
262
+ "eval_wer": 0.1730727267059447,
263
+ "step": 8500
264
+ },
265
+ {
266
+ "epoch": 4.79,
267
+ "learning_rate": 1.834107104297128e-05,
268
+ "loss": 1.1577,
269
+ "step": 9000
270
+ },
271
+ {
272
+ "epoch": 4.79,
273
+ "eval_loss": 0.28156954050064087,
274
+ "eval_runtime": 371.5997,
275
+ "eval_samples_per_second": 11.564,
276
+ "eval_steps_per_second": 0.363,
277
+ "eval_wer": 0.17605494768124527,
278
+ "step": 9000
279
+ },
280
+ {
281
+ "epoch": 5.06,
282
+ "learning_rate": 1.4300367091340963e-05,
283
+ "loss": 1.1448,
284
+ "step": 9500
285
+ },
286
+ {
287
+ "epoch": 5.06,
288
+ "eval_loss": 0.2757455110549927,
289
+ "eval_runtime": 373.7148,
290
+ "eval_samples_per_second": 11.498,
291
+ "eval_steps_per_second": 0.361,
292
+ "eval_wer": 0.17397674491100096,
293
+ "step": 9500
294
+ },
295
+ {
296
+ "epoch": 5.33,
297
+ "learning_rate": 1.0251565536601165e-05,
298
+ "loss": 1.1407,
299
+ "step": 10000
300
+ },
301
+ {
302
+ "epoch": 5.33,
303
+ "eval_loss": 0.2768358588218689,
304
+ "eval_runtime": 373.2273,
305
+ "eval_samples_per_second": 11.513,
306
+ "eval_steps_per_second": 0.362,
307
+ "eval_wer": 0.17976453962613131,
308
+ "step": 10000
309
+ },
310
+ {
311
+ "epoch": 5.59,
312
+ "learning_rate": 6.202763981861369e-06,
313
+ "loss": 1.1401,
314
+ "step": 10500
315
+ },
316
+ {
317
+ "epoch": 5.59,
318
+ "eval_loss": 0.2780420780181885,
319
+ "eval_runtime": 375.3753,
320
+ "eval_samples_per_second": 11.447,
321
+ "eval_steps_per_second": 0.36,
322
+ "eval_wer": 0.18160374907779753,
323
+ "step": 10500
324
+ },
325
+ {
326
+ "epoch": 5.86,
327
+ "learning_rate": 2.153962427121572e-06,
328
+ "loss": 1.1333,
329
+ "step": 11000
330
+ },
331
+ {
332
+ "epoch": 5.86,
333
+ "eval_loss": 0.2748132646083832,
334
+ "eval_runtime": 371.6352,
335
+ "eval_samples_per_second": 11.562,
336
+ "eval_steps_per_second": 0.363,
337
+ "eval_wer": 0.17495350021301578,
338
+ "step": 11000
339
+ },
340
+ {
341
+ "epoch": 6.0,
342
+ "step": 11262,
343
+ "total_flos": 3.063803468240281e+20,
344
+ "train_loss": 0.5975231319014925,
345
+ "train_runtime": 88923.2167,
346
+ "train_samples_per_second": 16.216,
347
+ "train_steps_per_second": 0.127
348
  }
349
  ],
350
+ "max_steps": 11262,
351
+ "num_train_epochs": 6,
352
+ "total_flos": 3.063803468240281e+20,
353
  "trial_name": null,
354
  "trial_params": null
355
  }