libri-smallw2v2-no-copy-mse-alpha-0.75-T-1
This model is a fine-tuned version of on the None dataset. It achieves the following results on the evaluation set:
- Loss: 299.5353
- Wer: 0.5607
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0002
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.2
- num_epochs: 100
- mixed_precision_training: Native AMP
Training results
Training Loss | Epoch | Step | Validation Loss | Wer |
---|---|---|---|---|
2621.6909 | 1.79 | 400 | 892.3344 | 1.0000 |
1182.3671 | 3.59 | 800 | 583.0377 | 0.9766 |
946.8591 | 5.38 | 1200 | 493.0189 | 0.9365 |
833.3073 | 7.17 | 1600 | 444.2031 | 0.9141 |
751.816 | 8.97 | 2000 | 400.5189 | 0.8937 |
698.0563 | 10.76 | 2400 | 369.2302 | 0.8649 |
644.5172 | 12.56 | 2800 | 345.4725 | 0.8458 |
600.5121 | 14.35 | 3200 | 323.2295 | 0.8151 |
561.2316 | 16.14 | 3600 | 307.8979 | 0.7950 |
523.6023 | 17.94 | 4000 | 291.1282 | 0.7720 |
496.2595 | 19.73 | 4400 | 279.5770 | 0.7517 |
466.5839 | 21.52 | 4800 | 267.7692 | 0.7263 |
440.6372 | 23.32 | 5200 | 258.0482 | 0.7028 |
417.6962 | 25.11 | 5600 | 254.2653 | 0.6902 |
392.6432 | 26.91 | 6000 | 252.7504 | 0.6774 |
378.9874 | 28.7 | 6400 | 246.1504 | 0.6714 |
361.5383 | 30.49 | 6800 | 243.2339 | 0.6630 |
345.9527 | 32.29 | 7200 | 242.4769 | 0.6494 |
335.28 | 34.08 | 7600 | 243.5441 | 0.6448 |
320.2823 | 35.87 | 8000 | 242.6982 | 0.6353 |
306.8673 | 37.67 | 8400 | 252.8057 | 0.6280 |
300.3173 | 39.46 | 8800 | 249.5198 | 0.6267 |
290.0972 | 41.26 | 9200 | 243.8252 | 0.6218 |
283.0466 | 43.05 | 9600 | 242.9062 | 0.6173 |
273.4801 | 44.84 | 10000 | 249.2953 | 0.6121 |
264.7652 | 46.64 | 10400 | 251.2528 | 0.6127 |
257.2499 | 48.43 | 10800 | 257.1318 | 0.6089 |
250.4637 | 50.22 | 11200 | 264.7531 | 0.6069 |
245.2013 | 52.02 | 11600 | 258.1169 | 0.6000 |
240.3053 | 53.81 | 12000 | 255.6198 | 0.5941 |
232.0262 | 55.61 | 12400 | 261.2134 | 0.5949 |
228.3372 | 57.4 | 12800 | 263.1902 | 0.5919 |
222.2224 | 59.19 | 13200 | 271.1831 | 0.5898 |
218.0379 | 60.99 | 13600 | 264.8538 | 0.5821 |
215.7583 | 62.78 | 14000 | 268.3278 | 0.5874 |
210.5649 | 64.57 | 14400 | 273.0802 | 0.5808 |
205.7684 | 66.37 | 14800 | 273.7237 | 0.5792 |
202.7417 | 68.16 | 15200 | 276.0477 | 0.5750 |
197.5293 | 69.96 | 15600 | 272.9896 | 0.5785 |
196.512 | 71.75 | 16000 | 274.7004 | 0.5777 |
193.9167 | 73.54 | 16400 | 275.8243 | 0.5727 |
190.2722 | 75.34 | 16800 | 283.2336 | 0.5743 |
187.5092 | 77.13 | 17200 | 284.5899 | 0.5723 |
182.8259 | 78.92 | 17600 | 287.1284 | 0.5732 |
184.1322 | 80.72 | 18000 | 285.4507 | 0.5704 |
180.0438 | 82.51 | 18400 | 283.7040 | 0.5684 |
179.6387 | 84.3 | 18800 | 288.0453 | 0.5670 |
176.7927 | 86.1 | 19200 | 290.4306 | 0.5662 |
173.8417 | 87.89 | 19600 | 298.6391 | 0.5663 |
174.4995 | 89.69 | 20000 | 299.5260 | 0.5634 |
171.9799 | 91.48 | 20400 | 301.1923 | 0.5653 |
170.4011 | 93.27 | 20800 | 299.7171 | 0.5667 |
170.4423 | 95.07 | 21200 | 296.5629 | 0.5639 |
169.7396 | 96.86 | 21600 | 299.2657 | 0.5622 |
166.9574 | 98.65 | 22000 | 299.5353 | 0.5607 |
Framework versions
- Transformers 4.25.1
- Pytorch 1.12.1
- Datasets 2.8.0
- Tokenizers 0.13.2
- Downloads last month
- 7