--- license: apache-2.0 tags: - generated_from_trainer metrics: - wer model-index: - name: whisper-base-nl-3 results: [] --- # whisper-base-nl-3 This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.5826 - Wer: 24.8548 - Cer: 8.2429 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 2 - eval_batch_size: 3 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 45000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Cer | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:-------:|:---------------:|:-------:| | 0.5761 | 0.06 | 1000 | 10.1154 | 0.5675 | 28.1532 | | 0.48 | 0.13 | 2000 | 9.6911 | 0.5239 | 26.4364 | | 0.4094 | 0.19 | 3000 | 9.1532 | 0.4925 | 24.8355 | | 0.4792 | 0.26 | 4000 | 8.8414 | 0.4702 | 24.1105 | | 0.3444 | 0.32 | 5000 | 8.8531 | 0.4544 | 23.9017 | | 0.3943 | 0.39 | 6000 | 8.3602 | 0.4446 | 22.7353 | | 0.4925 | 0.45 | 7000 | 8.3724 | 0.4348 | 22.1788 | | 0.4455 | 0.52 | 8000 | 8.2989 | 0.4270 | 21.7549 | | 0.3987 | 0.58 | 9000 | 7.9417 | 0.4139 | 20.8424 | | 0.3373 | 0.65 | 10000 | 7.8871 | 0.4116 | 21.2144 | | 0.3808 | 0.71 | 11000 | 7.6264 | 0.4016 | 20.5092 | | 0.4214 | 0.78 | 12000 | 7.4153 | 0.3949 | 20.0938 | | 0.3029 | 0.84 | 13000 | 7.3581 | 0.3902 | 19.7347 | | 0.3549 | 1.66 | 14000 | 7.1195 | 0.3908 | 19.4115 | | 0.3385 | 1.78 | 15000 | 7.7792 | 0.3906 | 20.2051 | | 0.3282 | 1.9 | 16000 | 7.1081 | 0.3923 | 19.2651 | | 0.3196 | 2.02 | 17000 | 7.2249 | 0.3923 | 19.3352 | | 0.3251 | 2.14 | 18000 | 7.1761 | 0.3981 | 19.4831 | | 0.4162 | 2.25 | 19000 | 7.0590 | 0.3958 | 19.0577 | | 0.2851 | 2.37 | 20000 | 7.0167 | 0.3953 | 19.2095 | | 0.2982 | 2.49 | 21000 | 6.8426 | 0.3929 | 18.8100 | | 0.3642 | 2.61 | 22000 | 6.8867 | 0.3954 | 18.6972 | | 0.2297 | 2.73 | 23000 | 6.9384 | 0.3916 | 18.7330 | | 0.2313 | 2.85 | 24000 | 6.7785 | 0.3930 | 18.6034 | | 0.2833 | 2.97 | 25000 | 6.8552 | 0.3910 | 18.5981 | | 0.2509 | 3.09 | 26000 | 6.8165 | 0.3949 | 18.5180 | | 0.2085 | 3.2 | 27000 | 6.8113 | 0.3985 | 18.6133 | | 0.2055 | 3.32 | 28000 | 6.8624 | 0.3995 | 18.7612 | | 0.175 | 3.44 | 29000 | 6.7727 | 0.4009 | 18.4814 | | 0.1701 | 3.56 | 30000 | 7.0136 | 0.3998 | 18.8344 | | 0.6832 | 33.81 | 31000 | 7.8509 | 0.5425 | 24.6216 | | 0.5676 | 34.9 | 32000 | 7.3776 | 0.5141 | 23.6790 | | 0.4863 | 35.99 | 33000 | 7.2441 | 0.5003 | 23.0542 | | 0.5007 | 37.08 | 34000 | 7.1545 | 0.4948 | 22.9234 | | 0.4519 | 38.17 | 35000 | 7.1257 | 0.4922 | 22.8248 | | 0.3674 | 39.26 | 36000 | 7.0104 | 0.4754 | 22.6642 | | 0.3481 | 40.35 | 37000 | 7.0311 | 0.4679 | 22.6314 | | 0.2992 | 41.44 | 38000 | 6.9465 | 0.4622 | 22.2595 | | 0.2505 | 42.53 | 39000 | 6.9198 | 0.4641 | 22.1937 | | 0.2477 | 43.62 | 40000 | 7.2008 | 0.4678 | 22.8279 | | 0.1994 | 44.71 | 41000 | 7.1179 | 0.4689 | 22.3808 | | 0.1865 | 45.8 | 42000 | 7.1351 | 0.4717 | 22.5664 | | 0.2307 | 46.89 | 43000 | 7.1364 | 0.4754 | 22.3722 | | 0.1705 | 47.98 | 44000 | 7.0830 | 0.4759 | 22.3863 | | 0.2007 | 49.07 | 45000 | 7.1187 | 0.4767 | 22.4849 | ### Framework versions - Transformers 4.26.0.dev0 - Pytorch 1.13.1+cu117 - Datasets 2.8.1.dev0 - Tokenizers 0.13.2