--- library_name: transformers --- # finetuned whisper-tiny model on custom dataset This model is a fine-tuned version of `openai/whisper-tiny` on Serbian Mozilla/Common Voice 13. It achieves the following results on the evaluation set: - **Loss**: 0.1628 - **Wer Ortho**: 0.1635 - **Wer**: 0.0556 ## Training Procedure ### Training Hyperparameters The following hyperparameters were used during training: - **learning_rate**: 3e-5 - **train_batch_size**: 32 - **eval_batch_size**: 32 - **gradient_accumulation_steps**: 2 - **total_train_batch_size**: 64 - **optimizer**: Adam with betas=(0.9,0.999) and epsilon=1e-08 - **lr_scheduler_type**: linear - **lr_scheduler_warmup_steps**: 100 - **training_steps**: 2000 - **mixed_precision_training**: Native AMP ### Training Results | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer | |---------------|-------|------|-----------------|-----------|-------| | 0.0600 | 1.34 | 500 | 0.1852 | 0.1800 | 0.0745| | 0.0285 | 2.67 | 1000 | 0.1715 | 0.1710 | 0.0640| | 0.0140 | 4.01 | 1500 | 0.1658 | 0.1685 | 0.0582| ## Framework Versions - **Transformers**: 4.41.2 - **Pytorch**: 2.3.0+cu121 - **Datasets**: 2.18.0 - **Tokenizers**: 0.19.1