Edit model card

Whisper Small - Swedish

This is a fine-tuned version of the openai/whisper-small model on the Common Voice 11.0 dataset.

The following results were achieved after training for 4000 optimization steps:

  • Training Loss: 0.003900
  • Validation Loss: 0.326255
  • WER: 19.894598

Training hyperparameters

The following hyperparameters were used during training:

  • train_batch_size: 16
  • gradient_accumulation_steps: 1
  • learning_rate: 1e-5
  • eval_batch_size: 8
  • max_steps: 4000
  • eval_steps: 1000

Framework Versions

  • Transformers 4.25.0
  • Pytorch 1.12.1
  • Datasets 2.7.1
Downloads last month
5

Spaces using khalidey/ID2223_Lab2_Whisper_SV 2