whisper-small-en / README.md
BlueRaccoon's picture
Update metadata with huggingface_hub
9c74d46
metadata
license: apache-2.0
tags:
  - whisper-event
  - generated_from_trainer
datasets:
  - google/fleurs
metrics:
  - wer
model-index:
  - name: Whisper Small English
    results:
      - task:
          type: automatic-speech-recognition
          name: Automatic Speech Recognition
        dataset:
          name: google/fleurs en_us
          type: google/fleurs
          config: en_us
          split: test
          args: en_us
        metrics:
          - type: wer
            value: 7.990755655157924
            name: Wer
      - task:
          type: automatic-speech-recognition
          name: Automatic Speech Recognition
        dataset:
          name: mozilla-foundation/common_voice_11_0
          type: mozilla-foundation/common_voice_11_0
          config: en
          split: test
        metrics:
          - type: wer
            value: 18.21
            name: WER

Whisper Small English

This model is a fine-tuned version of openai/whisper-small.en on the google/fleurs en_us dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6007
  • Wer: 7.9908

Model description

This model was created as part of the Whisper Fine-Tune Event. This is my first attempt at fine-tuning the Whisper neural network. Honestly, it's my second time ever trying anything related to training a neural network, and my first time was pretty bad (but I did get a lot of rather funny images out of it, so perhaps it wasn't entirely fruitless?), and it seems like the WER only went up after step 2000, so... I'm not sure if I did a good job or if I just wasted GPU cycles, but maybe I can try again and get a better score?

I'm learning.

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 64
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 500
  • training_steps: 5000
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Wer
0.0005 24.0 1000 0.5092 7.5566
0.0002 48.01 2000 0.5528 7.7526
0.0001 73.0 3000 0.5785 7.8507
0.0001 97.0 4000 0.5936 7.9908
0.0001 121.01 5000 0.6007 7.9908

Framework versions

  • Transformers 4.26.0.dev0
  • Pytorch 1.13.0+cu117
  • Datasets 2.7.1.dev0
  • Tokenizers 0.13.2