ylacombe reach-vb HF staff commited on
Commit
bb8db30
β€’
1 Parent(s): 45f16f2

Update README.md (#4)

Browse files

- Update README.md (c329f74454be12bf7e475f5406b456f7a5855f96)


Co-authored-by: Vaibhav Srivastav <reach-vb@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -119,8 +119,8 @@ Whisper is a state-of-the-art model for automatic speech recognition (ASR) and s
119
  et al. from OpenAI. Trained on >5M hours of labeled data, Whisper demonstrates a strong ability to generalise to many
120
  datasets and domains in a zero-shot setting.
121
 
122
- Whisper large-v3-turbo is a distilled version of [Whisper large-v3](https://huggingface.co/openai/whisper-large-v3). In other words, it's the exact same model, except that the number of decoding layers have reduced from 32 to 4.
123
- As a result, the model is way faster, at the expense of a minor quality degradation.
124
 
125
  **Disclaimer**: Content for this model card has partly been written by the πŸ€— Hugging Face team, and partly copied and
126
  pasted from the original model card.
@@ -148,7 +148,7 @@ from datasets import load_dataset
148
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
149
  torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32
150
 
151
- model_id = "ylacombe/whisper-large-v3-turbo"
152
 
153
  model = AutoModelForSpeechSeq2Seq.from_pretrained(
154
  model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True, use_safetensors=True
@@ -252,7 +252,7 @@ from datasets import Audio, load_dataset
252
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
253
  torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32
254
 
255
- model_id = "ylacombe/whisper-large-v3-turbo"
256
 
257
  model = AutoModelForSpeechSeq2Seq.from_pretrained(
258
  model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True
@@ -327,7 +327,7 @@ from datasets import load_dataset
327
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
328
  torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32
329
 
330
- model_id = "ylacombe/whisper-large-v3-turbo"
331
 
332
  model = AutoModelForSpeechSeq2Seq.from_pretrained(
333
  model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True
@@ -373,7 +373,7 @@ torch.set_float32_matmul_precision("high")
373
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
374
  torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32
375
 
376
- model_id = "ylacombe/whisper-large-v3-turbo"
377
 
378
  model = AutoModelForSpeechSeq2Seq.from_pretrained(
379
  model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True
@@ -472,7 +472,7 @@ checkpoints are summarised in the following table with links to the models on th
472
  | large | 1550 M | x | [βœ“](https://huggingface.co/openai/whisper-large) |
473
  | large-v2 | 1550 M | x | [βœ“](https://huggingface.co/openai/whisper-large-v2) |
474
  | large-v3 | 1550 M | x | [βœ“](https://huggingface.co/openai/whisper-large-v3) |
475
- | large-v3-turbo | 809 M | x | [βœ“](https://huggingface.co/ylacombe/whisper-large-v3-turbo) |
476
 
477
 
478
  ## Fine-Tuning
 
119
  et al. from OpenAI. Trained on >5M hours of labeled data, Whisper demonstrates a strong ability to generalise to many
120
  datasets and domains in a zero-shot setting.
121
 
122
+ Whisper large-v3-turbo is a finetuned version of a pruned [Whisper large-v3](https://huggingface.co/openai/whisper-large-v3). In other words, it's the exact same model, except that the number of decoding layers have reduced from 32 to 4.
123
+ As a result, the model is way faster, at the expense of a minor quality degradation. You can find more details about it [in this GitHub discussion](https://github.com/openai/whisper/discussions/2363).
124
 
125
  **Disclaimer**: Content for this model card has partly been written by the πŸ€— Hugging Face team, and partly copied and
126
  pasted from the original model card.
 
148
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
149
  torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32
150
 
151
+ model_id = "openai/whisper-large-v3-turbo"
152
 
153
  model = AutoModelForSpeechSeq2Seq.from_pretrained(
154
  model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True, use_safetensors=True
 
252
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
253
  torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32
254
 
255
+ model_id = "openai/whisper-large-v3-turbo"
256
 
257
  model = AutoModelForSpeechSeq2Seq.from_pretrained(
258
  model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True
 
327
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
328
  torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32
329
 
330
+ model_id = "openai/whisper-large-v3-turbo"
331
 
332
  model = AutoModelForSpeechSeq2Seq.from_pretrained(
333
  model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True
 
373
  device = "cuda:0" if torch.cuda.is_available() else "cpu"
374
  torch_dtype = torch.float16 if torch.cuda.is_available() else torch.float32
375
 
376
+ model_id = "openai/whisper-large-v3-turbo"
377
 
378
  model = AutoModelForSpeechSeq2Seq.from_pretrained(
379
  model_id, torch_dtype=torch_dtype, low_cpu_mem_usage=True
 
472
  | large | 1550 M | x | [βœ“](https://huggingface.co/openai/whisper-large) |
473
  | large-v2 | 1550 M | x | [βœ“](https://huggingface.co/openai/whisper-large-v2) |
474
  | large-v3 | 1550 M | x | [βœ“](https://huggingface.co/openai/whisper-large-v3) |
475
+ | large-v3-turbo | 809 M | x | [βœ“](https://huggingface.co/openai/whisper-large-v3-turbo) |
476
 
477
 
478
  ## Fine-Tuning