sanchit-gandhi HF staff commited on
Commit
c7bea3d
1 Parent(s): 796a56b
Files changed (1) hide show
  1. README.md +38 -12
README.md CHANGED
@@ -229,18 +229,6 @@ model = AutoModelForSpeechSeq2Seq.from_pretrained(model_id, torch_dtype=torch_dt
229
  + model = model.to_bettertransformer()
230
  ```
231
 
232
- ### 8bit & 4bit Quantization
233
-
234
- Coming soon ...
235
-
236
- ### Candle
237
-
238
- Coming soon ...
239
-
240
- ### Whisper.cpp
241
-
242
- Coming soon ...
243
-
244
  ### Running Whisper in `openai-whisper`
245
 
246
  To use the model in the original Whisper format, first ensure you have the [`openai-whisper`](https://pypi.org/project/openai-whisper/) package installed:
@@ -290,6 +278,44 @@ let output = await transcriber(url);
290
  See the [docs](https://huggingface.co/docs/transformers.js/api/pipelines#module_pipelines.AutomaticSpeechRecognitionPipeline) for more information.
291
 
292
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
293
  ## Model Details
294
 
295
  Distil-Whisper inherits the encoder-decoder architecture from Whisper. The encoder maps a sequence of speech vector
 
229
  + model = model.to_bettertransformer()
230
  ```
231
 
 
 
 
 
 
 
 
 
 
 
 
 
232
  ### Running Whisper in `openai-whisper`
233
 
234
  To use the model in the original Whisper format, first ensure you have the [`openai-whisper`](https://pypi.org/project/openai-whisper/) package installed:
 
278
  See the [docs](https://huggingface.co/docs/transformers.js/api/pipelines#module_pipelines.AutomaticSpeechRecognitionPipeline) for more information.
279
 
280
 
281
+
282
+ ### Candle
283
+
284
+ Through an integration with Hugging Face [Candle](https://github.com/huggingface/candle/tree/main) 🕯️, Distil-Whisper is
285
+ now available in the Rust library 🦀
286
+
287
+ Benefit from:
288
+ * Optimised CPU backend with optional MKL support for x86 and Accelerate for Macs
289
+ * CUDA backend for efficiently running on GPUs, multiple GPU distribution via NCCL
290
+ * WASM support: run Distil-Whisper in a browser
291
+
292
+ Steps for getting started:
293
+ 1. Install [`candle-core`](https://github.com/huggingface/candle/tree/main/candle-core) as explained [here](https://huggingface.github.io/candle/guide/installation.html)
294
+ 2. Clone the `candle` repository locally:
295
+ ```
296
+ git clone https://github.com/huggingface/candle.git
297
+ ```
298
+ 3. Enter the example directory for [Whisper](https://github.com/huggingface/candle/tree/main/candle-examples/examples/whisper):
299
+ ```
300
+ cd candle/candle-examples/examples/whisper
301
+ ```
302
+ 4. Run an example:
303
+ ```
304
+ cargo run --example whisper --release -- --model distil-medium.en
305
+ ```
306
+ 5. To specify your own audio file, add the `--input` flag:
307
+ ```
308
+ cargo run --example whisper --release -- --model distil-medium.en --input audio.wav
309
+ ```
310
+
311
+ ### 8bit & 4bit Quantization
312
+
313
+ Coming soon ...
314
+
315
+ ### Whisper.cpp
316
+
317
+ Coming soon ...
318
+
319
  ## Model Details
320
 
321
  Distil-Whisper inherits the encoder-decoder architecture from Whisper. The encoder maps a sequence of speech vector