asahi417 commited on
Commit
25dfa95
1 Parent(s): af651af

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -6
README.md CHANGED
@@ -48,17 +48,19 @@ We measure the inference speed of different kotoba-whisper-v1.0 implementations
48
  - OS Sonoma Version 14.4.1 (23E224)
49
 
50
  | audio file | audio duration (min)| [whisper.cpp](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml) (sec) | [faster-whisper](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-faster) (sec)| [hf pipeline](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0) (sec)
51
- |--------|------|-----|------|---|
52
- |audio 1 | 50.3 | 581 | 2601 | |
53
- |audio 2 | 5.6 | 41 | 73 | 61 |
54
- |audio 3 | 4.9 | 30 | 141 | 54 |
55
- |audio 4 | 5.6 | 35 | 126 | 69 |
56
 
57
  Scripts to re-run the experiment can be found bellow:
58
  * [whisper.cpp](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml/blob/main/benchmark.sh)
59
  * [faster-whisper](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-faster/blob/main/benchmark.sh)
60
  * [hf pipeline](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0/blob/main/benchmark.sh)
61
-
 
 
62
 
63
  ### Quantized Model
64
  To use the quantized model, download the quantized GGML weights:
 
48
  - OS Sonoma Version 14.4.1 (23E224)
49
 
50
  | audio file | audio duration (min)| [whisper.cpp](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml) (sec) | [faster-whisper](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-faster) (sec)| [hf pipeline](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0) (sec)
51
+ |--------|------|-----|------|-----|
52
+ |audio 1 | 50.3 | 581 | 2601 | 807 |
53
+ |audio 2 | 5.6 | 41 | 73 | 61 |
54
+ |audio 3 | 4.9 | 30 | 141 | 54 |
55
+ |audio 4 | 5.6 | 35 | 126 | 69 |
56
 
57
  Scripts to re-run the experiment can be found bellow:
58
  * [whisper.cpp](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-ggml/blob/main/benchmark.sh)
59
  * [faster-whisper](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-faster/blob/main/benchmark.sh)
60
  * [hf pipeline](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0/blob/main/benchmark.sh)
61
+ Also, currently whisper.cpp and faster-whisper support the [sequential long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#sequential-long-form),
62
+ and only Huggingface pipeline supports the [chunked long-form decoding](https://huggingface.co/distil-whisper/distil-large-v3#chunked-long-form), which we empirically
63
+ found better than the sequnential long-form decoding.
64
 
65
  ### Quantized Model
66
  To use the quantized model, download the quantized GGML weights: