asahi417 commited on
Commit
9f2df35
1 Parent(s): f969f24

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +52 -0
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: ja
3
+ tags:
4
+ - audio
5
+ - automatic-speech-recognition
6
+ license: mit
7
+ library_name: ctranslate2
8
+ ---
9
+
10
+ # Whisper kotoba-whisper-v2.0 model for CTranslate2
11
+
12
+ This repository contains the conversion of [kotoba-tech/kotoba-whisper-v2.0](https://huggingface.co/kotoba-tech/kotoba-whisper-v2.0) to the [CTranslate2](https://github.com/OpenNMT/CTranslate2) model format.
13
+
14
+ This model can be used in CTranslate2 or projects based on CTranslate2 such as [faster-whisper](https://github.com/systran/faster-whisper).
15
+
16
+ ## Example
17
+ Install library and download sample audio.
18
+ ```shell
19
+ pip install faster-whisper
20
+ wget https://huggingface.co/kotoba-tech/kotoba-whisper-v2.0-ggml/resolve/main/sample_ja_speech.wav
21
+ ```
22
+ Inference with the kotoba-whisper-v2.0-faster.
23
+
24
+ ```python
25
+ from faster_whisper import WhisperModel
26
+
27
+ model = WhisperModel("kotoba-tech/kotoba-whisper-v2.0-faster")
28
+
29
+ segments, info = model.transcribe("sample_ja_speech.wav", language="ja", chunk_length=15, condition_on_previous_text=False)
30
+ for segment in segments:
31
+ print("[%.2fs -> %.2fs] %s" % (segment.start, segment.end, segment.text))
32
+ ```
33
+
34
+ ### Benchmark
35
+ Please refer to the [kotoba-tech/kotoba-whisper-v1.0-faster](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-faster) for the detail of speed up [here](https://huggingface.co/kotoba-tech/kotoba-whisper-v1.0-faster#benchmark).
36
+
37
+
38
+
39
+ ## Conversion details
40
+
41
+ The original model was converted with the following command:
42
+
43
+ ```
44
+ ct2-transformers-converter --model kotoba-tech/kotoba-whisper-v2.0 --output_dir kotoba-whisper-v2.0-faster \
45
+ --copy_files tokenizer.json preprocessor_config.json --quantization float16
46
+ ```
47
+
48
+ Note that the model weights are saved in FP16. This type can be changed when the model is loaded using the [`compute_type` option in CTranslate2](https://opennmt.net/CTranslate2/quantization.html).
49
+
50
+ ## More information
51
+
52
+ For more information about the kotoba-whisper-v2.0, refer to the original [model card](https://huggingface.co/kotoba-tech/kotoba-whisper-v2.0).