Whisper large-v3 model for CTranslate2

This repository contains the conversion of whisper-turbo to the CTranslate2 model format.

Example

from huggingface_hub import snapshot_download
from faster_whisper import WhisperModel

repo_id = "jootanehorror/faster-whisper-large-v3-turbo-ct2"
local_dir = "faster-whisper-large-v3-turbo-ct2"
snapshot_download(repo_id=repo_id, local_dir=local_dir, repo_type="model")

model = WhisperModel(local_dir, device='cpu', compute_type='int8')

segments, info = model.transcribe("sample.mp3")

for segment in segments:
    print("[%.2fs -> %.2fs] %s" % (segment.start, segment.end, segment.text))

More information

For more information about the model, see its official github page.

Downloads last month
35
Inference Examples
Inference API (serverless) does not yet support ctranslate2 models for this pipeline type.