israelisraeli's picture
Create README.md
7a64aed verified
metadata
language:
  - he
base_model:
  - ivrit-ai/whisper-large-v3-turbo-d4-p1-take2
pipeline_tag: automatic-speech-recognition
tags:
  - faster-whisper

ivrit-faster-whisper-turbo-d4

This model is a conversion of the ivrit-ai/whisper-large-v3-turbo-d4-p1-take2 model to the Faster-Whisper format, offering significantly faster inference times.

Model Overview

All credits go to ivrit-ai for developing the original Whisper model.

How to Use the Model

To use the model in your projects, follow the steps below to load and transcribe audio:

# Import the Faster Whisper module
import faster_whisper

# Load the model from Hugging Face
model = faster_whisper.WhisperModel("israelisraeli/ivrit-faster-whisper-turbo-d4", device="cuda")

# Transcribe the audio file to JSON
segs, _ = model.transcribe("AUDIOFILE_efiTheTigger.mp3", language="he")

# Format the output into a list of dictionaries with timestamps and text
transcribed_segments_with_timestamps = [
    {"start": s.start, "end": s.end, "text": s.text} for s in segs
]

import json

# Save the result to a JSON file
with open("transcribed_segments_with_timestamps.json", "w", encoding="utf-8") as json_file:
    json.dump(
        transcribed_segments_with_timestamps, json_file, ensure_ascii=False, indent=4
    )

print("Transcription saved to transcribed_segments_with_timestamps.json")

Conversion process

Tokenizer Conversion

from transformers import AutoTokenizer

# Load the tokenizer from the original Whisper model files
tokenizer_directory = "path_to_whisper_model_files"
tokenizer = AutoTokenizer.from_pretrained(tokenizer_directory)

# Save the tokenizer into a single JSON file
tokenizer.save_pretrained("path_to_save_directory", legacy_format=False)

Model Conversion to Faster-Whisper

To convert the original ivrit-ai/whisper-large-v3-turbo-d4-p1-take2 model to the Faster-Whisper format, i used the CTranslate2 library. The following command was used for the conversion:

ct2-transformers-converter \
  --model ./whisper-large-v3-turbo-d4-p1-take2 \
  --output_dir ./ivrit-faster-whisper-turbo-d4 \
  --copy_files tokenizer.json preprocessor_config.json \