Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Inference

from transformers import WhisperFeatureExtractor, WhisperProcessor
import numpy as np
import librosa
import torch

model = WhisperForConditionalGeneration.from_pretrained("userdata/ud-whisper-medium-1").cuda()
processor = WhisperProcessor.from_pretrained( "userdata/ud-whisper-medium-1")

_ = model.eval()
model.config.forced_decoder_ids = None

sec = 30
target_sr = 16_000     
audio, sr = librosa.load('/home/userdata/ariff-wav2vec2/finetune/2887.mp3', sr=None)  
audio_array = librosa.resample(audio, orig_sr=sr, target_sr=target_sr)
chunk = [audio_array[i: i + (target_sr * sec)] for i in range(0, len(audio_array), target_sr * sec)]

with torch.no_grad():
    input_features = (processor(chunk[4], sampling_rate=16_000, return_tensors="pt").input_features).cuda()
    predicted_ids = model.generate(input_features)
    transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)

res = ''.join(list(transcription))
print(res)

Play Audio

import IPython.display as ipd
ipd.Audio(data=np.asarray(chunk), autoplay=True, rate=16000)
Downloads last month
3
Safetensors
Model size
764M params
Tensor type
BF16
·
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.