File size: 1,426 Bytes
5054447
 
 
 
 
 
2123e3c
5054447
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f0a0639
471b1ce
5054447
 
 
 
 
a80fa0a
5054447
 
 
 
 
 
471b1ce
3a172df
5054447
f0a0639
 
 
 
5054447
 
2123e3c
f0a0639
5054447
 
 
 
 
 
 
 
 
471b1ce
5054447
f0a0639
5054447
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
import torch

import gradio as gr
import pytube as pt
from transformers import pipeline

MODEL_NAME = "openai/whisper-small.en"

device = 0 if torch.cuda.is_available() else "cpu"

pipe = pipeline(
    task="automatic-speech-recognition",
    model=MODEL_NAME,
    chunk_length_s=30,
    device=device,
)


all_special_ids = pipe.tokenizer.all_special_ids
transcribe_token_id = all_special_ids[-5]
translate_token_id = all_special_ids[-6]


def transcribe(microphone, state, task="transcribe"):
    file = microphone

    pipe.model.config.forced_decoder_ids = [[2, transcribe_token_id if task=="transcribe" else translate_token_id]]

    text = pipe(file)["text"]

    return state + "\n" + text, state + "\n" + text



mf_transcribe = gr.Interface(
    fn=transcribe,
    inputs=[
        gr.Audio(source="microphone", type="filepath", optional=True),
        gr.State(value="")
    ],
    outputs=[
        gr.Textbox(lines=15),
        gr.State()]
    ,
    layout="horizontal",
    theme="huggingface",
    title="Whisper Small.EN: Transcribe Audio",
    live=True,
    description=(
        "Transcribe long-form microphone or audio inputs with the click of a button! Demo uses the"
        f" checkpoint [{MODEL_NAME}](https://huggingface.co/{MODEL_NAME}) and 🤗 Transformers to transcribe audio files"
        " of arbitrary length."
    ),
    allow_flagging="never",
)




mf_transcribe.launch(enable_queue=True)