aka7774 commited on
Commit
659d1cf
1 Parent(s): 8c0f874

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +15 -4
app.py CHANGED
@@ -1,18 +1,29 @@
1
  import gradio as gr
2
  import whisper
 
3
 
4
  model_size = 'large-v3'
5
- model = whisper.load_model(model_size)
 
 
 
 
 
 
 
6
 
7
  def speech_to_text(audio_file, _model_size):
8
  global model_size, model
9
  if model_size != _model_size:
10
  model_size = _model_size
11
- model = whisper.load_model(model_size)
 
12
 
13
- result = model.transcribe(audio_file)
 
14
 
15
- return result["text"]
 
16
 
17
  gr.Interface(
18
  fn=speech_to_text,
 
1
  import gradio as gr
2
  import whisper
3
+ from faster_whisper import WhisperModel
4
 
5
  model_size = 'large-v3'
6
+ #model = whisper.load_model(model_size)
7
+ #model = WhisperModel(model_size, device="cuda", compute_type="float16")
8
+ model = WhisperModel(model_size, compute_type="float16")
9
+
10
+ # or run on GPU with INT8
11
+ # model = WhisperModel(model_size, device="cuda", compute_type="int8_float16")
12
+ # or run on CPU with INT8
13
+ # model = WhisperModel(model_size, device="cpu", compute_type="int8")
14
 
15
  def speech_to_text(audio_file, _model_size):
16
  global model_size, model
17
  if model_size != _model_size:
18
  model_size = _model_size
19
+ #model = whisper.load_model(model_size)
20
+ model = WhisperModel(model_size, compute_type="float16")
21
 
22
+ #result = model.transcribe(audio_file)
23
+ segments, info = model.transcribe(audio_file, beam_size=5)
24
 
25
+ # return result["text"]
26
+ return "".join([segment.text for segment in segments])
27
 
28
  gr.Interface(
29
  fn=speech_to_text,