Update app.py
Browse files
app.py
CHANGED
@@ -74,6 +74,7 @@ def audioMode(audio):
|
|
74 |
print(type(audio))
|
75 |
print(audio)
|
76 |
text = pipe(audio, batch_size=8, generate_kwargs={"task": "transcribe"}, return_timestamps=True)["text"]
|
|
|
77 |
#sampling_rate = audio[0]
|
78 |
#audio_array = audio[1]
|
79 |
#print(sampling_rate)
|
@@ -81,7 +82,7 @@ def audioMode(audio):
|
|
81 |
#input_features = processor_audio(audio_array, sampling_rate=16000, return_tensors="pt").input_features
|
82 |
#predicted_ids = model_audio.generate(input_features)
|
83 |
#transcription = processor_audio.batch_decode(predicted_ids, skip_special_tokens=True)
|
84 |
-
return
|
85 |
|
86 |
|
87 |
interface_title = "TSAI-ERA-V1 - Capstone - Multimodal GPT Demo"
|
|
|
74 |
print(type(audio))
|
75 |
print(audio)
|
76 |
text = pipe(audio, batch_size=8, generate_kwargs={"task": "transcribe"}, return_timestamps=True)["text"]
|
77 |
+
pred_text = textMode(text, 50)
|
78 |
#sampling_rate = audio[0]
|
79 |
#audio_array = audio[1]
|
80 |
#print(sampling_rate)
|
|
|
82 |
#input_features = processor_audio(audio_array, sampling_rate=16000, return_tensors="pt").input_features
|
83 |
#predicted_ids = model_audio.generate(input_features)
|
84 |
#transcription = processor_audio.batch_decode(predicted_ids, skip_special_tokens=True)
|
85 |
+
return pred_text
|
86 |
|
87 |
|
88 |
interface_title = "TSAI-ERA-V1 - Capstone - Multimodal GPT Demo"
|