main
Browse files
app.py
CHANGED
@@ -4,13 +4,10 @@ import pandas as pd
|
|
4 |
import torch
|
5 |
import torchaudio
|
6 |
import time
|
7 |
-
|
8 |
from transformers import pipeline
|
9 |
-
# from speechbrain.inference.VAD import VAD
|
10 |
from speechbrain.inference.classifiers import EncoderClassifier
|
11 |
|
12 |
transcriber = pipeline("automatic-speech-recognition", model="openai/whisper-tiny")
|
13 |
-
# VAD = VAD.from_hparams(source="speechbrain/vad-crdnn-libriparty", savedir="pretrained_models/vad-crdnn-libriparty")
|
14 |
language_id = EncoderClassifier.from_hparams(source="speechbrain/lang-id-voxlingua107-ecapa")
|
15 |
|
16 |
data = []
|
@@ -123,5 +120,6 @@ demo = gr.Interface(
|
|
123 |
description="Speak into the microphone and see real-time audio processing results."
|
124 |
)
|
125 |
|
126 |
-
demo.launch()
|
127 |
|
|
|
|
|
|
4 |
import torch
|
5 |
import torchaudio
|
6 |
import time
|
|
|
7 |
from transformers import pipeline
|
|
|
8 |
from speechbrain.inference.classifiers import EncoderClassifier
|
9 |
|
10 |
transcriber = pipeline("automatic-speech-recognition", model="openai/whisper-tiny")
|
|
|
11 |
language_id = EncoderClassifier.from_hparams(source="speechbrain/lang-id-voxlingua107-ecapa")
|
12 |
|
13 |
data = []
|
|
|
120 |
description="Speak into the microphone and see real-time audio processing results."
|
121 |
)
|
122 |
|
|
|
123 |
|
124 |
+
if __name__ == "__main__":
|
125 |
+
demo.launch()
|