ysharma HF staff commited on
Commit
073aa56
1 Parent(s): 80a64d7
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -163,13 +163,13 @@ demo = gr.Blocks()
163
  with demo:
164
  gr.Markdown("<h1><center>Talk to Your Multilingual AI Assistant</center></h1>")
165
  gr.Markdown(
166
- """Model pipeline consisting of - <br>- [**Whisper**](https://github.com/openai/whisper)for Speech-to-text, <br>- [**Bloom**](https://huggingface.co/bigscience/bloom) for Text-generation, and <br>- [**CoquiTTS**](https://huggingface.co/coqui) for Text-To-Speech. <br><br> Front end is built using [**Gradio Block API**](https://gradio.app/docs/#blocks).<br>All three models are Multilingual, however, there are only these three overlapping languages among them - Spanish (es), French(fr), and English(en). Hence it would be suggested to test using these languages to get the best results out of this ML-App. If an English voice input is given then both the textbox on the left-hand side would show the same transcripts. However, if the input is either in Spanish or French, then the first textbox would show the language transcript, while the next one would show its English translations.
167
  """)
168
  with gr.Row():
169
  with gr.Column():
170
- in_audio = gr.Audio(source="microphone", type="filepath", label='Record your voice here') #type='filepath'
171
  b1 = gr.Button("AI response pipeline (Whisper - Bloom - Coqui pipeline)")
172
- out_transcript = gr.Textbox(label= 'As is Transcript using OpenAI Whisper')
173
  out_translation_en = gr.Textbox(label= 'English Translation of audio using OpenAI Whisper')
174
  with gr.Column():
175
  out_audio = gr.Audio(label='AI response in Audio form in your preferred language')
 
163
  with demo:
164
  gr.Markdown("<h1><center>Talk to Your Multilingual AI Assistant</center></h1>")
165
  gr.Markdown(
166
+ """Model pipeline consisting of - <br>- [**Whisper**](https://github.com/openai/whisper)for Speech-to-text, <br>- [**Bloom**](https://huggingface.co/bigscience/bloom) for Text-generation, and <br>- [**CoquiTTS**](https://huggingface.co/coqui) for Text-To-Speech. <br><br> Front end is built using [**Gradio Block API**](https://gradio.app/docs/#blocks).<br>All three models are Multilingual, however, there are only these three overlapping languages among them - Spanish (es), French(fr), and English(en). Hence it would be suggested to test using these languages to get the best results out of this ML-App. If an English voice input is given then both the textbox on the left-hand side would show the same transcripts. However, if the input is either in _Spanish_ or _French_, then the first textbox would show the language transcript, while the next one would show its English translations.
167
  """)
168
  with gr.Row():
169
  with gr.Column():
170
+ in_audio = gr.Audio(source="microphone", type="filepath", label='Record your voice here in English, Spanish or French for best results-') #type='filepath'
171
  b1 = gr.Button("AI response pipeline (Whisper - Bloom - Coqui pipeline)")
172
+ out_transcript = gr.Textbox(label= 'English/Spanish/French Transcript of your Audio using OpenAI Whisper')
173
  out_translation_en = gr.Textbox(label= 'English Translation of audio using OpenAI Whisper')
174
  with gr.Column():
175
  out_audio = gr.Audio(label='AI response in Audio form in your preferred language')