update
Browse files
app.py
CHANGED
@@ -163,13 +163,13 @@ demo = gr.Blocks()
|
|
163 |
with demo:
|
164 |
gr.Markdown("<h1><center>Talk to Your Multilingual AI Assistant</center></h1>")
|
165 |
gr.Markdown(
|
166 |
-
"""Model pipeline consisting of - <br>- [**Whisper**](https://github.com/openai/whisper)for Speech-to-text, <br>- [**Bloom**](https://huggingface.co/bigscience/bloom) for Text-generation, and <br>- [**CoquiTTS**](https://huggingface.co/coqui) for Text-To-Speech. <br><br> Front end is built using [**Gradio Block API**](https://gradio.app/docs/#blocks).<br>All three models are Multilingual, however, there are only these three overlapping languages among them - Spanish (es), French(fr), and English(en). Hence it would be suggested to test using these languages to get the best results out of this ML-App. If an English voice input is given then both the textbox on the left-hand side would show the same transcripts. However, if the input is either in
|
167 |
""")
|
168 |
with gr.Row():
|
169 |
with gr.Column():
|
170 |
-
in_audio = gr.Audio(source="microphone", type="filepath", label='Record your voice here') #type='filepath'
|
171 |
b1 = gr.Button("AI response pipeline (Whisper - Bloom - Coqui pipeline)")
|
172 |
-
out_transcript = gr.Textbox(label= '
|
173 |
out_translation_en = gr.Textbox(label= 'English Translation of audio using OpenAI Whisper')
|
174 |
with gr.Column():
|
175 |
out_audio = gr.Audio(label='AI response in Audio form in your preferred language')
|
|
|
163 |
with demo:
|
164 |
gr.Markdown("<h1><center>Talk to Your Multilingual AI Assistant</center></h1>")
|
165 |
gr.Markdown(
|
166 |
+
"""Model pipeline consisting of - <br>- [**Whisper**](https://github.com/openai/whisper)for Speech-to-text, <br>- [**Bloom**](https://huggingface.co/bigscience/bloom) for Text-generation, and <br>- [**CoquiTTS**](https://huggingface.co/coqui) for Text-To-Speech. <br><br> Front end is built using [**Gradio Block API**](https://gradio.app/docs/#blocks).<br>All three models are Multilingual, however, there are only these three overlapping languages among them - Spanish (es), French(fr), and English(en). Hence it would be suggested to test using these languages to get the best results out of this ML-App. If an English voice input is given then both the textbox on the left-hand side would show the same transcripts. However, if the input is either in _Spanish_ or _French_, then the first textbox would show the language transcript, while the next one would show its English translations.
|
167 |
""")
|
168 |
with gr.Row():
|
169 |
with gr.Column():
|
170 |
+
in_audio = gr.Audio(source="microphone", type="filepath", label='Record your voice here in English, Spanish or French for best results-') #type='filepath'
|
171 |
b1 = gr.Button("AI response pipeline (Whisper - Bloom - Coqui pipeline)")
|
172 |
+
out_transcript = gr.Textbox(label= 'English/Spanish/French Transcript of your Audio using OpenAI Whisper')
|
173 |
out_translation_en = gr.Textbox(label= 'English Translation of audio using OpenAI Whisper')
|
174 |
with gr.Column():
|
175 |
out_audio = gr.Audio(label='AI response in Audio form in your preferred language')
|