Vaibhav Srivastav
up
e7a921a
raw
history blame
No virus
3.42 kB
import gradio as gr
from TTS.api import TTS
tts = TTS("tts_models/multilingual/multi-dataset/xtts_v1")
tts.to("cuda")
def predict(prompt, language, audio_file_pth, agree):
if agree == True:
tts.tts_to_file(
text=prompt,
file_path="output.wav",
speaker_wav=audio_file_pth,
language=language,
)
return gr.make_waveform(
audio="output.wav",
), gr.Audio(value="output.wav")
else:
gr.Warning("Please accept the Terms & Condition!")
title = "Coqui🐸 XTTS"
description = """
<a href="https://huggingface.co/coqui/XTTS-v1">XTTS</a> is a Voice generation model that lets you clone voices into different languages by using just a quick 3-second audio clip.
<br/>
Built on Tortoise, XTTS has important model changes that make cross-language voice cloning and multi-lingual speech generation super easy.
<br/>
This is the same model that powers Coqui Studio, and Coqui API, however we apply a few tricks to make it faster and support streaming inference.
<br/>
<br/>
<p>For faster inference without waiting in the queue, you should duplicate this space and upgrade to GPU via the settings.
<br/>
<a href="https://huggingface.co/spaces/coqui/xtts?duplicate=true">
<img style="margin-top: 0em; margin-bottom: 0em" src="https://bit.ly/3gLdBN6" alt="Duplicate Space"></a>
</p>
"""
article = """
<div style='margin:20px auto;'>
<p>By using this demo you agree to the terms of the Coqui Public Model License at https://coqui.ai/cpml</p>
</div>
"""
examples = [
# [
# "Once when I was six years old I saw a magnificent picture.",
# "en",
# "examples/female.wav",
# True
# ],
[
"Lorsque j'avais six ans j'ai vu, une fois, une magnifique image.",
"fr",
"examples/male.wav",
True
],
# [
# "Un tempo lontano, quando avevo sei anni, vidi un magnifico disegno.",
# "it",
# "examples/female.wav",
# True
# ],
]
gr.Interface(
fn=predict,
inputs=[
gr.Textbox(
label="Text Prompt",
info="One or two sentences at a time is better",
value="It took me quite a long time to develop a voice, and now that I have it I'm not going to be silent.",
),
gr.Dropdown(
label="Language",
info="Select an output language for the synthesised speech",
choices=[
"en",
"es",
"fr",
"de",
"it",
"pt",
"pl",
"tr",
"ru",
"nl",
"cz",
"ar",
"zh",
],
max_choices=1,
value="en",
),
gr.Audio(
label="Reference Audio",
info="Click on the ✎ button to upload your own target speaker audio",
type="filepath",
value="examples/male.wav",
),
gr.Checkbox(label="Agree", value=True, info="I agree to the terms of the Coqui Public Model License at https://coqui.ai/cpml")
],
outputs=[
gr.Video(label="Synthesised Waveform"),
gr.Audio(label="Synthesised Audio"),
],
title=title,
description=description,
article=article,
examples=examples,
).queue().launch(debug=True)