top_md = """ # VALL-E X [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1yyD_sz531QntLKowMHo-XxorsFBCfKul?usp=sharing) [![GitHub](https://img.shields.io/badge/github-%23121011.svg?style=for-the-badge&logo=github&logoColor=white)](https://github.com/Plachtaa/vallex-webui) Unofficial implementation of Microsoft's [VALL-E X](https://arxiv.org/pdf/2303.03926).
VALL-E X can synthesize high-quality personalized speech with only a 3-second enrolled recording of an unseen speaker as an acoustic prompt, even in another language for a monolingual speaker.
This implementation supports zero-shot, mono-lingual/cross-lingual text-to-speech functionality of three languages (English, Chinese, Japanese)
See this [demo](https://plachtaa.github.io/) page for more details. """ infer_from_audio_md = """ Upload a speech of 3~10 seconds as the audio prompt and type in the text you'd like to synthesize.
The model will synthesize speech of given text with the same voice of your audio prompt.
The model also tends to preserve the emotion & acoustic environment of your given speech.
For faster inference, please use **"Make prompt"** to get a `.npz` file as the encoded audio prompt, and use it by **"Infer from prompt"** """ make_prompt_md = """ Upload a speech of 3~10 seconds as the audio prompt.
Get a `.npz` file as the encoded audio prompt. Use it by **"Infer with prompt"** """ infer_from_prompt_md = """ Faster than **"Infer from audio"**.
You need to **"Make prompt"** first, and upload the encoded prompt (a `.npz` file) """