Spaces:
Sleeping
Sleeping
top_md = """ | |
# VALL-E X | |
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1yyD_sz531QntLKowMHo-XxorsFBCfKul?usp=sharing) | |
[![GitHub](https://img.shields.io/badge/github-%23121011.svg?style=for-the-badge&logo=github&logoColor=white)](https://github.com/Plachtaa/vallex-webui) | |
Unofficial implementation of Microsoft's [VALL-E X](https://arxiv.org/pdf/2303.03926).<br> | |
VALL-E X can synthesize high-quality personalized speech with only a 3-second enrolled recording of | |
an unseen speaker as an acoustic prompt, even in another language for a monolingual speaker.<br> | |
This implementation supports zero-shot, mono-lingual/cross-lingual text-to-speech functionality of three languages (English, Chinese, Japanese)<br> | |
See this [demo](https://plachtaa.github.io/) page for more details. | |
""" | |
infer_from_audio_md = """ | |
Upload a speech of 3~10 seconds as the audio prompt and type in the text you'd like to synthesize.<br> | |
The model will synthesize speech of given text with the same voice of your audio prompt.<br> | |
The model also tends to preserve the emotion & acoustic environment of your given speech.<br> | |
For faster inference, please use **"Make prompt"** to get a `.npz` file as the encoded audio prompt, and use it by **"Infer from prompt"** | |
""" | |
make_prompt_md = """ | |
Upload a speech of 3~10 seconds as the audio prompt.<br> | |
Get a `.npz` file as the encoded audio prompt. Use it by **"Infer with prompt"** | |
""" | |
infer_from_prompt_md = """ | |
Faster than **"Infer from audio"**.<br> | |
You need to **"Make prompt"** first, and upload the encoded prompt (a `.npz` file) | |
""" |