Tortoise-TTS-de / README.md
AOLCDROM's picture
Update README.md
02ad57c
Tortoise TTS AR model fine-tuned for German
Trained on 3 speakers; 2 LibriVox readers, and Thorsten Mueller's dataset https://github.com/thorstenMueller/Thorsten-Voice
***THE NEWEST VERSIONS***: v# indicates the number of training sessions, #e is how many epochs.
9/5 training session uploaded
Requires the tokenizer file placed in the tokenizers/ directory
Voice latents are pre-computed in voices/ for some uploaded versions. Voice samples to recompute latents are uploaded.
For use in MRQ Voice Cloning WebUI:
Requires the tokenizer used in training, and code changes to disable text cleaners. At minimum, change english_cleaners to basic_cleaners.
Code changes:
modules\tortoise-tts\tortoise\utils\tokenizer.py
Change Line 201: txt = english_cleaners(txt) and replace it
with txt = basic_cleaners(txt)
modules\tortoise-tts\build\lib\tortoise\utils\tokenizer.py
Change Line 201: txt = english_cleaners(txt) and replace it
with txt = basic_cleaners(txt)
\modules\dlas\dlas\data\audio\paired_voice_audio_dataset.py
Line 133: return text_to_sequence(txt, ['english_cleaners'])
and replace it with: return text_to_sequence(txt, ['basic_cleaners'])
modules\dlas\dlas\data\audio\voice_tokenizer.py
Line 14: from dlas.models.audio.tts.tacotron2.text.cleaners import
english_cleaners
to: from dlas.models.audio.tts.tacotron2.text.cleaners import
english_cleaners, basic_cleaners
Line 85: txt = english_cleaners(txt) to txt =
basic_cleaners(txt)
Line 134: word = english_cleaners(word) to basic_cleaners(word)
Copy and paste German text into the tokenizer tester on the utilities
tab, and you should see it tokenized with all of the special
characters, and no [UNK].
---
license: other
language:
- de
---