YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Tortoise TTS AR model fine-tuned for German

Trained on 3 speakers; 2 LibriVox readers, and Thorsten Mueller's dataset https://github.com/thorstenMueller/Thorsten-Voice

THE NEWEST VERSIONS: v# indicates the number of training sessions, #e is how many epochs. 9/5 training session uploaded

Requires the tokenizer file placed in the tokenizers/ directory

Voice latents are pre-computed in voices/ for some uploaded versions. Voice samples to recompute latents are uploaded.

For use in MRQ Voice Cloning WebUI:

Requires the tokenizer used in training, and code changes to disable text cleaners. At minimum, change english_cleaners to basic_cleaners.

Code changes: modules\tortoise-tts\tortoise\utils\tokenizer.py Change Line 201: txt = english_cleaners(txt) and replace it with txt = basic_cleaners(txt)

modules\tortoise-tts\build\lib\tortoise\utils\tokenizer.py Change Line 201: txt = english_cleaners(txt) and replace it with txt = basic_cleaners(txt)

\modules\dlas\dlas\data\audio\paired_voice_audio_dataset.py Line 133: return text_to_sequence(txt, ['english_cleaners']) and replace it with: return text_to_sequence(txt, ['basic_cleaners'])

modules\dlas\dlas\data\audio\voice_tokenizer.py Line 14: from dlas.models.audio.tts.tacotron2.text.cleaners import english_cleaners to: from dlas.models.audio.tts.tacotron2.text.cleaners import english_cleaners, basic_cleaners Line 85: txt = english_cleaners(txt) to txt = basic_cleaners(txt) Line 134: word = english_cleaners(word) to basic_cleaners(word)

Copy and paste German text into the tokenizer tester on the utilities tab, and you should see it tokenized with all of the special characters, and no [UNK].

license: other language: - de

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.