mirror of https://github.com/coqui-ai/TTS.git
* Update BaseDatasetConfig - Add dataset_name - Chane name to formatter_name * Update compute_embedding - Allow entering dataset by args - Use released model by default - Use the new key format * Update loading * Update recipes * Update other dep code * Update tests * Fixup * Load multiple embedding files * Fix argument names in dep code * Update docs * Fix argument name * Fix linter |
||
---|---|---|
.. | ||
align_tts | ||
glow_tts | ||
hifigan | ||
multiband_melgan | ||
speedy_speech | ||
tacotron2-DDC | ||
univnet | ||
vits_tts | ||
wavegrad | ||
wavernn | ||
README.md | ||
download_thorsten_DE.sh |
README.md
🐸💬 TTS Thorsten Recipes
For running the recipes you need the Thorsten-Voice dataset.
You can download it manually from the official website or use download_thorsten_de.sh
alternatively running any of the train_modelX.pyscripts will download the dataset if not already present.
Then, go to your desired model folder and run the training.
Running Python files. (Choose the desired GPU ID for your run and set ```CUDA_VISIBLE_DEVICES```)
```terminal
CUDA_VISIBLE_DEVICES="0" python train_modelX.py
```
💡 Note that these runs are just templates to help you start training your first model. They are not optimized for the best result. Double-check the configurations and feel free to share your experiments to find better parameters together 💪.