Multi-Speaker Adaptive Text-to-Speech Generation with python
StyleSpeech PyTorch Implementation of Meta-StyleSpeech : Multi-Speaker Adaptive Text-to-Speech Generation. [x] StyleSpeech (naive branch) [x] Meta-StyleSpeech (main branch) Dependencies You can install the Python dependencies with pip3 install -r requirements.txt Inference You have to download pretrained models and put them in output/ckpt/LibriTTS/. For English single-speaker TTS, run python3 synthesize.py –text “YOUR_DESIRED_TEXT” –ref_audio path/to/reference_audio.wav –restore_step 200000 –mode single -p config/LibriTTS/preprocess.yaml -m config/LibriTTS/model.yaml -t config/LibriTTS/train.yaml The generated utterances will be put in output/result/. Your synthesized speech will have ref_audio‘s style. Batch Inference […]
Read more