VoiceSmith makes it possible to train and infer on both single and multispeaker models without any coding experience. It fine-tunes a pretty solid text to speech pipeline based on a modified version of DelightfulTTS and UnivNet on your dataset. Both models were pretrained on a proprietary 5000 speaker dataset. It also provides some tools for dataset preprocessing like automatic text normalization.
If you want to play around with a model trained on a highly emotional emotional 60 speaker dataset using an earlier version of this software click here.

Make sure you have the latest version of Node.js installed
Clone the repository
git clone https://github.com/dunky11/voicesmith
Install dependencies, this can take a minute
cd voicesmith
npm install
Click here, select the folder with the latest version, download all files and place them inside the repositories assets folder.
Start the project
npm start
Follow steps 1 - 4 from above.
Run make, this will create a folder named out/make with an installer inside. The installer will be different based on your operating system.
npm make
VoiceSmith currently uses a two-stage modified DelightfulTTS and UnivNet pipeline.

Show your support by the project. Pull requests are always welcome.
This project is licensed under the Apache-2.0 license - see the LICENSE.md file for details.