-
Notifications
You must be signed in to change notification settings - Fork 1.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[TTS]Update VITS to support VITS and its voice cloning training on AIShell-3 #2268
Conversation
Thank you for your contribution. I've been busy publishing the version recently. I'll have a large amount of time to review your code at the end of September. If you are in Chinese Mainland, you can scan the wechat QR code on the homepage to join our user group and add '子龙' for discussion. |
This pull request is now in conflict :( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
PR types
New features | Bug fixes
PR changes
APIs | Docs
Describe
New features:
Bug fixes:
VITS inference code:
feats_lengths = paddle.to_tensor([paddle.shape(feats)[2]])
->feats_lengths = paddle.to_tensor(paddle.shape(feats)[2])
Extra:
I added the docs for traning VITS or VITS-VC on aishell-3 dataset, but leave a
TODO
label for pretrained model. I trained the two new examples with a little modified on aistudio single v100 16GB card withbatch_size=24
for 25000 step. This is to verify that training can be performed normally.I hope the official could release the official version of the pre-trained model trained on 4 cards in the future.
Here are some outputs of my 25000 step models:
test_vits.zip
test_e2e_vits.zip
test_vits_vc.zip
vc_syn_vits_vc_src_text.zip
vc_syn_vits_vc_src_audio.zip