-
Notifications
You must be signed in to change notification settings - Fork 8.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bacth size relation with memory #587
Comments
what is the best way of stopping the training so you can test if the model is any good? |
All of the utterances in the training batch are padded to match the length of the longest one. To reduce memory consumption you can also reduce You can stop the training at every time, but you will lose all progress since the last saved checkpoint. Use |
Thank you very much! |
I preproccesed and embedded with 'with metadata_fpath.open("r", encoding="latin-1") as metadata_file:' |
I think im importing it wrong because the wavs folder gives me some what understandable audio files. |
Tacotron uses "teacher forcing" training which means the decoder cheats and uses the ground truth mel to predict subsequent frames. It's necessary to ensure the input and output have same length which facilitates a loss calculation. For inference you do not have a ground truth, so the model has to rely on what it learned during training. Toolbox output with your new model will not be intelligible until about 20-50k steps. You need 150-250k steps not to have serious bugs during inference. A few words about dataset quality: if your dataset is good, it may converge in fewer steps. And if it is bad, your inference will never be good. Many free datasets are bad quality. I will recommend that you modify synthesizer/utils/symbols.py to contain all letters of your alphabet. Anything that is not in that list will be ignored by the model, and make it very hard to generalize. There's a learning curve to this. Since you are training a new model, it will be very helpful to study the code carefully and figure out what it's doing. If there is something you don't understand, read the papers linked in the README for context. |
Thank you very much! |
Hi,
while training the synthesizer with batch size 32 I get an error after 30 steps on average with the error of my GPU ran out of memory.
I tried reducing the batch size to 32 but I'm still getting an error.
what are other things I can do, can I go with a batch size below 32?
The text was updated successfully, but these errors were encountered: