-
Notifications
You must be signed in to change notification settings - Fork 41
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Prosody Loss #15
Comments
Hi @inconnu11 , thanks for your attention. My intention was to prevent the prosody encoder learning meaningless representations at the first few training steps. But you can remove |
Hmm, it's weird. If you have room for that, could you please do some sanity checks on your side? For example, removing some part of the code to make it simpler until the nan loss disappear would be one. It will definitely be helpful for others interested in this issue. |
I'd like to do so. But it takes too long to train it. I have to train the model for 7days with one gpu T4. Are there any parts |
The MDN loss (i.e. negative log-likelihood) can be negative value. However, in your log, it is almost 0 before becoming nan. |
Hi, I am adding your MDN prosody modeling code segment to my tacotron but I encountered several problems about the code segment about prosody modeling. First, the prosody loss is added into the total loss only after the
prosody_loss_enable_steps
but in the training steps before theprosody_loss_enable_steps
the prosody representation is already added with the text encoding. Does it means in the training steps before theprosody_loss_enable_steps
, the prosody representation is optimized without the prosody loss?Second, in the training steps, the backward gradient of training prosody predictor should be acted like "stop gradient" but it seems little relevant code.
Thanks!
The text was updated successfully, but these errors were encountered: