Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How good are the pretrained model? #27

Open
BuaaAlban opened this issue Jun 18, 2020 · 1 comment
Open

How good are the pretrained model? #27

BuaaAlban opened this issue Jun 18, 2020 · 1 comment

Comments

@BuaaAlban
Copy link

Can I get the quality of end2end https://melgan-neurips.github.io/ in the demo using the preovided model ?
I have tried Fastspeech and Tacotron2 to generated Mel spectrogram, and use the pretrained melgan vocoder to generate wav, but the result can't reach the quality of the demo and it's no better than waveglow.
What should I do to improve the performance?

@rijulg
Copy link

rijulg commented Aug 20, 2020

@BuaaAlban have you tried using an audio file directly to make sure that the quality degradation is coming from the vocoder and not tacotron?
You can run the following pipeline:

audio -> MelVocoder() -> mel -> MelVocoder.inverse() -> y_audio

and check whether the quality is good enough or not. In my experience so far, it is pretty good.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Development

No branches or pull requests

2 participants