Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Could you kindly share how to train new model? #20

Open
a-ejiri opened this issue Nov 25, 2022 · 1 comment
Open

Could you kindly share how to train new model? #20

a-ejiri opened this issue Nov 25, 2022 · 1 comment
Assignees
Labels
documentation Improvements or additions to documentation question Further information is requested

Comments

@a-ejiri
Copy link

a-ejiri commented Nov 25, 2022

Hi!
I appreciate for sharing your marine. I used your pretrained model trained by JSUT.
Though pretrained model works very well in ordinary corpus, when I used it in my original corpus (corpus about games or animations), its accuracy was about 50%.
So I wanted to build my model by using my corpus set and testify accuracy.

I know you described it will come soon. I'm very looking forward to it, and want to know when you plan to share it.
Thank you.

@6gsn
Copy link
Owner

6gsn commented Dec 16, 2022

Hi, @a-ejiri!
Thank you for your interest and comments! Here are my answers to your comments;

I used your pretrained model trained by JSUT. Though pretrained model works very well in ordinary corpus, when I used it in my original corpus (corpus about games or animations), its accuracy was about 50%.

That's a pretty predictable result! As you know, the JSUT corpus covers all of the main pronunciations of daily-use Japanese characters [Sonobe, 2017]. However, your test set seems to include a lot of named entities regarding specific domains (e.g., game, animation). Hence, the current pre-trained model may not cover the samples without fine-tuning.

I'm working on preparing the documents and the recipes for building a model with our codes! And the jobs will be finished in serval weeks. I would appreciate your patience in a few whiles!

Thanks!

[Sonobe, 2017] Ryosuke Sonobe, Shinnosuke Takamichi and Hiroshi Saruwatari, "JSUT corpus: free large-scale Japanese speech corpus for end-to-end speech synthesis," arXiv preprint, 1711.00354, 2017.

@6gsn 6gsn added the question Further information is requested label Dec 16, 2022
@6gsn 6gsn self-assigned this Dec 16, 2022
@6gsn 6gsn added the documentation Improvements or additions to documentation label Dec 16, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants