-
Notifications
You must be signed in to change notification settings - Fork 419
This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can I train the Chinese model? #70
Comments
Look at issue #41 to check the current progress. |
You can, but with the current PL-BERT in English the quality won’t be as good it’s originally proposed to be. I’m working on multilingual PL-BERT now and it may take one or two months to finish. |
See yl4579/StyleTTS#10 for more details. |
@yl4579 I trained styletts2 successfully using Chinese data, it sound very good. As wavlm-base-plus only supporting English, I used a Chinese Hubert model as SLM. When I want to train a model both for Chinese and English, I can not find a pre-trained model sopport Chinese and English at the same time. About SLM,Do you have any suggestions ? |
You can try whisper encoder that was trained with multiple languages. You can also try multilingual wav2vec2.0: https://huggingface.co/facebook/wav2vec2-large-xlsr-53 |
Did you use the English PL-BERT or did you train PL-BERT with Chinese data? |
train PL-BERT with Chinese data |
What is your modeling unit? IPA or Pinyin? |
I trained Chinese PL-BERT without pinyin tones. But maybe PL-BERT with tones will also work normally, so you can try. |
How many samples did you use to train Chinese PL-BERT? |
@zhouyong64 I used about 84,000,000 text sentences to train the Chinese PL-BERT model. |
@Moonmore @zhouyong64 Sorry for the wrong information of yesterday, I tained PL-BERT with tones, and trained asr without tones.
|
So can I understand that all text-related models are trained using the same phoneme unit, and the characteristics of each minimum pronunciation modeling unit are obtained. like(ni3 hao3 -> n i3 h ao3), The input length is 4, and the output length of the model is also 4. text encoder and the bert model. and how to construct the plbert label? |
@Moonmore |
@hermanseu Thank you for your reply. |
This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
I want to train the Chinese model. Do you support mixed input in Chinese and English?
The text was updated successfully, but these errors were encountered: