Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Availability of using other BERT alternatives for current langauges #145

Open
gordon0414 opened this issue Jun 20, 2024 · 3 comments
Open

Comments

@gordon0414
Copy link

gordon0414 commented Jun 20, 2024

Hello, I was wondering if we can replace the BERT for current supported models

    Languages.JP: BASE_DIR / "bert" / "deberta-v2-large-japanese-char-wwm",
    Languages.EN: BASE_DIR / "bert" / "deberta-v3-large",
    Languages.ZH: BASE_DIR / "bert" / "chinese-roberta-wwm-ext-large",

Since the current models are mostly large variants of the models, but I think smaller BERT models will still be capable for our use case.

I mean by smaller variants such as
- English: microsoft/deberta-v3-small
- Japanese: ku-nlp/deberta-v2-base-japanese-char-wwm

I suppose these variants won't cause any issues since the tokenizers are identical to the original ones

My hypothesis is that if we fine-tune our model with the new BERT models, it will seamlessly incorporate with the current system.

@gordon0414 gordon0414 changed the title Availability using other BERT alternatives for current langauges Availability of using other BERT alternatives for current langauges Jun 20, 2024
@litagin02
Copy link
Owner

I haven't checked, but maybe as you guess we can, and you can try it.

@gordon0414
Copy link
Author

okay, maybe I should try it this weekend!

@gordon0414
Copy link
Author

The dimensional discrepancy between the base and large models (768 vs. 1024) seems to be a significant problem. We need to come up with techniques such as knowledge distillation or other methods.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants