You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I suppose these variants won't cause any issues since the tokenizers are identical to the original ones
My hypothesis is that if we fine-tune our model with the new BERT models, it will seamlessly incorporate with the current system.
The text was updated successfully, but these errors were encountered:
gordon0414
changed the title
Availability using other BERT alternatives for current langauges
Availability of using other BERT alternatives for current langauges
Jun 20, 2024
The dimensional discrepancy between the base and large models (768 vs. 1024) seems to be a significant problem. We need to come up with techniques such as knowledge distillation or other methods.
Hello, I was wondering if we can replace the BERT for current supported models
Since the current models are mostly
large
variants of the models, but I think smaller BERT models will still be capable for our use case.I mean by smaller variants such as
- English: microsoft/deberta-v3-small
- Japanese: ku-nlp/deberta-v2-base-japanese-char-wwm
I suppose these variants won't cause any issues since the tokenizers are identical to the original ones
My hypothesis is that if we fine-tune our model with the new BERT models, it will seamlessly incorporate with the current system.
The text was updated successfully, but these errors were encountered: