-
Notifications
You must be signed in to change notification settings - Fork 2.1k
Closed
Description
Hi,
By now we may all have heard about the Llama model. However, it was train on a limited set of languages. I want to "transfer" the Llama knowledge to a new language that it wasn't trained on, say Korean. My thought is to expand the Llama token vocab (~32K) by adding Korean tokens, resulting a bigger token embedding map. Then, I will freeze all Llama layers except the word embedding layer and the LMHead layer, to train a CausalLM model on Korean text. I only have a 3090 GPU so I am thinking about peft/LORA.
My question: is there an example or guide to do this? Like setting LORA module targets to an Embedding Layer?
Thanks,
Metadata
Metadata
Assignees
Labels
No labels