-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MultiheadAttention out_projection #3
Comments
Yep, I also find this question. It is because the |
Yep, so |
Sounds like a good idea. I will try to fix it (maybe after two weeks, i am busy with some ddls currently). |
did you manage to fix it? |
P.S. thanks for this implementation! |
If I set enable_lora: list = ['q', 'k', 'v','o'], the problem mentioned in #7 still exists. This may be due to the need to pass the with_nn parameter during the recursive calls. |
@mounchiliu Thanks for your suggestion. I have fixed this problem. |
@marcomistretta @ghost Sorry for the late reply. The LoRA of out_proj is not updated because of the wrong init of the LoRA rather than the use of NonDynamicallyQuantizableLinear. I have fixed this problem. |
Hello,
thanks for this implementation - very useful.
I had a question regarding
MultiheadAttention
class - it seems likeout_proj.weight
is not updated or I am missing something?Thanks!
The text was updated successfully, but these errors were encountered: