-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Using llama.cpp #14
Comments
Hi @PrithivirajDamodaran! Sorry I missed your last issue. Forgot to turn on notifications for this repo. For everyday stuff I just use the As for the MLM issue. Right now, |
Hey @iamlemec , Thanks for taking the time out and all the awesome work you are doing.
As we speak I am working on a fork for the community to take full advantage of all the awesome work that been done in this space🙏. Will share more soon. Cheers |
@PrithivirajDamodaran Looks cool! Yeah, so I haven't done benchmarks in a bit, but the main reason it should be faster is that Yup, pooling options are great, epecially with some of the new approaches coming out like GritLM. |
I am trying to use llama.cpp as you suggested its merged there for the same baai 1.5 embedding models , could you please help me how should I get started. I cant figure out the equivalent of bert_tokenize part there.
Thanks
The text was updated successfully, but these errors were encountered: