UPSTREAM PR #18607: model : add LFM2-ColBert-350M#820
Open
Conversation
|
Explore the complete analysis inside the Version Insights Perfect! I've successfully generated a comprehensive summary report for your project. The report shows that Pull Request #820 for the llama.cpp repository demonstrates significant performance improvements across the board: Key Highlights:✨ Top Performance Gains:
🎯 Main Impact Areas:
📈 Recommendation: This PR shows excellent performance improvements and should be strongly considered for merging, pending functional verification. The report includes detailed tables showing the top 10 functions by both response time and throughput improvements, along with analysis and recommendations for next steps. |
f85d458 to
67c372e
Compare
048ad94 to
6c1fde6
Compare
ef7afbe to
d4c3480
Compare
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Mirrored from ggml-org/llama.cpp#18607
PR adds support for LFM2-ColBert-350M by introducing
n_embd_out- a separate output embedding dimension that can differ from the input embedding dimension (n_embd).Initially, I introduced
LLAMA_POOLING_TYPE_TOKEN, which was applyingcls_outand outputting all embedding, but then switched ton_embd_out.n_embd_outwill be used in future multimodal models as well.New GGUF key and API:
LLM_KV_EMBEDDING_LENGTH_OUT- stores output embedding dimensionhparams.n_embd_outif set and fallbacks tohparams.n_embdTesting
Convert
Launch server
Run the attached Python script
rerank.py
cc: @ngxson