-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Add qwen2.5 vl #2995
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Add qwen2.5 vl #2995
Conversation
|
Nice work, does the result verified? Also, would consider add quantization support maybe |
|
It's still on work though.. |
|
Oh. no |
|
Nice Work. This is definitely needed. But I see that the model uses conv3d which is not there yet in Candle. How are you planning to handle that? |
|
@akshayballal95 I’ve been struggling this for a while. While flattening the temporal dimension and using Conv2D (reshaping [B, T, C, H, W] to [B*T, C, H, W]) may be possible, it changes the computation and is incompatible with pretrained weights. Implementing native Conv3D support in Candle seems to be the proper solution. |
|
Well, we have been asking for it |
|
@maximizemaxwell you can look at: https://github.com/EricLBuehler/mistral.rs/blob/4608202c128da44b84157573dbc8ff1a1146f64c/mistralrs-core/src/layers.rs#L1965-L2036 This is written under the assumption that the temporal patch size== |
|
Gosh I didn’t realize there was already an issue — I just created a new one with the same content, thinking it didn’t exist. |
|
We need someone to review these features as already implemented in the community. Please merge it into candle so that other people could bring more broad model support to candle! |
What does this PR do?
Add support for Qwen-2.5-VL
Part of Issue
#2814