Skip to content

[Bugfix][Model] Fix PixtralForConditionalGeneration LoRA#36963

Merged
vllm-bot merged 6 commits intomainfrom
fix-pixtral-lora
Mar 30, 2026
Merged

[Bugfix][Model] Fix PixtralForConditionalGeneration LoRA#36963
vllm-bot merged 6 commits intomainfrom
fix-pixtral-lora

Conversation

@jeejeelee
Copy link
Copy Markdown
Collaborator

@jeejeelee jeejeelee commented Mar 13, 2026

Purpose

Fix #34591

Test Plan

  • Testing with the real LoRA adapter

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request refactors the Pixtral vision encoder to utilize vLLM's parallel layers, which is a significant and well-executed change to enable tensor parallelism, quantization, and LoRA support. The code is mostly clean and consistent. I have one suggestion to improve the robustness of the weight loading logic for future compatibility with features like quantization.

Comment thread vllm/model_executor/models/pixtral.py
@jeejeelee jeejeelee marked this pull request as draft March 13, 2026 07:59
@jeejeelee jeejeelee marked this pull request as ready for review March 16, 2026 01:52
@mergify
Copy link
Copy Markdown
Contributor

mergify Bot commented Mar 19, 2026

This pull request has merge conflicts that must be resolved before it can be
merged. Please rebase the PR, @jeejeelee.

https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@mergify mergify Bot added the needs-rebase label Mar 19, 2026
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
@mergify mergify Bot removed the needs-rebase label Mar 27, 2026
@ywang96 ywang96 added the ready ONLY add when PR is ready to merge/full CI is needed label Mar 30, 2026
@ywang96
Copy link
Copy Markdown
Member

ywang96 commented Mar 30, 2026

I don't think this will break the base model but FYI @patrickvonplaten

@jeejeelee jeejeelee enabled auto-merge (squash) March 30, 2026 04:52
@vllm-bot vllm-bot merged commit ac30a83 into main Mar 30, 2026
59 of 61 checks passed
@vllm-bot vllm-bot deleted the fix-pixtral-lora branch March 30, 2026 06:59
neweyes pushed a commit to neweyes/vllm that referenced this pull request Mar 31, 2026
…t#36963)

Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Roger Wang <hey@rogerw.io>
Signed-off-by: neweyes <328719365@qq.com>
puririshi98 pushed a commit to puririshi98/vllm that referenced this pull request Apr 7, 2026
…t#36963)

Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Roger Wang <hey@rogerw.io>
Signed-off-by: Rishi Puri <riship@nvidia.com>
mtparet pushed a commit to blackfuel-ai/vllm that referenced this pull request Apr 9, 2026
…t#36963)

Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Co-authored-by: Roger Wang <hey@rogerw.io>
iboiko-habana pushed a commit to vllm-project/vllm-gaudi that referenced this pull request Apr 10, 2026
…xtral, MoE and Granite regressions (#1311)

## Summary
This PR fixes a set of regressions introduced by recent upstream changes
and observed in vLLM-Gaudi hourly validation.

The branch now includes:
- Pixtral HPUAttention projection path fix,
- MoE dispatch and method override alignment updates for fused MoE and
compressed tensors,
- unit test updates to match the new MoE runner API usage,
- fix hybrid model page size alignment for Granite 4.0-H.

## Related upstream PRs that introduced the regressions
- vllm-project/vllm#37234
- vllm-project/vllm#35153
- vllm-project/vllm#36963
- vllm-project/vllm#38960
- vllm-project/vllm#35326
- vllm-project/vllm#37467

---------

Signed-off-by: Paweł Olejniczak <pawelx.olejniczak@intel.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

bug Something isn't working ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[Bug]: Ministral-3 LoRA adapter fails silently

3 participants