Skip to content

[Bugfix] Fix Basic Models Test#34818

Merged
vllm-bot merged 26 commits intovllm-project:mainfrom
MatthewBonanni:fix_basic_extra_init
Feb 19, 2026
Merged

[Bugfix] Fix Basic Models Test#34818
vllm-bot merged 26 commits intovllm-project:mainfrom
MatthewBonanni:fix_basic_extra_init

Conversation

@MatthewBonanni
Copy link
Collaborator

@MatthewBonanni MatthewBonanni commented Feb 18, 2026

Purpose

Fixes #34806, #34810, #34814, and #34819

Test Plan

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
@mergify mergify bot added speculative-decoding bug Something isn't working labels Feb 18, 2026
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request introduces two small but useful changes. The first is an update to the model registry in tests/models/registry.py to provide a more specific reason for skipping tests related to H2OVLChatModel, which will improve test stability. The second change, in vllm/model_executor/models/minicpm_eagle.py, explicitly disallows the use of inputs_embeds for the EagleMiniCPMForCausalLM model by raising a NotImplementedError. This is a good defensive measure to prevent incorrect usage of the model and provides a clear error message. Both changes are correct and improve the robustness of the codebase. I approve these changes.

@mgoin mgoin added ready ONLY add when PR is ready to merge/full CI is needed ci-failure Issue about an unexpected test failure in CI labels Feb 18, 2026
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
@DarkLight1337
Copy link
Member

Hmm this simply skips the tests, which we didn't have to do before #33600

@MatthewBonanni MatthewBonanni changed the title [Bugfix] Fix Basic Models Test (Extra Initialization) [Bugfix] Fix Basic Models Test Feb 18, 2026
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
@MatthewBonanni
Copy link
Collaborator Author

MatthewBonanni commented Feb 18, 2026

@DarkLight1337 good point - I changed this in 6ac74b8

@DarkLight1337 DarkLight1337 requested a review from mgoin February 18, 2026 17:29
@DarkLight1337
Copy link
Member

DarkLight1337 commented Feb 18, 2026

Extra Initialization tests still fail

@mgoin
Copy link
Member

mgoin commented Feb 18, 2026

The extra init tests seem to be failing because of OOM, so unsure of clear debug

@DarkLight1337
Copy link
Member

DarkLight1337 commented Feb 18, 2026

Might it be because #33600 caused the overrides/patches specific to the model initialization tests to not function correctly?

@MatthewBonanni
Copy link
Collaborator Author

@DarkLight1337 yes, I believe it's because _update_block_size_for_backend causes CUDA initialization

Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
@MatthewBonanni
Copy link
Collaborator Author

MatthewBonanni commented Feb 18, 2026

As a workaround, I've decided to just only run _update_block_size_for_backend for MLA models. This was the behavior pre-33600. All standard attention backends support the default block size of 16. This avoids importing attention backends that initialize CUDA.

After this change, if the user sets a bad block size for a standard attention model, the attention backend selection won't factor this in. This was the pre-33600 behavior, though, so it should be alright.

I'll work on a fix that lets us get rid of the use_mla check in a follow up

EDIT: This workaround ended up not being any simpler than the full fix

Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Copy link
Member

@mgoin mgoin left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The compromise is good with me to fix CI, thanks Matt

@github-project-automation github-project-automation bot moved this to Ready in NVIDIA Feb 18, 2026
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
@vllm-bot vllm-bot merged commit 662205d into vllm-project:main Feb 19, 2026
64 of 67 checks passed
@github-project-automation github-project-automation bot moved this from Ready to Done in NVIDIA Feb 19, 2026
mgoin added a commit to neuralmagic/vllm that referenced this pull request Feb 20, 2026
Signed-off-by: mgoin <mgoin64@gmail.com>
LucasWilkinson added a commit that referenced this pull request Feb 20, 2026
jmamou pushed a commit to jmamou/vllm that referenced this pull request Feb 23, 2026
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Co-authored-by: Robert Shaw <114415538+robertgshaw2-redhat@users.noreply.github.com>
MatthewBonanni added a commit to MatthewBonanni/vllm that referenced this pull request Feb 23, 2026
commit 4f9b8be
Author: mgoin <mgoin64@gmail.com>
Date:   Fri Feb 20 17:14:56 2026 +0000

    Cleanup

    Signed-off-by: mgoin <mgoin64@gmail.com>

commit feed637
Author: mgoin <mgoin64@gmail.com>
Date:   Fri Feb 20 17:08:37 2026 +0000

    Fix block_size mismatch for MLA models after vllm-project#34818

    Signed-off-by: mgoin <mgoin64@gmail.com>

Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
llsj14 pushed a commit to llsj14/vllm that referenced this pull request Mar 1, 2026
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Co-authored-by: Robert Shaw <114415538+robertgshaw2-redhat@users.noreply.github.com>
tunglinwood pushed a commit to tunglinwood/vllm that referenced this pull request Mar 4, 2026
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Co-authored-by: Robert Shaw <114415538+robertgshaw2-redhat@users.noreply.github.com>
sharvil10 pushed a commit to sharvil10/vllm that referenced this pull request Mar 4, 2026
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Co-authored-by: Robert Shaw <114415538+robertgshaw2-redhat@users.noreply.github.com>
askliar pushed a commit to askliar/vllm that referenced this pull request Mar 9, 2026
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Co-authored-by: Robert Shaw <114415538+robertgshaw2-redhat@users.noreply.github.com>
Signed-off-by: Andrii Skliar <askliar@nvidia.com>
Copilot AI pushed a commit to machov/vllm that referenced this pull request Mar 10, 2026
Signed-off-by: Matthew Bonanni <mbonanni@redhat.com>
Co-authored-by: Robert Shaw <114415538+robertgshaw2-redhat@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

bug Something isn't working ci-failure Issue about an unexpected test failure in CI multi-modality Related to multi-modality (#4194) nvidia ready ONLY add when PR is ready to merge/full CI is needed speculative-decoding v1

Projects

Status: Done
Status: Done

Development

Successfully merging this pull request may close these issues.

[CI Failure]: models/test_initialization.py::test_can_initialize_large_subset[EagleMiniCPMForCausalLM]

7 participants