Skip to content

[Model] Introduce verify_and_update_model_config for VerifyAndUpdateConfig.#31131

Merged
noooop merged 14 commits intovllm-project:mainfrom
noooop:verify_and_update_model_config
Dec 24, 2025
Merged

[Model] Introduce verify_and_update_model_config for VerifyAndUpdateConfig.#31131
noooop merged 14 commits intovllm-project:mainfrom
noooop:verify_and_update_model_config

Conversation

@noooop
Copy link
Collaborator

@noooop noooop commented Dec 22, 2025

Purpose

VerifyAndUpdateConfig.verify_and_update_config too late to _set_default_chunked_prefill_and_prefix_caching_args.
Causing the logs to not match the actual behavior, and requiring the use of some non-obvious failback to get the correct results (which sounds like a bug)

e.g.
nvidia/llama-nemotron-embed-1b-v2
nvidia/llama-nemotron-rerank-1b-v2
google/embeddinggemma-300m

Let's introduce verify_and_update_model_config for VerifyAndUpdateConfig to make the logic here clearer.

Test Plan

tests/models/language/pooling_mteb_test/
tests/entrypoints/pooling/

Test Result

pass


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Signed-off-by: wang.yuqi <yuqi.wang@daocloud.io>
@mergify mergify bot added the qwen Related to Qwen models label Dec 22, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request refactors the model configuration update mechanism to allow for earlier updates. It introduces a new method verify_and_update_model_config on ModelConfig and migrates several model-specific configuration classes to use it. The changes also include simplifying ModelInfo classes in tests by removing default_pooling_type.

While the refactoring is a good idea, it is incomplete. Several configuration classes in vllm/model_executor/models/config.py have not been migrated to the new mechanism. This will lead to lost functionality for the corresponding models if the old configuration update path is removed. I've left a critical comment detailing the issue and suggesting potential solutions.

Signed-off-by: wang.yuqi <yuqi.wang@daocloud.io>
Signed-off-by: wang.yuqi <yuqi.wang@daocloud.io>
Signed-off-by: wang.yuqi <yuqi.wang@daocloud.io>
Signed-off-by: wang.yuqi <yuqi.wang@daocloud.io>
@noooop noooop changed the title [Model] Add verify_and_update_model_config for VerifyAndUpdateConfig. [Model] Introduce verify_and_update_model_config for VerifyAndUpdateConfig. Dec 23, 2025
noooop and others added 3 commits December 24, 2025 10:29
Signed-off-by: wang.yuqi <noooop@126.com>
Signed-off-by: wang.yuqi <noooop@126.com>
@mergify mergify bot added the llama Related to Llama models label Dec 24, 2025
noooop and others added 3 commits December 24, 2025 12:06
Signed-off-by: wang.yuqi <noooop@126.com>
Signed-off-by: wang.yuqi <noooop@126.com>
@chatgpt-codex-connector
Copy link

Codex usage limits have been reached for code reviews. Please check with the admins of this repo to increase the limits by adding credits.

@DarkLight1337 DarkLight1337 added the ready ONLY add when PR is ready to merge/full CI is needed label Dec 24, 2025
@noooop noooop enabled auto-merge (squash) December 24, 2025 08:43
@noooop noooop merged commit bd89ce1 into vllm-project:main Dec 24, 2025
59 of 60 checks passed
@noooop noooop deleted the verify_and_update_model_config branch December 24, 2025 09:58
yiliu30 pushed a commit to yiliu30/vllm-fork that referenced this pull request Dec 30, 2025
…onfig. (vllm-project#31131)

Signed-off-by: wang.yuqi <yuqi.wang@daocloud.io>
Signed-off-by: wang.yuqi <noooop@126.com>
dsuhinin pushed a commit to dsuhinin/vllm that referenced this pull request Jan 21, 2026
…onfig. (vllm-project#31131)

Signed-off-by: wang.yuqi <yuqi.wang@daocloud.io>
Signed-off-by: wang.yuqi <noooop@126.com>
Signed-off-by: dsuhinin <suhinin.dmitriy@gmail.com>
ItzDEXX pushed a commit to ItzDEXX/vllm that referenced this pull request Feb 19, 2026
…onfig. (vllm-project#31131)

Signed-off-by: wang.yuqi <yuqi.wang@daocloud.io>
Signed-off-by: wang.yuqi <noooop@126.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

llama Related to Llama models qwen Related to Qwen models ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants