Skip to content

[[V0 deprecation]]Remove VLLM_USE_V1 env#28204

Merged
mgoin merged 1 commit intovllm-project:mainfrom
wangxiyuan:remove_VLLM_USE_V1
Nov 12, 2025
Merged

[[V0 deprecation]]Remove VLLM_USE_V1 env#28204
mgoin merged 1 commit intovllm-project:mainfrom
wangxiyuan:remove_VLLM_USE_V1

Conversation

@wangxiyuan
Copy link
Copy Markdown
Contributor

@wangxiyuan wangxiyuan commented Nov 6, 2025

Purpose

This is the last PR to remove VLLM_USE_V1 env totally. Currently, there is no place to use this env and is safe to remote it.

Test Plan

Test Result


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Copy link
Copy Markdown

@chatgpt-codex-connector chatgpt-codex-connector bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codex Review

Here are some automated review suggestions for this pull request.

ℹ️ About Codex in GitHub

Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you

  • Open a pull request for review
  • Mark a draft as ready
  • Comment "@codex review".

If Codex has suggestions, it will comment; otherwise it will react with 👍.

When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request completes the removal of the deprecated VLLM_USE_V1 environment variable as part of the V0 deprecation. The changes are straightforward, correctly removing the variable's definition, its usage in tests, and its collection for usage statistics. The modifications are consistent and self-contained. I have reviewed the changes and found no issues of high or critical severity.

@mergify mergify bot added the ci/build label Nov 6, 2025
Copy link
Copy Markdown
Member

@DarkLight1337 DarkLight1337 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks

@DarkLight1337 DarkLight1337 enabled auto-merge (squash) November 6, 2025 12:10
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Nov 6, 2025
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
auto-merge was automatically disabled November 6, 2025 12:16

Head branch was pushed to by a user without write access

@mergify
Copy link
Copy Markdown

mergify bot commented Nov 6, 2025

Documentation preview: https://vllm--28204.org.readthedocs.build/en/28204/

@mergify mergify bot added documentation Improvements or additions to documentation qwen Related to Qwen models speculative-decoding kv-connector labels Nov 6, 2025
@wangxiyuan
Copy link
Copy Markdown
Contributor Author

wangxiyuan commented Nov 6, 2025

@DarkLight1337 For example part, I left mlpspeculator and qwen2.5-omni there instead of removing them, since that I notice there are some PRs to support them in V1 already. These examples can be useful then.

@wangxiyuan
Copy link
Copy Markdown
Contributor Author

@DarkLight1337 CI failure looks doesn't relate to this change.

@DarkLight1337
Copy link
Copy Markdown
Member

Retrying flaky test

Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It seems we should delete this example rather than allowing it to continue on, if V0 is required cc @ywang96

Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ditto, we should delete the example if it doesn't work on V1

@wangxiyuan
Copy link
Copy Markdown
Contributor Author

@mgoin I left them there since I notice there are some PRs to support it for V1. So maybe it can be reused later.
mlp_speculator: #21276
qwen2.5-omni: #26334

@mgoin mgoin merged commit e171039 into vllm-project:main Nov 12, 2025
50 of 52 checks passed
geodavic pushed a commit to geodavic/vllm that referenced this pull request Nov 16, 2025
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: George D. Torres <gdavtor@gmail.com>
devpatelio pushed a commit to SumanthRH/vllm that referenced this pull request Nov 29, 2025
Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build documentation Improvements or additions to documentation kv-connector qwen Related to Qwen models ready ONLY add when PR is ready to merge/full CI is needed speculative-decoding

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants