[V0 Deprecation] Remove VLLM_USE_V1 from docs and scripts#26336
[V0 Deprecation] Remove VLLM_USE_V1 from docs and scripts#26336DarkLight1337 merged 1 commit intovllm-project:mainfrom
VLLM_USE_V1 from docs and scripts#26336Conversation
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
There was a problem hiding this comment.
Code Review
This pull request is a good step towards deprecating the V0 engine by removing the VLLM_USE_V1 environment variable from various documentation files and scripts. The changes are correct and align with the goal of making V1 the default engine.
However, the deprecation appears to be incomplete. The core Python logic for selecting the engine version and falling back to V0 still exists in the codebase, for instance in vllm/engine/arg_utils.py. This could be misleading, as it suggests V0 is still a fallback option. For example, the _raise_or_fallback function still contains the message "Falling back to V0".
To complete the deprecation, I recommend creating a follow-up pull request to remove this V0/V1 switching logic. This would likely involve simplifying create_engine_config to always use the V1 engine and updating _is_v1_supported_oracle to raise errors for unsupported features instead of attempting a fallback.
The changes within this pull request are correct for its stated scope, and I have no further comments on the modified files.
hmellor
left a comment
There was a problem hiding this comment.
LGTM, one nit about TPU test naming
There was a problem hiding this comment.
Maybe also rename these files to remove the V1? Also the job names in test pipeline?
There was a problem hiding this comment.
I think that is just to keep the same naming convension as the "V1 Tests" we have now. We can move them in another PR.
…ject#26336) Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk> Signed-off-by: Dhruvil Bhatt <bhattdbh@amazon.com>
…ject#26336) Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
…ject#26336) Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
…ject#26336) Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
…ject#26336) Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
Purpose
Remove
VLLM_USE_V1flag from all docs and scripts, apart from those that don't support V1:run-cpu-test.sh: There is a commented-out script that uses V0. cc @Isotr0py @bigPYJ1151examples/offline_inference/mlpspeculator.py: Not sure who to ping for this. Maybe @WoosukKwon ?examples/offline_inference/qwen2_5_omni: We are working on this in [MM][Feat] Add support for audio in video in Qwen2.5-Omni #26334examples/others/lmcache/cpu_offload_lmcache.py: cc @robertgshaw2-redhatTest Plan
Test Result
Essential Elements of an Effective PR Description Checklist
supported_models.mdandexamplesfor a new model.