Skip to content

Remove use of VLLM_USE_V1.#1083

Merged
QiliangCui merged 1 commit intomainfrom
fix-use-vllm
Nov 12, 2025
Merged

Remove use of VLLM_USE_V1.#1083
QiliangCui merged 1 commit intomainfrom
fix-use-vllm

Conversation

@QiliangCui
Copy link
Copy Markdown
Collaborator

@QiliangCui QiliangCui commented Nov 12, 2025

Description

Remove VLLM_USE_V1 because it is removed from vllm.main
vllm-project/vllm#28204 removed "VLLM_USE_V1".

FIXES: b/460101498

Tests

Manual launch

# 1 not set VLLM_USE_V1
TPU_BACKEND_TYPE=jax vllm serve Qwen/Qwen2.5-1.5B-Instruct \
 --seed 42 \
 --disable-log-requests \
 --max-num-seqs 512 \
 --max-num-batched-tokens 4096 \
 --tensor-parallel-size 1 \
 --no-enable-prefix-caching \
 --download_dir /mnt/disks/persist \
 --max-model-len 2048

vllm bench serve \
    --backend vllm \
    --model "Qwen/Qwen2.5-1.5B-Instruct" \
    --dataset-name "sonnet" \
    --dataset-path "benchmarks/sonnet_4x.txt" \
    --sonnet-input-len 1200 \
    --sonnet-output-len 1 \
    --num-prompts 1 \
    --request-rate inf \
    --percentile-metrics "ttft,tpot,itl,e2el" \
    --ignore-eos

# 2: still set VLLM_USE_V1=1 making sure it runs as backward compatible. 
VLLM_USE_V1=1 TPU_BACKEND_TYPE=jax vllm serve Qwen/Qwen2.5-1.5B-Instruct \
 --seed 42 \
 --disable-log-requests \
 --max-num-seqs 512 \
 --max-num-batched-tokens 4096 \
 --tensor-parallel-size 1 \
 --no-enable-prefix-caching \
 --download_dir /mnt/disks/persist \
 --max-model-len 2048

vllm bench serve \
    --backend vllm \
    --model "Qwen/Qwen2.5-1.5B-Instruct" \
    --dataset-name "sonnet" \
    --dataset-path "benchmarks/sonnet_4x.txt" \
    --sonnet-input-len 1200 \
    --sonnet-output-len 1 \
    --num-prompts 1 \
    --request-rate inf \
    --percentile-metrics "ttft,tpot,itl,e2el" \
    --ignore-eos

CIT

https://buildkite.com/tpu-commons/tpu-inference-ci/builds/5249

The two broken tests are tracked by other bug as known issue
b/460112990 and b/459771818

@github-actions
Copy link
Copy Markdown

Description

Start with a short description of what the PR does and how this is a change from
the past.

The rest of the description includes relevant details and context, examples:

  • why is this change being made,
  • the problem being solved and any relevant context,
  • why this is a good solution,
  • some information about the specific implementation,
  • shortcomings of the solution and possible future improvements.

If the change fixes a bug or a Github issue, please include a link, e.g.,:
FIXES: b/123456
FIXES: #123456

Tests

Please describe how you tested this change, and include any instructions and/or
commands to reproduce.

Checklist

Before submitting this PR, please make sure:

  • I have performed a self-review of my code.
  • I have necessary comments in my code, particularly in hard-to-understand areas.
  • I have made or will make corresponding changes to any relevant documentation.

Copy link
Copy Markdown
Collaborator

@wdhongtw wdhongtw left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM. Just a small indent issue and need some noqa comment (for ruff tool) if we want to minimize the line of changes.

@QiliangCui QiliangCui force-pushed the fix-use-vllm branch 2 times, most recently from 00441ac to 7f03966 Compare November 12, 2025 16:54
Copy link
Copy Markdown
Collaborator

@wdhongtw wdhongtw left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Signed-off-by: Qiliang Cui <derrhein@gmail.com>
@QiliangCui QiliangCui merged commit cb8734c into main Nov 12, 2025
3 checks passed
@wdhongtw wdhongtw deleted the fix-use-vllm branch April 7, 2026 09:53
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants