Skip to content

[Deprecation] Remove deprecated environment variables#32812

Merged
ProExpertProg merged 2 commits intomainfrom
wentao-remove-deprecated-envs
Jan 22, 2026
Merged

[Deprecation] Remove deprecated environment variables#32812
ProExpertProg merged 2 commits intomainfrom
wentao-remove-deprecated-envs

Conversation

@yewentao256
Copy link
Member

@yewentao256 yewentao256 commented Jan 21, 2026

Purpose

As v0.14.0 has been out, we can remove these deprecated envs.

CC: @MatthewBonanni

Signed-off-by: yewentao256 <zhyanwentao@126.com>
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request correctly removes several deprecated environment variables related to attention configuration. The changes are consistent across the codebase, with updates to configuration classes, environment variable definitions, and direct usages in tests and platform-specific code. The migration from environment variables to fields within the AttentionConfig dataclass is well-executed and improves configuration management. I found no high or critical issues in these changes.

Copy link

@cursor cursor bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Cursor Bugbot has reviewed your changes and found 1 potential issue.

Bugbot Autofix is OFF. To automatically fix reported issues with Cloud Agents, enable Autofix in the Cursor dashboard.

Comment @cursor review or bugbot run to trigger another review on this PR

if envs.VLLM_V1_USE_PREFILL_DECODE_ATTENTION:
from vllm.config import get_current_vllm_config

vllm_config = get_current_vllm_config()
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you make sure that this doesn't cause Current vLLM config is not set. log spam?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

There is no log spam now, raise AssertionError()

def get_current_vllm_config() -> VllmConfig:
    if _current_vllm_config is None:
        raise AssertionError(
            "Current vLLM config is not set. This typically means "

Copy link
Collaborator

@MatthewBonanni MatthewBonanni left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for doing this! Can you make sure to double check that you got everything? Looks like there are some uses in test-amd.yaml, e.g.

- VLLM_ATTENTION_BACKEND=ROCM_ATTN bash v1/kv_connector/nixl_integration/config_sweep_accuracy_test.sh

Signed-off-by: yewentao256 <zhyanwentao@126.com>
@yewentao256
Copy link
Member Author

Thanks for doing this! Can you make sure to double check that you got everything? Looks like there are some uses in test-amd.yaml, e.g.

- VLLM_ATTENTION_BACKEND=ROCM_ATTN bash v1/kv_connector/nixl_integration/config_sweep_accuracy_test.sh

OK everything checked, thanks!

Copy link
Collaborator

@ProExpertProg ProExpertProg left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Beautiful

@ProExpertProg ProExpertProg enabled auto-merge (squash) January 22, 2026 00:23
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Jan 22, 2026
@ProExpertProg ProExpertProg merged commit 6437ff1 into main Jan 22, 2026
53 checks passed
@ProExpertProg ProExpertProg deleted the wentao-remove-deprecated-envs branch January 22, 2026 02:25
commands:
- uv pip install --system -r /vllm-workspace/requirements/kv_connectors_rocm.txt
- VLLM_ATTENTION_BACKEND=ROCM_ATTN bash v1/kv_connector/nixl_integration/config_sweep_accuracy_test.sh
- bash v1/kv_connector/nixl_integration/config_sweep_accuracy_test.sh --attention-backend ROCM_ATTN
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Does this actually work? It doesn't seem that the extra arguments are forwarded through the bash script

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Copy link
Contributor

@mawong-amd mawong-amd Jan 22, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It doesn't. This PR #32837 fixes it, along with another error caused by the deprecation of VLLM_V1_USE_PREFILL_DECODE_ATTENTION

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Does this actually work? It doesn't seem that the extra arguments are forwarded through the bash script

They indeed are forwarded. Thanks for the cc btw :)

monajafi-amd pushed a commit to monajafi-amd/vllm that referenced this pull request Jan 23, 2026
…2812)

Signed-off-by: yewentao256 <zhyanwentao@126.com>
Signed-off-by: mohammad najafi <mohammad.najafi@amd.com>
cwazai pushed a commit to cwazai/vllm that referenced this pull request Jan 25, 2026
…2812)

Signed-off-by: yewentao256 <zhyanwentao@126.com>
Signed-off-by: 陈建华 <1647430658@qq.com>
@NickLucche
Copy link
Collaborator

how's one supposed to set FA version manually now that VLLM_FLASH_ATTN_VERSION is gone?

@MatthewBonanni
Copy link
Collaborator

how's one supposed to set FA version manually now that VLLM_FLASH_ATTN_VERSION is gone?

You can use --attention-config.flash_attn_version in the CLI, or pass attention_config={'flash_attn_version': 2} in the LLM() API.

lapy pushed a commit to lapy/vllm that referenced this pull request Jan 27, 2026
ItzDEXX pushed a commit to ItzDEXX/vllm that referenced this pull request Feb 19, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build ready ONLY add when PR is ready to merge/full CI is needed rocm Related to AMD ROCm speculative-decoding v1

Projects

None yet

Development

Successfully merging this pull request may close these issues.

7 participants