Skip to content

[MISC] Clean up useless env USE_OPTIMIZED_MODEL#6618

Merged
wangxiyuan merged 1 commit intovllm-project:mainfrom
wangxiyuan:cleanup_env
Feb 9, 2026
Merged

[MISC] Clean up useless env USE_OPTIMIZED_MODEL#6618
wangxiyuan merged 1 commit intovllm-project:mainfrom
wangxiyuan:cleanup_env

Conversation

@wangxiyuan
Copy link
Copy Markdown
Collaborator

@wangxiyuan wangxiyuan commented Feb 9, 2026

Clean up uesless env USE_OPTIMIZED_MODEL

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
@github-actions
Copy link
Copy Markdown
Contributor

github-actions Bot commented Feb 9, 2026

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

@gemini-code-assist
Copy link
Copy Markdown
Contributor

Summary of Changes

Hello @wangxiyuan, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request focuses on code cleanup by removing the USE_OPTIMIZED_MODEL environment variable, which was identified as obsolete. This change simplifies the environment configuration and eliminates unnecessary conditional logic from the multi-modal encoder attention mechanism, contributing to a cleaner and more maintainable system.

Highlights

  • Environment Variable Removal: The USE_OPTIMIZED_MODEL environment variable, along with its associated comments, has been completely removed from vllm_ascend/envs.py.
  • Dependency Cleanup: The mm_encoder_attention.py module no longer imports vllm_ascend.envs and its enable_pad logic has been updated to remove the dependency on the USE_OPTIMIZED_MODEL flag.
  • Code Standardization: A minor cleanup was performed in vllm_ascend/envs.py by standardizing the environment variable access for HCCL_SO_PATH from os.environ.get to os.getenv.

🧠 New Feature in Public Preview: You can now enable Memory to help Gemini Code Assist learn from your team's feedback. This makes future code reviews more consistent and personalized to your project's style. Click here to enable Memory in your admin console.

Changelog
  • vllm_ascend/envs.py
    • Removed the USE_OPTIMIZED_MODEL environment variable definition and its descriptive comments.
    • Standardized the retrieval of HCCL_SO_PATH to use os.getenv instead of os.environ.get.
  • vllm_ascend/ops/mm_encoder_attention.py
    • Removed the unused import statement for vllm_ascend.envs.
    • Modified the enable_pad condition within the forward_oot function to no longer reference the removed USE_OPTIMIZED_MODEL flag.
Activity
  • No specific activity (comments, reviews, or progress updates) has been recorded for this pull request yet.
Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request aims to clean up a supposedly useless environment variable, USE_OPTIMIZED_MODEL. It removes the variable definition and its usage in the codebase. While the cleanup is straightforward, I have a significant concern about removing the ability to disable the associated optimization, as the original comments suggest it might be necessary for certain use cases like RLHF training. This could be a breaking change for some users. A minor consistency improvement is also included.

I am having trouble creating individual review comments. Click here to see my feedback.

vllm_ascend/envs.py (69-72)

high

The removal of the USE_OPTIMIZED_MODEL environment variable could introduce a breaking change. The previous implementation allowed disabling an optimization that was noted as potentially unsuitable for use cases like RLHF training. By removing this flag, this control is lost. If this optimization is not universally safe, removing the flag could be a regression. Please either restore this flag or provide a clear justification in the pull request description for why it is no longer necessary.

@wangxiyuan wangxiyuan changed the title [MISC] Clean up useless env [MISC] Clean up useless env USE_OPTIMIZED_MODEL Feb 9, 2026
@wangxiyuan wangxiyuan merged commit 9c6d031 into vllm-project:main Feb 9, 2026
22 checks passed
845473182 pushed a commit to 845473182/vllm-ascend that referenced this pull request Feb 11, 2026
…to qwen3next_rebase

* 'main' of https://github.com/vllm-project/vllm-ascend:
  [Feat] 310p support MoE W8A8 quantizaition (vllm-project#6641)
  [TEST]add a qwen3-30b acc case with mooncake mempool (vllm-project#6244)
  [MOE Refactor] Remove QuantType in prepare_finalize.py (vllm-project#6534)
  [EPLB] Avoiding eplb's dependency on a specified model (vllm-project#6528)
  [Doc][Misc] Restructure tutorial documentation (vllm-project#6501)
  implement batch invariant with ascendc (vllm-project#6590)
  [Refact]Refact MLA/SFA weight prefetch to consist with moe weight prefetch (vllm-project#6629)
  [Misc] upgrade to vllm main (vllm-project#6646)
  [main][Docs] Fix spelling errors across documentation (vllm-project#6649)
  [bugfix]Fix no attribute 'data' when MLAPO is enable  (vllm-project#6601)
  [DOC]Add Memcache Usage Guide (vllm-project#6476)
  [main][bugfix] Fix spec acceptance rate problem in vllm_0.15.0 (vllm-project#6606)
  [Test][LoRA] Add e2e test for base model inference (vllm-project#6624)
  [refactor]Optimized the kvcache usage of Deepseek v3.2 (vllm-project#6610)
  [Feat](sfa,dcp) support dcp for sfa (vllm-project#6563)
  [BugFix] Add support for rotary_dim parameter when using partial rope in rotary_embedding (vllm-project#6581)
  [fix bug] fix tensor mismatch bug in sigmoid operate test case (vllm-project#6619)
  [Kernel]: Optimize DispatchFFNCombine performance (vllm-project#6468)
  [MISC] Clean up useless env USE_OPTIMIZED_MODEL (vllm-project#6618)
chenchuw886 pushed a commit to chenchuw886/vllm-ascend that referenced this pull request Feb 12, 2026
Clean up uesless env `USE_OPTIMIZED_MODEL`

- vLLM version: v0.15.0
- vLLM main:
vllm-project/vllm@d7e17aa

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: momochenchuw <chenchuw@huawei.com>
@wangxiyuan wangxiyuan mentioned this pull request Feb 24, 2026
ZRJ026 pushed a commit to ZRJ026/vllm-ascend that referenced this pull request Feb 28, 2026
Clean up uesless env `USE_OPTIMIZED_MODEL`

- vLLM version: v0.15.0
- vLLM main:
vllm-project/vllm@d7e17aa

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: zrj026 <zhangrunjiang026@gmail.com>
maoxx241 pushed a commit to maoxx241/vllm-ascend that referenced this pull request Mar 2, 2026
Clean up uesless env `USE_OPTIMIZED_MODEL`

- vLLM version: v0.15.0
- vLLM main:
vllm-project/vllm@d7e17aa

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
ZRJ026 pushed a commit to ZRJ026/vllm-ascend that referenced this pull request Mar 4, 2026
Clean up uesless env `USE_OPTIMIZED_MODEL`

- vLLM version: v0.15.0
- vLLM main:
vllm-project/vllm@d7e17aa

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Signed-off-by: zrj026 <zhangrunjiang026@gmail.com>
LCAIZJ pushed a commit to LCAIZJ/vllm-ascend that referenced this pull request Mar 7, 2026
Clean up uesless env `USE_OPTIMIZED_MODEL`

- vLLM version: v0.15.0
- vLLM main:
vllm-project/vllm@d7e17aa

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant