Skip to content

Revert "moe_gating_top_k"#5512

Merged
wangxiyuan merged 1 commit intomainfrom
revert-5271-main
Dec 30, 2025
Merged

Revert "moe_gating_top_k"#5512
wangxiyuan merged 1 commit intomainfrom
revert-5271-main

Conversation

@zzzzwwjj
Copy link
Copy Markdown
Collaborator

@zzzzwwjj zzzzwwjj commented Dec 30, 2025

Reverts #5271

It breaks e2e test

@github-actions
Copy link
Copy Markdown
Contributor

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request reverts the moe_gating_top_k custom operator. The changes primarily consist of removing the implementation files for this operator. A key modification is in vllm_ascend/ops/fused_moe/experts_selector.py, where the removed custom operator is replaced with torch_npu.npu_moe_gating_top_k. This change also adjusts the logic to align with the native implementation, which appears to be a correct and beneficial update. The pull request is clean and serves its purpose of reverting the feature.

@wangxiyuan wangxiyuan merged commit 71f729a into main Dec 30, 2025
18 checks passed
@zzzzwwjj zzzzwwjj deleted the revert-5271-main branch December 30, 2025 07:06
fuzhihong699 pushed a commit to fuzhihong699/vllm-ascend that referenced this pull request Dec 31, 2025
Reverts vllm-project#5271

It breaks e2e test

- vLLM version: v0.13.0
- vLLM main:
vllm-project/vllm@45c1ca1

Signed-off-by: f00824209 <fuzhihong4@huawei.com>
845473182 pushed a commit to 845473182/vllm-ascend that referenced this pull request Dec 31, 2025
…to FIA_rebase

* 'main' of https://github.com/vllm-project/vllm-ascend: (88 commits)
  [1/N] Refactor nightly test structure (vllm-project#5479)
  Docs: Remove deprecated --task parameter for embedding models (vllm-project#5257)
  Revert "moe_gating_top_k" (vllm-project#5512)
  [Doc] Fix issue link for 0.12.0 (vllm-project#5500)
  [CI]update triton ascend version (vllm-project#5392)
  moe_gating_top_k (vllm-project#5271)
  [refactor] refactor model runner capture model (vllm-project#5230)
  Update corresponding vllm commit ID to 12 29 (vllm-project#5475)
  [Kernel]update csrc cmakelist for open-source cann (vllm-project#5458)
  [OP] add custom op aclnnMoeInitRoutingCustom (vllm-project#5251)
  [Refactor][EAGLE] 1/N delete __init__ in mtp_proposer (vllm-project#5176)
  [Refactor][Triton] Move reject sample triton kernels into ops/triton (vllm-project#5324)
  [Feature] support eager mode in model runner v2 (vllm-project#5210)
  [feature] fia support sliding windows (vllm-project#5239)
  Optimize some rejectsampler functions to make npu op launch non-blocking (vllm-project#4587)
  [Feature] Support to use fullgraph with eagle (vllm-project#5118)
  [EPLB][refactor] Modification of the initialization logic for expert_map and log2phy(depend on pr5285) (vllm-project#5311)
  [Refactor]6/N Extract common code of class AscendMLAImpl (vllm-project#5314)
  [Refactor] cache cos/sin in mla & remove parameter model in builder. (vllm-project#5277)
  update vllm pin to 12.27 (vllm-project#5412)
  ...
shenchuxiaofugui pushed a commit to shenchuxiaofugui/vllm-ascend that referenced this pull request Dec 31, 2025
Reverts vllm-project#5271

It breaks e2e test

- vLLM version: v0.13.0
- vLLM main:
vllm-project/vllm@45c1ca1
ZRJ026 pushed a commit to ZRJ026/vllm-ascend that referenced this pull request Feb 28, 2026
Reverts vllm-project#5271

It breaks e2e test

- vLLM version: v0.13.0
- vLLM main:
vllm-project/vllm@45c1ca1

Signed-off-by: zrj026 <zhangrunjiang026@gmail.com>
maoxx241 pushed a commit to maoxx241/vllm-ascend that referenced this pull request Mar 2, 2026
Reverts vllm-project#5271

It breaks e2e test

- vLLM version: v0.13.0
- vLLM main:
vllm-project/vllm@45c1ca1
ZRJ026 pushed a commit to ZRJ026/vllm-ascend that referenced this pull request Mar 4, 2026
Reverts vllm-project#5271

It breaks e2e test

- vLLM version: v0.13.0
- vLLM main:
vllm-project/vllm@45c1ca1

Signed-off-by: zrj026 <zhangrunjiang026@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants