Skip to content

Revert "feat(moe): Add is_act_and_mul=False support for Triton MoE kernels"#31978

Merged
vllm-bot merged 1 commit intomainfrom
revert-31645-is_act_and_mul
Jan 8, 2026
Merged

Revert "feat(moe): Add is_act_and_mul=False support for Triton MoE kernels"#31978
vllm-bot merged 1 commit intomainfrom
revert-31645-is_act_and_mul

Conversation

@mgoin
Copy link
Copy Markdown
Member

@mgoin mgoin commented Jan 8, 2026

Reverts #31645, my reasoning is here #31645 (comment)

@mgoin mgoin added the ready ONLY add when PR is ready to merge/full CI is needed label Jan 8, 2026
Copy link
Copy Markdown
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request reverts the is_act_and_mul=False feature for Triton MoE kernels. The changes are mostly correct, but the revert appears to be incomplete. I've identified dead code in vllm/model_executor/layers/fused_moe/layer.py that should be removed to finalize the revert. Additionally, to fully remove the feature, the is_act_and_mul attribute should also be removed from FusedMoEConfig in vllm/model_executor/layers/fused_moe/config.py.

Comment on lines +603 to 606
if not current_platform.is_cuda():
raise NotImplementedError(
"is_act_and_mul=False is supported only for CUDA, or ROCm "
"(when AITER MoE is disabled) for now"
"is_act_and_mul=False is supported only for CUDA for now"
)
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

critical

The surrounding if not self.moe_config.is_act_and_mul: block (starting at line 584) is now dead code because this PR reverts the is_act_and_mul=False feature. With this revert, self.moe_config.is_act_and_mul will always be True. To complete the revert and improve code clarity, this entire if block (lines 584-607) should be removed. As a follow-up, the is_act_and_mul attribute should also be removed from FusedMoEConfig in vllm/model_executor/layers/fused_moe/config.py.

Copy link
Copy Markdown
Collaborator

@tjtanaa tjtanaa left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you for bringing this up. I didn't notice there is a parallel PR.

@tjtanaa tjtanaa added the rocm Related to AMD ROCm label Jan 8, 2026
@vllm-bot vllm-bot merged commit 87e07a6 into main Jan 8, 2026
59 of 60 checks passed
@vllm-bot vllm-bot deleted the revert-31645-is_act_and_mul branch January 8, 2026 19:31
yugong333 pushed a commit to yugong333/vllm that referenced this pull request Jan 9, 2026
akh64bit pushed a commit to akh64bit/vllm that referenced this pull request Jan 16, 2026
dsuhinin pushed a commit to dsuhinin/vllm that referenced this pull request Jan 21, 2026
…rnels" (vllm-project#31978)

Signed-off-by: dsuhinin <suhinin.dmitriy@gmail.com>
ItzDEXX pushed a commit to ItzDEXX/vllm that referenced this pull request Feb 19, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ready ONLY add when PR is ready to merge/full CI is needed rocm Related to AMD ROCm

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants