From 9c0a5be7ae652e0c164555b671703e3af741f09e Mon Sep 17 00:00:00 2001 From: zhyncs Date: Mon, 8 Dec 2025 22:54:05 +0000 Subject: [PATCH] Revert "[Bug] fix not desired disable fused share experts caused by rocm logic (#14432)" This reverts commit 2ecee7571cdfe10a124ae313fe020166ad5c56b8. --- python/sglang/srt/models/deepseek_v2.py | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/python/sglang/srt/models/deepseek_v2.py b/python/sglang/srt/models/deepseek_v2.py index 9fdfbcb6a255..94e08789484e 100644 --- a/python/sglang/srt/models/deepseek_v2.py +++ b/python/sglang/srt/models/deepseek_v2.py @@ -3275,10 +3275,8 @@ def determine_num_fused_shared_experts( "Only Deepseek V3/R1 on NV-platform with capability >= 80 " "or AMD-platform with capability >= gfx942(MI30x) can use shared experts fusion optimization." ) - elif ( - get_moe_expert_parallel_world_size() > 1 - and _is_hip - and torch.cuda.get_device_capability("cuda") < (9, 4) + elif get_moe_expert_parallel_world_size() > 1 and ( + not _is_hip or torch.cuda.get_device_capability("cuda") < (9, 4) ): disable_reason = "Only Deepseek V3/R1 on AMD-platform with capability >= gfx942(MI30x) can use shared experts fusion optimization under expert parallelism." elif disable_reason is None and get_moe_a2a_backend().is_deepep():