Skip to content

Commit 4e3a7fe

Browse files
committed
Update fused_moe.py
1 parent 633f013 commit 4e3a7fe

File tree

1 file changed

+0
-12
lines changed

1 file changed

+0
-12
lines changed

vllm_ascend/ops/fused_moe.py

Lines changed: 0 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -197,18 +197,6 @@ def apply(
197197
shared_experts=shared_experts,
198198
need_trans=True)
199199

200-
# return unified_fused_experts_eager(hidden_states=x,
201-
# w1=layer.w13_weight,
202-
# w2=layer.w2_weight,
203-
# topk_weights=topk_weights,
204-
# topk_ids=topk_ids,
205-
# row_idx=row_idx,
206-
# expert_map=expert_map,
207-
# shared_experts=shared_experts,
208-
# mc2_mask=kwargs.get(
209-
# "mc2_mask", None),
210-
# with_quant=False)
211-
212200

213201
class AscendFusedMoE(FusedMoE):
214202

0 commit comments

Comments
 (0)