Skip to content

Commit 0b31d41

Browse files
committed
syntax error fixes
Signed-off-by: Neta Zmora <[email protected]>
1 parent 90e748e commit 0b31d41

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

tensorrt_llm/_torch/auto_deploy/transform/library/fused_moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -710,7 +710,7 @@ def get_param_or_buffer(target):
710710
# Create new node with get_attr for stacked parameters
711711
with graph.inserting_before(node):
712712
new_node = graph.call_function(
713-
torch.ops.auto_deploy.trtllm_quant_fp8moe_fused,
713+
torch.ops.auto_deploy.trtllm_quant_fp8_moe_fused,
714714
args=(
715715
hidden_states,
716716
selected_experts,

tests/unittest/_torch/auto_deploy/unit/singlegpu/custom_ops/test_trtllm_moe.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -376,7 +376,7 @@ def dequantize_weights(w31_weight, w2_weight, w31_scales, w2_scales):
376376

377377
activation_type = _activation_type_from_str(activation_func)
378378

379-
ad_test_output = torch.ops.auto_deploy.trtllm_quant_fp8moe_fused(
379+
ad_test_output = torch.ops.auto_deploy.trtllm_quant_fp8_moe_fused(
380380
x, # Note! unquantized input is expected
381381
selected_experts.to(torch.int),
382382
routing_weights,

0 commit comments

Comments
 (0)