Skip to content

Commit 5f6b5c1

Browse files
committed
upd
Signed-off-by: Siyuan Fu <[email protected]>
1 parent bc289df commit 5f6b5c1

File tree

1 file changed

+4
-7
lines changed

1 file changed

+4
-7
lines changed

tests/moe/test_trtllm_gen_routed_fused_moe.py

Lines changed: 4 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -140,8 +140,6 @@ def test_trtllm_gen_routed_fused_moe(
140140
)
141141
w13_global_scale = 1.0
142142
w2_global_scale = 1.0
143-
bias13 = torch.randn(num_experts, intermediate_size * 2, device=device) * 10
144-
bias2 = torch.randn(num_experts, intermediate_size * 2, device=device) * 10
145143

146144
output1_scale_scalar = torch.tensor(
147145
[hidden_states_global_scale * w13_global_scale] * num_experts, device=device
@@ -160,13 +158,13 @@ def test_trtllm_gen_routed_fused_moe(
160158
hidden_states_scale,
161159
w13,
162160
w13_scale,
163-
bias13,
161+
None,
164162
None, # gemm1_alpha
165163
None, # gemm1_beta
166164
None, # gemm1_clamp_limit
167165
w2,
168166
w2_scale,
169-
bias2,
167+
None,
170168
output1_scale_scalar,
171169
output1_scale_gate_scalar,
172170
output2_scale_scalar,
@@ -209,19 +207,18 @@ def test_trtllm_gen_routed_fused_moe(
209207

210208
output = trtllm_fp4_block_scale_routed_moe(
211209
packed_tensor,
212-
expert_weights,
213210
None, # routing_bias
214211
hidden_states,
215212
hidden_states_scale,
216213
w13,
217214
w13_scale,
218-
bias13,
215+
None,
219216
None, # gemm1_alpha
220217
None, # gemm1_beta
221218
None, # gemm1_clamp_limit
222219
w2,
223220
w2_scale,
224-
bias2,
221+
None,
225222
output1_scale_scalar,
226223
output1_scale_gate_scalar,
227224
output2_scale_scalar,

0 commit comments

Comments
 (0)