File tree Expand file tree Collapse file tree 1 file changed +4
-7
lines changed Expand file tree Collapse file tree 1 file changed +4
-7
lines changed Original file line number Diff line number Diff line change @@ -140,8 +140,6 @@ def test_trtllm_gen_routed_fused_moe(
140140 )
141141 w13_global_scale = 1.0
142142 w2_global_scale = 1.0
143- bias13 = torch .randn (num_experts , intermediate_size * 2 , device = device ) * 10
144- bias2 = torch .randn (num_experts , intermediate_size * 2 , device = device ) * 10
145143
146144 output1_scale_scalar = torch .tensor (
147145 [hidden_states_global_scale * w13_global_scale ] * num_experts , device = device
@@ -160,13 +158,13 @@ def test_trtllm_gen_routed_fused_moe(
160158 hidden_states_scale ,
161159 w13 ,
162160 w13_scale ,
163- bias13 ,
161+ None ,
164162 None , # gemm1_alpha
165163 None , # gemm1_beta
166164 None , # gemm1_clamp_limit
167165 w2 ,
168166 w2_scale ,
169- bias2 ,
167+ None ,
170168 output1_scale_scalar ,
171169 output1_scale_gate_scalar ,
172170 output2_scale_scalar ,
@@ -209,19 +207,18 @@ def test_trtllm_gen_routed_fused_moe(
209207
210208 output = trtllm_fp4_block_scale_routed_moe (
211209 packed_tensor ,
212- expert_weights ,
213210 None , # routing_bias
214211 hidden_states ,
215212 hidden_states_scale ,
216213 w13 ,
217214 w13_scale ,
218- bias13 ,
215+ None ,
219216 None , # gemm1_alpha
220217 None , # gemm1_beta
221218 None , # gemm1_clamp_limit
222219 w2 ,
223220 w2_scale ,
224- bias2 ,
221+ None ,
225222 output1_scale_scalar ,
226223 output1_scale_gate_scalar ,
227224 output2_scale_scalar ,
You can’t perform that action at this time.
0 commit comments