Skip to content

Commit e0efad8

Browse files
committed
fix conflict
Signed-off-by: jiahanc <[email protected]>
2 parents 1b12e88 + c9e42ca commit e0efad8

File tree

3 files changed

+20
-4
lines changed

3 files changed

+20
-4
lines changed

csrc/trtllm_fused_moe_kernel_launcher.cu

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -355,6 +355,7 @@ void trtllm_fp8_block_scale_moe_launcher(
355355
<< "routing_bias has incorrect shape.";
356356
}
357357

358+
<<<<<<< HEAD
358359
if (n_group.has_value() && n_group.value() != 0) {
359360
TVM_FFI_ICHECK(static_cast<RoutingMethodType>(routing_method_type) ==
360361
RoutingMethodType::DeepSeekV3)
@@ -382,6 +383,20 @@ void trtllm_fp8_block_scale_moe_launcher(
382383
TVM_FFI_ICHECK_EQ(top_k, 1)
383384
<< "Current routing kernel (no groups, Llama4) only supports top_k=1.";
384385
}
386+
=======
387+
// if (n_group <= 0 || topk_group <= 0) {
388+
// TVM_FFI_ICHECK_EQ(top_k, 1) << "Current routing kernel (no groups) only supports top_k=1.";
389+
// } else {
390+
// TVM_FFI_ICHECK_LE(top_k, 8) << "Current routing kernel (with groups) only supports
391+
// top_k<=8."; TVM_FFI_ICHECK_LE(topk_group, 4)
392+
// << "Current routing kernel (with groups) only supports topk_group<=4.";
393+
// TVM_FFI_ICHECK_LE(topk_group, n_group) << "n_group must not be smaller than topk_group.";
394+
// TVM_FFI_ICHECK_EQ(num_experts % n_group, 0) << "num_experts must be divisible by n_group";
395+
// // This check ensures we have enough experts in the selected groups to handle the top_k
396+
// routing TVM_FFI_ICHECK_LT(top_k, (topk_group * num_experts / n_group))
397+
// << "top_k must be less than total number of experts in selected groups";
398+
// }
399+
>>>>>>> c9e42cacef0506c57777d8d8efcf859219529951
385400
TVM_FFI_ICHECK_EQ(num_experts % 4, 0)
386401
<< "Routing kernel expects that num_experts must be divisible by 4";
387402
TVM_FFI_ICHECK_GT(num_experts, top_k) << "num_experts must be greater than top_k";

csrc/trtllm_fused_moe_routing_renormalize.cu

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -464,6 +464,7 @@ void run(Data const& data, void* stream) {
464464
}
465465
cudaDeviceSynchronize();
466466
cudaError_t result = cudaGetLastError();
467+
std::cout << "cudaGetLastError: " << cudaGetErrorString(result) << std::endl;
467468
}
468469

469470
// void run(Data const& data, void* stream) {

tests/moe/test_trtllm_gen_fused_moe.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1836,9 +1836,9 @@ def cache_permute_indices():
18361836
return _cache_permute_indices
18371837

18381838

1839-
@pytest.mark.parametrize("num_tokens", [1, 8, 1024,512])
1839+
@pytest.mark.parametrize("num_tokens", [1, 8, 1024])
18401840
@pytest.mark.parametrize("hidden_size", [1024, 2048, 8192])
1841-
@pytest.mark.parametrize("intermediate_size", [2048, 1024, 768, 384, 512])
1841+
@pytest.mark.parametrize("intermediate_size", [2048, 1024, 5120, 768, 384])
18421842
@pytest.mark.parametrize(
18431843
"moe_impl",
18441844
[
@@ -1905,8 +1905,8 @@ def cache_permute_indices():
19051905
),
19061906
pytest.param(
19071907
{
1908-
"num_experts": 256,
1909-
"top_k": 8,
1908+
"num_experts": 512,
1909+
"top_k": 10,
19101910
"padding": 8,
19111911
"n_groups": None,
19121912
"top_k_groups": None,

0 commit comments

Comments
 (0)