Skip to content

Commit ef6d38a

Browse files
committed
fix bug of using fp4 on sm120
Signed-off-by: bhsueh <[email protected]>
1 parent d9a3530 commit ef6d38a

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

cpp/tensorrt_llm/thop/attentionOp.cpp

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -671,7 +671,8 @@ bool attention_supports_nvfp4_output(int64_t const num_heads, int64_t const num_
671671
bool const use_paged_context_fmha, bool is_mla_enable)
672672
{
673673
// Only Blackwell supports NVFP4 output.
674-
if (tensorrt_llm::common::getSMVersion() < 100)
674+
// SM 120 does not support NVFP4 output.
675+
if (tensorrt_llm::common::getSMVersion() < 100 || tensorrt_llm::common::getSMVersion() == 120)
675676
{
676677
return false;
677678
}

0 commit comments

Comments
 (0)