From fd8f8b8e821d31e45715ebc0092e46a7bc5a3ee3 Mon Sep 17 00:00:00 2001 From: "Chendi.Xue" Date: Sat, 12 Jul 2025 02:19:44 +0300 Subject: [PATCH] [BUGFIX][NON_CUDA] Fix failing introduced by #20061 when import bitsandbytes Signed-off-by: Chendi.Xue --- vllm/model_executor/layers/quantization/bitsandbytes.py | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/vllm/model_executor/layers/quantization/bitsandbytes.py b/vllm/model_executor/layers/quantization/bitsandbytes.py index 20625f587f51..539e1acc7eeb 100644 --- a/vllm/model_executor/layers/quantization/bitsandbytes.py +++ b/vllm/model_executor/layers/quantization/bitsandbytes.py @@ -5,7 +5,11 @@ import torch -from vllm.model_executor.layers.fused_moe import fused_experts +from vllm.triton_utils import HAS_TRITON + +if HAS_TRITON: + from vllm.model_executor.layers.fused_moe import fused_experts + from vllm.model_executor.layers.fused_moe.layer import (FusedMoE, FusedMoEMethodBase) from vllm.model_executor.layers.linear import (LinearBase, LinearMethodBase,