We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 80c7b08 commit 34a0e96Copy full SHA for 34a0e96
vllm/envs.py
@@ -352,7 +352,7 @@ def get_default_config_root():
352
os.path.join(get_default_cache_root(), "vllm", "xla_cache"),
353
)),
354
"VLLM_FUSED_MOE_CHUNK_SIZE":
355
- lambda: int(os.getenv("VLLM_FUSED_MOE_CHUNK_SIZE", "65536")),
+ lambda: int(os.getenv("VLLM_FUSED_MOE_CHUNK_SIZE", "32768")),
356
357
# If set, vllm will skip the deprecation warnings.
358
"VLLM_NO_DEPRECATION_WARNING":
0 commit comments