Skip to content
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
6 changes: 5 additions & 1 deletion tensorrt_llm/llmapi/llm_args.py
Original file line number Diff line number Diff line change
Expand Up @@ -2230,7 +2230,11 @@ def validate_cuda_graph_config(self) -> 'TorchLlmArgs':
else:
config.max_batch_size = max(config.batch_sizes)
else:
max_batch_size = config.max_batch_size or 128
# Use the max batch size from:
# 1. cuda_graph_config.max_batch_size, if provided,
# 2. base_llm_args.max_batch_size, if provided,
# 3. default value 128.
max_batch_size = config.max_batch_size or self.max_batch_size or 128
generated_sizes = CudaGraphConfig._generate_cuda_graph_batch_sizes(
max_batch_size, config.enable_padding)
config.batch_sizes = generated_sizes
Expand Down