diff --git a/tests/unittest/llmapi/test_llm_args.py b/tests/unittest/llmapi/test_llm_args.py index c1bfdcc4001..801a2bf12a9 100644 --- a/tests/unittest/llmapi/test_llm_args.py +++ b/tests/unittest/llmapi/test_llm_args.py @@ -372,18 +372,18 @@ class TestTorchLlmArgs: def test_runtime_sizes(self): llm = TorchLLM( llama_model_path, - max_beam_width=4, + max_beam_width=1, max_num_tokens=256, max_seq_len=128, max_batch_size=8, ) - assert llm.args.max_beam_width == 4 + assert llm.args.max_beam_width == 1 assert llm.args.max_num_tokens == 256 assert llm.args.max_seq_len == 128 assert llm.args.max_batch_size == 8 - assert llm._executor_config.max_beam_width == 4 + assert llm._executor_config.max_beam_width == 1 assert llm._executor_config.max_num_tokens == 256 assert llm._executor_config.max_seq_len == 128 assert llm._executor_config.max_batch_size == 8