diff --git a/vllm/v1/engine/detokenizer.py b/vllm/v1/engine/detokenizer.py index 006d53d8f128..330a3f6dad90 100644 --- a/vllm/v1/engine/detokenizer.py +++ b/vllm/v1/engine/detokenizer.py @@ -161,7 +161,7 @@ def __init__(self, tokenizer: PreTrainedTokenizerFast, prompt_suffix = request.prompt_token_ids prompt_len = len(prompt_suffix) if prompt_len > 4: - for i in range(4, max(prompt_len + 1, 32)): + for i in range(4, min(prompt_len + 1, 24)): suffix = request.prompt_token_ids[-i:] if '�' not in self.tokenizer.decode(suffix): prompt_suffix = suffix