Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
16 changes: 14 additions & 2 deletions tests/test_regression.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,7 @@
import pytest
import torch

from tests.utils import large_gpu_mark
from vllm import LLM, SamplingParams
from vllm.platforms import current_platform

Expand All @@ -32,10 +33,21 @@ def test_duplicated_ignored_sequence_group():
assert len(prompts) == len(outputs)


def test_max_tokens_none():
@pytest.mark.parametrize(
"model",
[
pytest.param(
"distilbert/distilgpt2",
Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This model only has 88.2M params, how does it OOM?

Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can you add it to #37736?

marks=[
*([large_gpu_mark(min_gb=80)] if current_platform.is_rocm() else []),
],
),
],
)
def test_max_tokens_none(model):
sampling_params = SamplingParams(temperature=0.01, top_p=0.1, max_tokens=None)
llm = LLM(
model="distilbert/distilgpt2",
model=model,
max_num_batched_tokens=4096,
tensor_parallel_size=1,
)
Expand Down
Loading