Skip to content

Commit fdd3275

Browse files
authored
[CI/Build] Cleanup LoRA test (#26752)
Signed-off-by: Jee Jee Li <[email protected]>
1 parent c715ba3 commit fdd3275

File tree

3 files changed

+3
-11
lines changed

3 files changed

+3
-11
lines changed

tests/lora/test_chatglm3_tp.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,6 @@ def test_chatglm3_lora(chatglm3_lora_files):
5858
max_loras=4,
5959
max_lora_rank=64,
6060
trust_remote_code=True,
61-
enable_chunked_prefill=True,
6261
)
6362

6463
output1 = do_sample(llm, chatglm3_lora_files, lora_id=1)
@@ -70,7 +69,6 @@ def test_chatglm3_lora(chatglm3_lora_files):
7069

7170

7271
@multi_gpu_test(num_gpus=4)
73-
@create_new_process_for_each_test()
7472
def test_chatglm3_lora_tp4(chatglm3_lora_files):
7573
llm = vllm.LLM(
7674
MODEL_PATH,
@@ -81,7 +79,6 @@ def test_chatglm3_lora_tp4(chatglm3_lora_files):
8179
tensor_parallel_size=4,
8280
trust_remote_code=True,
8381
fully_sharded_loras=False,
84-
enable_chunked_prefill=True,
8582
)
8683

8784
output1 = do_sample(llm, chatglm3_lora_files, lora_id=1)
@@ -93,7 +90,6 @@ def test_chatglm3_lora_tp4(chatglm3_lora_files):
9390

9491

9592
@multi_gpu_test(num_gpus=4)
96-
@create_new_process_for_each_test()
9793
def test_chatglm3_lora_tp4_fully_sharded_loras(chatglm3_lora_files):
9894
# https://github.com/NVIDIA/nccl/issues/1790, set a lower value for
9995
# gpu_memory_utilization here because NCCL >= 2.26.3 seems to use
@@ -107,7 +103,6 @@ def test_chatglm3_lora_tp4_fully_sharded_loras(chatglm3_lora_files):
107103
tensor_parallel_size=4,
108104
trust_remote_code=True,
109105
fully_sharded_loras=True,
110-
enable_chunked_prefill=True,
111106
gpu_memory_utilization=0.85,
112107
)
113108
output1 = do_sample(llm, chatglm3_lora_files, lora_id=1)

tests/lora/test_llama_tp.py

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -113,7 +113,6 @@ def test_llama_lora(sql_lora_files):
113113

114114

115115
@multi_gpu_test(num_gpus=4)
116-
@create_new_process_for_each_test()
117116
def test_llama_lora_tp4(sql_lora_files):
118117
llm = vllm.LLM(
119118
MODEL_PATH,
@@ -127,7 +126,6 @@ def test_llama_lora_tp4(sql_lora_files):
127126

128127

129128
@multi_gpu_test(num_gpus=4)
130-
@create_new_process_for_each_test()
131129
def test_llama_lora_tp4_fully_sharded_loras(sql_lora_files):
132130
llm = vllm.LLM(
133131
MODEL_PATH,
@@ -142,7 +140,6 @@ def test_llama_lora_tp4_fully_sharded_loras(sql_lora_files):
142140

143141

144142
@multi_gpu_test(num_gpus=2)
145-
@create_new_process_for_each_test()
146143
def test_tp2_serialize_and_deserialize_lora(
147144
tmp_path, sql_lora_files, sql_lora_huggingface_id
148145
):

tests/lora/test_minicpmv_tp.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@
88
from vllm.lora.request import LoRARequest
99
from vllm.platforms import current_platform
1010

11-
from ..utils import create_new_process_for_each_test
11+
from ..utils import multi_gpu_test
1212

1313
MODEL_PATH = "openbmb/MiniCPM-Llama3-V-2_5"
1414

@@ -88,7 +88,7 @@ def test_minicpmv_lora(minicpmv_lora_files):
8888
current_platform.is_rocm(),
8989
reason="MiniCPM-V dependency xformers incompatible with ROCm",
9090
)
91-
@create_new_process_for_each_test()
91+
@multi_gpu_test(num_gpus=4)
9292
def test_minicpmv_tp4_wo_fully_sharded_loras(minicpmv_lora_files):
9393
llm = vllm.LLM(
9494
MODEL_PATH,
@@ -112,7 +112,7 @@ def test_minicpmv_tp4_wo_fully_sharded_loras(minicpmv_lora_files):
112112
current_platform.is_rocm(),
113113
reason="MiniCPM-V dependency xformers incompatible with ROCm",
114114
)
115-
@create_new_process_for_each_test()
115+
@multi_gpu_test(num_gpus=4)
116116
def test_minicpmv_tp4_fully_sharded_loras(minicpmv_lora_files):
117117
llm = vllm.LLM(
118118
MODEL_PATH,

0 commit comments

Comments
 (0)