4
4
import ray
5
5
6
6
import vllm
7
- from tests .utils import fork_new_process_for_each_test
8
7
from vllm .lora .request import LoRARequest
9
8
10
- from ..utils import multi_gpu_test
9
+ from ..utils import create_new_process_for_each_test , multi_gpu_test
11
10
12
11
MODEL_PATH = "meta-llama/Llama-2-7b-hf"
13
12
@@ -82,7 +81,7 @@ def v1(run_with_both_engines_lora):
82
81
83
82
# V1 Test: Failing due to numerics on V1.
84
83
@pytest .mark .skip_v1
85
- @fork_new_process_for_each_test
84
+ @create_new_process_for_each_test ()
86
85
def test_llama_lora (sql_lora_files ):
87
86
88
87
llm = vllm .LLM (MODEL_PATH ,
@@ -97,7 +96,7 @@ def test_llama_lora(sql_lora_files):
97
96
# Skipping for v1 as v1 doesn't have a good way to expose the num_gpu_blocks
98
97
# used by the engine yet.
99
98
@pytest .mark .skip_v1
100
- @fork_new_process_for_each_test
99
+ @create_new_process_for_each_test ()
101
100
def test_llama_lora_warmup (sql_lora_files ):
102
101
"""Test that the LLM initialization works with a warmup LORA path and
103
102
is more conservative"""
@@ -128,7 +127,7 @@ def get_num_gpu_blocks_no_lora():
128
127
# V1 Test: Failing due to numerics on V1.
129
128
@pytest .mark .skip_v1
130
129
@multi_gpu_test (num_gpus = 4 )
131
- @fork_new_process_for_each_test
130
+ @create_new_process_for_each_test ()
132
131
def test_llama_lora_tp4 (sql_lora_files ):
133
132
134
133
llm = vllm .LLM (
@@ -143,7 +142,7 @@ def test_llama_lora_tp4(sql_lora_files):
143
142
144
143
145
144
@multi_gpu_test (num_gpus = 4 )
146
- @fork_new_process_for_each_test
145
+ @create_new_process_for_each_test ()
147
146
def test_llama_lora_tp4_fully_sharded_loras (sql_lora_files ):
148
147
149
148
llm = vllm .LLM (
@@ -159,7 +158,7 @@ def test_llama_lora_tp4_fully_sharded_loras(sql_lora_files):
159
158
160
159
161
160
@multi_gpu_test (num_gpus = 4 )
162
- @fork_new_process_for_each_test
161
+ @create_new_process_for_each_test ()
163
162
def test_llama_lora_tp4_fully_sharded_enable_bias (sql_lora_files ):
164
163
165
164
llm = vllm .LLM (
0 commit comments