Skip to content

Commit a64a844

Browse files
authored
[2/n][ci] S3: Use full model path (#13564)
Signed-off-by: <>
1 parent aa1e62d commit a64a844

File tree

8 files changed

+26
-19
lines changed

8 files changed

+26
-19
lines changed

tests/basic_correctness/test_cumem.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -121,7 +121,7 @@ def model(x):
121121
"model, use_v1",
122122
[
123123
# sleep mode with safetensors
124-
(f"{MODEL_WEIGHTS_S3_BUCKET}/Llama-3.2-1B", True),
124+
(f"{MODEL_WEIGHTS_S3_BUCKET}/meta-llama/Llama-3.2-1B", True),
125125
# sleep mode with pytorch checkpoint
126126
("facebook/opt-125m", False),
127127
])

tests/conftest.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -746,8 +746,7 @@ def __init__(
746746
**kwargs,
747747
) -> None:
748748
if model_name in MODELS_ON_S3 and not load_format:
749-
model_name = (f"s3://vllm-ci-model-weights/"
750-
f"{model_name.split('/')[-1]}")
749+
model_name = (f"{MODEL_WEIGHTS_S3_BUCKET}/{model_name}")
751750
load_format = LoadFormat.RUNAI_STREAMER
752751
if not load_format:
753752
load_format = LoadFormat.AUTO

tests/engine/test_computed_prefix_blocks.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,8 @@
1010
from ..conftest import MODEL_WEIGHTS_S3_BUCKET
1111

1212

13-
@pytest.mark.parametrize("model", [f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2"])
13+
@pytest.mark.parametrize("model",
14+
[f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2"])
1415
@pytest.mark.parametrize("block_size", [16])
1516
def test_computed_prefix_blocks(model: str, block_size: int):
1617
# This test checks if we are able to run the engine to completion

tests/engine/test_detokenization.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,8 @@
99
from ..conftest import MODEL_WEIGHTS_S3_BUCKET
1010

1111

12-
@pytest.mark.parametrize("model", [f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2"])
12+
@pytest.mark.parametrize("model",
13+
[f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2"])
1314
def test_computed_prefix_blocks(model: str):
1415
# This test checks if the engine generates completions both with and
1516
# without optional detokenization, that detokenization includes text

tests/engine/test_executor.py

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,8 @@ def collective_rpc(self,
3838
CustomUniExecutorAsync = CustomUniExecutor
3939

4040

41-
@pytest.mark.parametrize("model", [f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2"])
41+
@pytest.mark.parametrize("model",
42+
[f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2"])
4243
def test_custom_executor_type_checking(model):
4344
with pytest.raises(ValueError):
4445
engine_args = EngineArgs(model=model,
@@ -51,7 +52,8 @@ def test_custom_executor_type_checking(model):
5152
AsyncLLMEngine.from_engine_args(engine_args)
5253

5354

54-
@pytest.mark.parametrize("model", [f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2"])
55+
@pytest.mark.parametrize("model",
56+
[f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2"])
5557
def test_custom_executor(model, tmp_path):
5658
cwd = os.path.abspath(".")
5759
os.chdir(tmp_path)
@@ -75,7 +77,8 @@ def test_custom_executor(model, tmp_path):
7577
os.chdir(cwd)
7678

7779

78-
@pytest.mark.parametrize("model", [f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2"])
80+
@pytest.mark.parametrize("model",
81+
[f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2"])
7982
def test_custom_executor_async(model, tmp_path):
8083
cwd = os.path.abspath(".")
8184
os.chdir(tmp_path)
@@ -103,7 +106,8 @@ async def t():
103106
os.chdir(cwd)
104107

105108

106-
@pytest.mark.parametrize("model", [f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2"])
109+
@pytest.mark.parametrize("model",
110+
[f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2"])
107111
def test_respect_ray(model):
108112
# even for TP=1 and PP=1,
109113
# if users specify ray, we should use ray.

tests/engine/test_skip_tokenizer_init.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,8 @@
99
from ..conftest import MODEL_WEIGHTS_S3_BUCKET
1010

1111

12-
@pytest.mark.parametrize("model", [f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2"])
12+
@pytest.mark.parametrize("model",
13+
[f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2"])
1314
def test_skip_tokenizer_initialization(model: str):
1415
# This test checks if the flag skip_tokenizer_init skips the initialization
1516
# of tokenizer and detokenizer. The generated output is expected to contain

tests/test_config.py

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -14,13 +14,14 @@
1414
@pytest.mark.parametrize(
1515
("model_id", "expected_runner_type", "expected_task"),
1616
[
17-
(f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2", "generate", "generate"),
18-
(f"{MODEL_WEIGHTS_S3_BUCKET}/e5-mistral-7b-instruct", "pooling",
19-
"embed"),
20-
(f"{MODEL_WEIGHTS_S3_BUCKET}/Qwen2.5-1.5B-apeach", "pooling",
17+
(f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2", "generate",
18+
"generate"),
19+
(f"{MODEL_WEIGHTS_S3_BUCKET}/intfloat/e5-mistral-7b-instruct",
20+
"pooling", "embed"),
21+
(f"{MODEL_WEIGHTS_S3_BUCKET}/jason9693/Qwen2.5-1.5B-apeach", "pooling",
2122
"classify"),
22-
(f"{MODEL_WEIGHTS_S3_BUCKET}/ms-marco-MiniLM-L-6-v2", "pooling",
23-
"score"),
23+
(f"{MODEL_WEIGHTS_S3_BUCKET}/cross-encoder/ms-marco-MiniLM-L-6-v2",
24+
"pooling", "score"),
2425
("Qwen/Qwen2.5-Math-RM-72B", "pooling", "reward"),
2526
("openai/whisper-small", "transcription", "transcription"),
2627
],

tests/test_regression.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ def test_duplicated_ignored_sequence_group():
2121
sampling_params = SamplingParams(temperature=0.01,
2222
top_p=0.1,
2323
max_tokens=256)
24-
llm = LLM(model=f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2",
24+
llm = LLM(model=f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2",
2525
load_format=LoadFormat.RUNAI_STREAMER,
2626
max_num_batched_tokens=4096,
2727
tensor_parallel_size=1)
@@ -35,7 +35,7 @@ def test_max_tokens_none():
3535
sampling_params = SamplingParams(temperature=0.01,
3636
top_p=0.1,
3737
max_tokens=None)
38-
llm = LLM(model=f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2",
38+
llm = LLM(model=f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2",
3939
load_format=LoadFormat.RUNAI_STREAMER,
4040
max_num_batched_tokens=4096,
4141
tensor_parallel_size=1)
@@ -46,7 +46,7 @@ def test_max_tokens_none():
4646

4747

4848
def test_gc():
49-
llm = LLM(model=f"{MODEL_WEIGHTS_S3_BUCKET}/distilgpt2",
49+
llm = LLM(model=f"{MODEL_WEIGHTS_S3_BUCKET}/distilbert/distilgpt2",
5050
load_format=LoadFormat.RUNAI_STREAMER,
5151
enforce_eager=True)
5252
del llm

0 commit comments

Comments
 (0)