Skip to content

Commit 2a4c825

Browse files
authored
[CI] Skip the pooling models that do not support transformers v4.55 (#22411)
Signed-off-by: wang.yuqi <[email protected]>
1 parent 4be02a3 commit 2a4c825

File tree

4 files changed

+28
-1
lines changed

4 files changed

+28
-1
lines changed

tests/models/language/pooling/test_embedding.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@
77
from vllm.config import PoolerConfig
88
from vllm.platforms import current_platform
99

10-
from ...utils import check_embeddings_close
10+
from ...utils import check_embeddings_close, check_transformers_version
1111

1212

1313
@pytest.fixture(autouse=True)
@@ -56,6 +56,9 @@ def test_models(
5656
model,
5757
monkeypatch,
5858
) -> None:
59+
if model == "Alibaba-NLP/gte-Qwen2-1.5B-instruct":
60+
check_transformers_version(model, max_transformers_version="4.53.2")
61+
5962
if model == "BAAI/bge-multilingual-gemma2" and current_platform.is_rocm():
6063
# ROCm Triton FA does not currently support sliding window attention
6164
# switch to use ROCm CK FA backend

tests/models/language/pooling/test_gte.py

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,7 @@
44

55
import pytest
66

7+
from ...utils import check_transformers_version
78
from .embed_utils import EmbedModelInfo, correctness_test_embed_models
89
from .mteb_utils import mteb_test_embed_models
910

@@ -60,6 +61,10 @@
6061
@pytest.mark.parametrize("model_info", MODELS)
6162
def test_embed_models_mteb(hf_runner, vllm_runner,
6263
model_info: EmbedModelInfo) -> None:
64+
if model_info.name == "Alibaba-NLP/gte-Qwen2-1.5B-instruct":
65+
check_transformers_version(model_info.name,
66+
max_transformers_version="4.53.2")
67+
6368
vllm_extra_kwargs: dict[str, Any] = {}
6469
if model_info.architecture == "GteNewModel":
6570
vllm_extra_kwargs["hf_overrides"] = {"architectures": ["GteNewModel"]}
@@ -72,6 +77,10 @@ def test_embed_models_mteb(hf_runner, vllm_runner,
7277
def test_embed_models_correctness(hf_runner, vllm_runner,
7378
model_info: EmbedModelInfo,
7479
example_prompts) -> None:
80+
if model_info.name == "Alibaba-NLP/gte-Qwen2-1.5B-instruct":
81+
check_transformers_version(model_info.name,
82+
max_transformers_version="4.53.2")
83+
7584
vllm_extra_kwargs: dict[str, Any] = {}
7685
if model_info.architecture == "GteNewModel":
7786
vllm_extra_kwargs["hf_overrides"] = {"architectures": ["GteNewModel"]}

tests/models/language/pooling/test_reward.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@
1010
from vllm.platforms import current_platform
1111

1212
from ....conftest import HfRunner
13+
from ...utils import check_transformers_version
1314

1415

1516
@pytest.fixture(autouse=True)
@@ -86,6 +87,9 @@ def test_prm_models(
8687
dtype: str,
8788
monkeypatch,
8889
) -> None:
90+
check_transformers_version("Qwen/Qwen2.5-Math-PRM-7B",
91+
max_transformers_version="4.53.2")
92+
8993
if current_platform.is_cpu() and os.environ.get("VLLM_USE_V1", "0") == "0":
9094
pytest.skip("CPU only supports V1")
9195

tests/models/utils.py

Lines changed: 11 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -412,3 +412,14 @@ def dummy_hf_overrides(
412412
})
413413

414414
return hf_config
415+
416+
417+
def check_transformers_version(model: str,
418+
min_transformers_version: Optional[str] = None,
419+
max_transformers_version: Optional[str] = None):
420+
from .registry import _HfExamplesInfo
421+
422+
return _HfExamplesInfo(model,
423+
min_transformers_version=min_transformers_version,
424+
max_transformers_version=max_transformers_version
425+
).check_transformers_version(on_fail="skip")

0 commit comments

Comments
 (0)