10
10
from ...utils import check_embeddings_close , check_transformers_version
11
11
12
12
13
- @pytest .fixture (autouse = True )
14
- def v1 (run_with_both_engines ):
15
- # Simple autouse wrapper to run both engines for each test
16
- # This can be promoted up to conftest.py to run for every
17
- # test in a package
18
- pass
19
-
20
-
21
13
@pytest .mark .parametrize (
22
14
"model" ,
23
15
[
@@ -32,21 +24,15 @@ def v1(run_with_both_engines):
32
24
"intfloat/e5-mistral-7b-instruct" ,
33
25
# CPU v1 doesn't support sliding window
34
26
marks = [pytest .mark .core_model ]),
35
- # the qwen models interfere with each other (see PR
36
- # https://github.com/vllm-project/vllm/pull/18720).
37
- # To avoid this problem, for now we skip v0 since it will be
38
- # deprecated anyway.
39
27
pytest .param ("ssmits/Qwen2-7B-Instruct-embed-base" ,
40
- marks = [pytest .mark .skip_v0 , pytest . mark . cpu_model ]),
28
+ marks = [pytest .mark .cpu_model ]),
41
29
# [Encoder-only]
42
30
pytest .param ("BAAI/bge-base-en-v1.5" , marks = [pytest .mark .core_model ]),
43
31
pytest .param ("sentence-transformers/all-MiniLM-L12-v2" ),
44
32
pytest .param ("intfloat/multilingual-e5-small" ),
45
- pytest .param ("Alibaba-NLP/gte-Qwen2-1.5B-instruct" ,
46
- marks = [pytest .mark .skip_v1 ]),
33
+ pytest .param ("Alibaba-NLP/gte-Qwen2-1.5B-instruct" ),
47
34
# [Cross-Encoder]
48
- pytest .param ("sentence-transformers/stsb-roberta-base-v2" ,
49
- marks = [pytest .mark .skip_v1 ]),
35
+ pytest .param ("sentence-transformers/stsb-roberta-base-v2" ),
50
36
],
51
37
)
52
38
def test_models (
0 commit comments