@@ -85,108 +85,7 @@ jobs:
85
85
run : |
86
86
# We found that if running aclgraph tests in batch, it will cause AclmdlRICaptureBegin error. So we run
87
87
# the test separately.
88
-
89
- pytest -sv tests/e2e/singlecard/test_aclgraph.py
90
- pytest -sv tests/e2e/singlecard/test_ascend_scheduler.py
91
- pytest -sv tests/e2e/singlecard/test_camem.py
92
- pytest -sv tests/e2e/singlecard/test_chunked.py
93
- pytest -sv tests/e2e/singlecard/test_embedding.py
94
- pytest -sv tests/e2e/singlecard/test_guided_decoding.py
95
- pytest -sv tests/e2e/singlecard/test_ilama_lora.py
96
- pytest -sv tests/e2e/singlecard/test_profile_execute_duration.py
97
- pytest -sv tests/e2e/singlecard/test_quantization.py
98
- pytest -sv tests/e2e/singlecard/test_sampler.py
99
- pytest -sv tests/e2e/singlecard/test_vlm.py
100
-
101
- # ------------------------------------ v1 spec decode test ------------------------------------ #
88
+ npu-smi info
102
89
pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_mtp_correctness.py
103
- pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_mtp_torchair_correctness.py
104
- pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_spec_decode.py
105
-
106
- pytest -sv tests/e2e/singlecard/ops/
107
-
108
- e2e-2-cards :
109
- name : multicard
110
- runs-on : ${{ inputs.runner }}-2
111
- container :
112
- image : ${{ inputs.image }}
113
- env :
114
- VLLM_LOGGING_LEVEL : ERROR
115
- VLLM_USE_MODELSCOPE : True
116
- steps :
117
- - name : Check npu and CANN info
118
- run : |
119
90
npu-smi info
120
- cat /usr/local/Ascend/ascend-toolkit/latest/"$(uname -i)"-linux/ascend_toolkit_install.info
121
-
122
- - name : Config mirrors
123
- run : |
124
- sed -Ei 's@(ports|archive)[email protected] :8081@g' /etc/apt/sources.list
125
- pip config set global.index-url http://cache-service.nginx-pypi-cache.svc.cluster.local/pypi/simple
126
- pip config set global.trusted-host cache-service.nginx-pypi-cache.svc.cluster.local
127
- apt-get update -y
128
- apt install git -y
129
-
130
- - name : Checkout vllm-project/vllm-ascend repo
131
- uses : actions/checkout@v4
132
-
133
- - name : Install system dependencies
134
- run : |
135
- apt-get -y install `cat packages.txt`
136
- apt-get -y install gcc g++ cmake libnuma-dev
137
-
138
- - name : Checkout vllm-project/vllm repo
139
- uses : actions/checkout@v4
140
- with :
141
- repository : vllm-project/vllm
142
- ref : ${{ inputs.vllm }}
143
- path : ./vllm-empty
144
- fetch-depth : 1
145
-
146
- - name : Install vllm-project/vllm from source
147
- working-directory : ./vllm-empty
148
- run : |
149
- VLLM_TARGET_DEVICE=empty pip install -e .
150
-
151
- - name : Install vllm-project/vllm-ascend
152
- env :
153
- PIP_EXTRA_INDEX_URL : https://mirrors.huaweicloud.com/ascend/repos/pypi
154
- run : |
155
- pip install -r requirements-dev.txt
156
- pip install -v -e .
157
-
158
- - name : Run vllm-project/vllm-ascend test (light)
159
- env :
160
- VLLM_WORKER_MULTIPROC_METHOD : spawn
161
- VLLM_USE_MODELSCOPE : True
162
- if : ${{ inputs.type == 'light' }}
163
- run : |
164
- pytest -sv tests/e2e/multicard/test_qwen3_moe.py::test_models_distributed_Qwen3_MOE_TP2_WITH_EP
165
-
166
- - name : Run vllm-project/vllm-ascend test (full)
167
- env :
168
- VLLM_WORKER_MULTIPROC_METHOD : spawn
169
- VLLM_USE_MODELSCOPE : True
170
- if : ${{ inputs.type == 'full' }}
171
- run : |
172
- pytest -sv tests/e2e/multicard/test_data_parallel.py
173
- pytest -sv tests/e2e/multicard/test_expert_parallel.py
174
- # external_launcher test is not stable enough. Fix it later
175
- # pytest -sv tests/e2e/multicard/test_external_launcher.py
176
- pytest -sv tests/e2e/multicard/test_fused_moe_allgather_ep.py
177
- pytest -sv tests/e2e/multicard/test_ilama_lora_tp2.py
178
-
179
- # To avoid oom, we need to run the test in a single process.
180
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_QwQ
181
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_multistream_moe
182
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_Qwen3_W8A8
183
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_Qwen3_W4A8DYNAMIC
184
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_W4A8DYNAMIC
185
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_sp_for_qwen3_moe
186
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_Qwen_Dense_with_flashcomm_v1
187
- pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_Qwen_Dense_with_prefetch_mlp_weight
188
-
189
- #pytest -sv tests/e2e/multicard/test_pipeline_parallel.py
190
- pytest -sv tests/e2e/multicard/test_prefix_caching.py
191
- pytest -sv tests/e2e/multicard/test_qwen3_moe.py
192
- pytest -sv tests/e2e/multicard/test_torchair_graph_mode.py
91
+ pytest -sv tests/e2e/singlecard/spec_decode_v1/test_v1_spec_decode.py
0 commit comments