Skip to content

Commit 0cd1812

Browse files
committed
[Qwen-moe] Remove the minor operation arange
Signed-off-by: s30076806 <[email protected]>
2 parents 5f4adf5 + 950c4b2 commit 0cd1812

File tree

124 files changed

+13845
-3074
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

124 files changed

+13845
-3074
lines changed

.github/ISSUE_TEMPLATE/750-RFC.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@ body:
4040
attributes:
4141
label: Any Other Things.
4242
description: >
43-
Any other things you would like to mention.
43+
Any other things you would like to mention, such as feature branch request.
4444
validations:
4545
required: false
4646
- type: markdown

.github/workflows/format_pr_body.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@ jobs:
4646
echo "VLLM_COMMIT=https://github.com/vllm-project/vllm/commit/$VLLM_COMMIT" >> $GITHUB_ENV
4747
4848
- name: Checkout repository
49-
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
49+
uses: actions/checkout@ff7abcd0c3c05ccf6adc123a8cd1fd4fb30fb493 # v4.2.2
5050

5151
- name: Set up Python
5252
uses: actions/setup-python@42375524e23c412d93fb67b49958b491fce71c38 # v5.4.0

.github/workflows/release_code.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ jobs:
4343
matrix:
4444
python-version: ["3.11"]
4545
steps:
46-
- uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
46+
- uses: actions/checkout@ff7abcd0c3c05ccf6adc123a8cd1fd4fb30fb493 # v4.2.2
4747

4848
- name: Print
4949
run: |

.github/workflows/release_whl.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ jobs:
5252
) }}
5353
runs-on: ${{ matrix.os }}
5454
steps:
55-
- uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
55+
- uses: actions/checkout@ff7abcd0c3c05ccf6adc123a8cd1fd4fb30fb493 # v4.2.2
5656

5757
- name: Print
5858
run: |

.github/workflows/vllm_ascend_test.yaml

Lines changed: 5 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -81,7 +81,7 @@ jobs:
8181
VLLM_USE_MODELSCOPE: True
8282
strategy:
8383
matrix:
84-
vllm_version: [main, v0.10.0]
84+
vllm_version: [v0.10.1.1, main]
8585
steps:
8686
- name: Install packages
8787
run: |
@@ -137,7 +137,7 @@ jobs:
137137
max-parallel: 2
138138
matrix:
139139
os: [linux-aarch64-a2-1]
140-
vllm_version: [main, v0.10.0]
140+
vllm_version: [v0.10.1.1, main]
141141
name: singlecard e2e test
142142
runs-on: ${{ matrix.os }}
143143
container:
@@ -185,17 +185,14 @@ jobs:
185185
run: |
186186
pip install -r requirements-dev.txt
187187
pip install -v -e .
188-
if [[ "${{ matrix.vllm_version }}" == "v0.10.0" ]]; then
189-
pip install "transformers<4.54.0"
190-
fi
191188
192189
- name: Run e2e test
193190
env:
194191
VLLM_WORKER_MULTIPROC_METHOD: spawn
195192
VLLM_USE_MODELSCOPE: True
196193
run: |
197194
pytest -sv tests/e2e/singlecard/test_offline_inference.py
198-
pytest -sv tests/e2e/singlecard/test_ilama_lora.py
195+
# pytest -sv tests/e2e/singlecard/test_ilama_lora.py
199196
pytest -sv tests/e2e/singlecard/test_guided_decoding.py
200197
pytest -sv tests/e2e/singlecard/test_camem.py
201198
pytest -sv tests/e2e/singlecard/test_embedding.py
@@ -222,7 +219,7 @@ jobs:
222219
max-parallel: 2
223220
matrix:
224221
os: [linux-aarch64-a2-2]
225-
vllm_version: [main, v0.10.0]
222+
vllm_version: [v0.10.1.1, main]
226223
name: multicard e2e test
227224
runs-on: ${{ matrix.os }}
228225
container:
@@ -270,21 +267,17 @@ jobs:
270267
run: |
271268
pip install -r requirements-dev.txt
272269
pip install -v -e .
273-
if [[ "${{ matrix.vllm_version }}" == "v0.10.0" ]]; then
274-
pip install "transformers<4.54.0"
275-
fi
276270
277271
- name: Run vllm-project/vllm-ascend test
278272
env:
279273
VLLM_WORKER_MULTIPROC_METHOD: spawn
280274
VLLM_USE_MODELSCOPE: True
281275
run: |
282-
pytest -sv tests/e2e/multicard/test_ilama_lora_tp2.py
276+
# pytest -sv tests/e2e/multicard/test_ilama_lora_tp2.py
283277
# Fixme: run VLLM_USE_MODELSCOPE=True pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py will raise error.
284278
# To avoid oom, we need to run the test in a single process.
285279
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_multistream_moe
286280
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_QwQ
287-
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeek_dbo
288281
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_DeepSeekV3_dbo
289282
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_alltoallv
290283
pytest -sv tests/e2e/multicard/test_offline_inference_distributed.py::test_models_distributed_Qwen3_W4A8DYNAMIC

.github/workflows/vllm_ascend_test_310p.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,7 @@ jobs:
5353
max-parallel: 2
5454
matrix:
5555
os: [linux-aarch64-310p-1, linux-aarch64-310p-4]
56-
vllm_version: [main, v0.10.0]
56+
vllm_version: [v0.10.1.1, main]
5757
name: 310p e2e test
5858
runs-on: ${{ matrix.os }}
5959
container:

Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ RUN pip config set global.index-url ${PIP_INDEX_URL}
3737

3838
# Install vLLM
3939
ARG VLLM_REPO=https://github.com/vllm-project/vllm.git
40-
ARG VLLM_TAG=v0.10.0
40+
ARG VLLM_TAG=v0.10.1.1
4141
RUN git clone --depth 1 $VLLM_REPO --branch $VLLM_TAG /vllm-workspace/vllm
4242
# In x86, triton will be installed by vllm. But in Ascend, triton doesn't work correctly. we need to uninstall it.
4343
RUN VLLM_TARGET_DEVICE="empty" python3 -m pip install -v -e /vllm-workspace/vllm/ --extra-index https://download.pytorch.org/whl/cpu/ && \

Dockerfile.310p

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ RUN pip config set global.index-url ${PIP_INDEX_URL}
3737

3838
# Install vLLM
3939
ARG VLLM_REPO=https://github.com/vllm-project/vllm.git
40-
ARG VLLM_TAG=v0.10.0
40+
ARG VLLM_TAG=v0.10.1.1
4141
RUN git clone --depth 1 $VLLM_REPO --branch $VLLM_TAG /vllm-workspace/vllm
4242
# In x86, triton will be installed by vllm. But in Ascend, triton doesn't work correctly. we need to uninstall it.
4343
RUN VLLM_TARGET_DEVICE="empty" python3 -m pip install -v -e /vllm-workspace/vllm/ --extra-index https://download.pytorch.org/whl/cpu/ && \

Dockerfile.310p.openEuler

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@ COPY . /vllm-workspace/vllm-ascend/
3434

3535
# Install vLLM
3636
ARG VLLM_REPO=https://github.com/vllm-project/vllm.git
37-
ARG VLLM_TAG=v0.10.0
37+
ARG VLLM_TAG=v0.10.1.1
3838

3939
RUN git clone --depth 1 $VLLM_REPO --branch $VLLM_TAG /vllm-workspace/vllm
4040
# In x86, triton will be installed by vllm. But in Ascend, triton doesn't work correctly. we need to uninstall it.

Dockerfile.a3

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ RUN pip config set global.index-url ${PIP_INDEX_URL}
3737

3838
# Install vLLM
3939
ARG VLLM_REPO=https://github.com/vllm-project/vllm.git
40-
ARG VLLM_TAG=v0.10.0
40+
ARG VLLM_TAG=v0.10.1.1
4141
RUN git clone --depth 1 $VLLM_REPO --branch $VLLM_TAG /vllm-workspace/vllm
4242
# In x86, triton will be installed by vllm. But in Ascend, triton doesn't work correctly. we need to uninstall it.
4343
RUN VLLM_TARGET_DEVICE="empty" python3 -m pip install -v -e /vllm-workspace/vllm/ --extra-index https://download.pytorch.org/whl/cpu/ && \

0 commit comments

Comments
 (0)