Skip to content

Commit f0b00bb

Browse files
[ci] remove flash attn installation in ete test workflow (#3908)
* update * update * update
1 parent fae8a98 commit f0b00bb

File tree

6 files changed

+3
-883
lines changed

6 files changed

+3
-883
lines changed

.github/workflows/benchmark.yml

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -103,9 +103,6 @@ jobs:
103103
name: my-artifact-${{ github.run_id }}-py310
104104
- name: Install lmdeploy - dependency
105105
run: |
106-
# manually install flash attn
107-
# the install packeage from. https://github.com/Dao-AILab/flash-attention/releases
108-
python3 -m pip install /root/packages/flash_attn-*.whl
109106
python3 -m pip install -r /nvme/qa_test_models/offline_pkg/requirements.txt
110107
- name: Install lmdeploy
111108
if: ${{github.event_name == 'schedule' || !inputs.offline_mode}}
@@ -119,7 +116,6 @@ jobs:
119116
python3 -m pip install -r requirements/test.txt
120117
- name: Check env
121118
run: |
122-
pip uninstall -y nvidia-nccl-cu11
123119
python3 -m pip list
124120
lmdeploy check_env
125121
mkdir ${{env.REPORT_DIR}}/allure-results/.pytest_cache -p

.github/workflows/daily_ete_test.yml

Lines changed: 1 addition & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -152,18 +152,15 @@ jobs:
152152
cp -r ${{env.TEST_CODE_PATH}}/. .
153153
- name: Install lmdeploy - dependency
154154
run: |
155-
# manually install flash attn
156-
# the install packeage from. https://github.com/Dao-AILab/flash-attention/releases
157-
python3 -m pip install /root/packages/flash_attn-*.whl
158155
python3 -m pip install -r ${{env.OFFLINE_REQUIREMENTS}}
159156
- name: Install lmdeploy
160157
run: |
161158
python3 -m pip uninstall lmdeploy -y && python3 -m pip install lmdeploy-*.whl --no-deps
162159
python3 -m pip install -r requirements/test.txt
160+
python3 -m pip install transformers==4.53.1
163161
pip install ${{env.DEEPSEEK_VL}} --no-deps
164162
- name: Check env
165163
run: |
166-
pip uninstall -y nvidia-nccl-cu11
167164
python3 -m pip list
168165
lmdeploy check_env
169166
rm -rf allure-results
@@ -237,9 +234,6 @@ jobs:
237234
cp -r ${{env.TEST_CODE_PATH}}/. .
238235
- name: Install lmdeploy - dependency
239236
run: |
240-
# manually install flash attn
241-
# the install packeage from. https://github.com/Dao-AILab/flash-attention/releases
242-
python3 -m pip install /root/packages/flash_attn-*.whl
243237
python3 -m pip install -r ${{env.OFFLINE_REQUIREMENTS}}
244238
- name: Install lmdeploy
245239
run: |
@@ -249,7 +243,6 @@ jobs:
249243
pip install ${{env.DEEPSEEK_VL}} --no-deps
250244
- name: Check env
251245
run: |
252-
pip uninstall -y nvidia-nccl-cu11
253246
python3 -m pip list
254247
lmdeploy check_env
255248
cp -r /root/lora .
@@ -343,9 +336,6 @@ jobs:
343336
cp -r ${{env.TEST_CODE_PATH}}/. .
344337
- name: Install lmdeploy - dependency
345338
run: |
346-
# manually install flash attn
347-
# the install packeage from. https://github.com/Dao-AILab/flash-attention/releases
348-
python3 -m pip install /root/packages/flash_attn-*.whl
349339
python3 -m pip install -r ${{env.OFFLINE_REQUIREMENTS}}
350340
- name: Install lmdeploy
351341
run: |
@@ -354,7 +344,6 @@ jobs:
354344
pip install ${{env.DEEPSEEK_VL}} --no-deps
355345
- name: Check env
356346
run: |
357-
pip uninstall -y nvidia-nccl-cu11
358347
python3 -m pip list
359348
lmdeploy check_env
360349
rm -rf allure-results
@@ -430,9 +419,6 @@ jobs:
430419
cp -r ${{env.TEST_CODE_PATH}}/. .
431420
- name: Install lmdeploy - dependency
432421
run: |
433-
# manually install flash attn
434-
# the install packeage from. https://github.com/Dao-AILab/flash-attention/releases
435-
python3 -m pip install /root/packages/flash_attn-*.whl
436422
python3 -m pip install -r ${{env.OFFLINE_REQUIREMENTS}}
437423
- name: Install lmdeploy
438424
run: |
@@ -441,7 +427,6 @@ jobs:
441427
pip install ${{env.DEEPSEEK_VL}} --no-deps
442428
- name: Check env
443429
run: |
444-
pip uninstall -y nvidia-nccl-cu11
445430
python3 -m pip list
446431
lmdeploy check_env
447432
rm -rf allure-results
@@ -490,9 +475,6 @@ jobs:
490475
cp -r ${{env.TEST_CODE_PATH}}/. .
491476
- name: Install lmdeploy - dependency
492477
run: |
493-
# manually install flash attn
494-
# the install packeage from. https://github.com/Dao-AILab/flash-attention/releases
495-
python3 -m pip install /root/packages/flash_attn-*.whl
496478
python3 -m pip install -r ${{env.OFFLINE_REQUIREMENTS}}
497479
- name: Install lmdeploy
498480
run: |
@@ -501,7 +483,6 @@ jobs:
501483
pip install ${{env.DEEPSEEK_VL}} --no-deps
502484
- name: Check env
503485
run: |
504-
pip uninstall -y nvidia-nccl-cu11
505486
python3 -m pip list
506487
lmdeploy check_env
507488
rm -rf allure-results
@@ -552,9 +533,6 @@ jobs:
552533
cp -r ${{env.TEST_CODE_PATH}}/. .
553534
- name: Install lmdeploy - dependency
554535
run: |
555-
# manually install flash attn
556-
# the install packeage from. https://github.com/Dao-AILab/flash-attention/releases
557-
python3 -m pip install /root/packages/flash_attn-*.whl
558536
python3 -m pip install sentence_transformers==2.2.2 --no-deps
559537
python3 -m pip install -r ${{env.OFFLINE_REQUIREMENTS}}
560538
- name: Install lmdeploy
@@ -571,7 +549,6 @@ jobs:
571549
echo "OPENCOMPASS_DIR=$(pwd)" >> $GITHUB_ENV
572550
- name: Check env
573551
run: |
574-
pip uninstall -y nvidia-nccl-cu11
575552
python3 -m pip list
576553
lmdeploy check_env
577554
rm -rf allure-results

.github/workflows/daily_ete_test_3090.yml

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -153,10 +153,10 @@ jobs:
153153
- name: Install lmdeploy
154154
run: |
155155
python3 -m pip install lmdeploy-*.whl --no-deps
156+
python3 -m pip install transformers==4.53.1
156157
python3 -m pip install -r requirements/test.txt
157158
- name: Check env
158159
run: |
159-
pip uninstall -y nvidia-nccl-cu11
160160
python3 -m pip list
161161
lmdeploy check_env
162162
rm -rf allure-results
@@ -229,7 +229,6 @@ jobs:
229229
python3 -m pip install -r requirements/test.txt
230230
- name: Check env
231231
run: |
232-
pip uninstall -y nvidia-nccl-cu11
233232
python3 -m pip list
234233
lmdeploy check_env
235234
rm -rf allure-results
@@ -294,7 +293,6 @@ jobs:
294293
python3 -m pip install -r requirements/test.txt
295294
- name: Check env
296295
run: |
297-
pip uninstall -y nvidia-nccl-cu11
298296
python3 -m pip list
299297
lmdeploy check_env
300298
rm -rf allure-results

.github/workflows/evaluate.yml

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ on:
1717
required: true
1818
description: 'Tested TurboMind models list. eg. [internlm_chat_7b,internlm_chat_7b_w8a16]'
1919
type: string
20-
default: '[turbomind_internlm2_5_7b_chat, pytorch_internlm2_5_7b_chat, turbomind_internlm2_5_7b_chat_batch1, turbomind_internlm2_5_7b_chat_batch1_4bits, turbomind_internlm3_8b_instruct, pytorch_internlm3_8b_instruct, turbomind_internlm2_5_20b_chat, pytorch_internlm2_5_20b_chat, turbomind_qwen1_5_7b_chat, pytorch_qwen1_5_7b_chat, turbomind_llama3_8b_instruct, pytorch_llama3_8b_instruct, turbomind_llama3_1_8b_instruct, pytorch_llama3_1_8b_instruct, turbomind_qwen2_7b_instruct, pytorch_qwen2_7b_instruct, turbomind_qwen2_5_7b_instruct, pytorch_qwen2_5_7b_instruct, turbomind_qwen2_5_32b_instruct, pytorch_qwen2_5_32b_instruct, turbomind_llama2_7b_chat, pytorch_qwen1_5_moe_2_7b_chat, pytorch_gemma_2_9b_it, pytorch_gemma_2_27b_it, turbomind_internlm2_5_7b_chat_4bits, turbomind_internlm2_5_7b_chat_kvint4, turbomind_internlm2_5_7b_chat_kvint8, pytorch_internlm2_5_7b_chat_w8a8, turbomind_internlm3_8b_instruct_4bits, turbomind_internlm3_8b_instruct_kvint4, turbomind_internlm3_8b_instruct_kvint8, pytorch_internlm3_8b_instruct_w8a8, turbomind_llama3_8b_instruct_4bits, turbomind_llama3_8b_instruct_kvint4, turbomind_llama3_1_8b_instruct_4bits, turbomind_llama3_1_8b_instruct_kvint4, turbomind_llama3_1_8b_instruct_kvint8,turbomind_llama3_8b_instruct_kvint8, pytorch_llama3_1_8b_instruct_w8a8, turbomind_qwen2_7b_instruct_4bits, turbomind_qwen2_7b_instruct_kvint8, pytorch_qwen2_7b_instruct_w8a8, turbomind_qwen2_5_7b_instruct_4bits, turbomind_qwen2_5_7b_instruct_kvint8, pytorch_qwen2_5_7b_instruct_w8a8, turbomind_qwen2_5_32b_instruct_4bits, turbomind_qwen2_5_32b_instruct_kvint8, pytorch_qwen2_5_32b_instruct_w8a8,turbomind_llama2_7b_chat_4bits, turbomind_llama2_7b_chat_kvint4, turbomind_llama2_7b_chat_kvint8]'
20+
default: '[turbomind_internlm2_5_7b_chat, pytorch_internlm2_5_7b_chat, turbomind_internlm2_5_7b_chat_batch1, turbomind_internlm2_5_7b_chat_batch1_4bits, turbomind_internlm3_8b_instruct, pytorch_internlm3_8b_instruct, turbomind_internlm2_5_20b_chat, pytorch_internlm2_5_20b_chat, turbomind_qwen1_5_7b_chat, pytorch_qwen1_5_7b_chat, turbomind_llama3_8b_instruct, pytorch_llama3_8b_instruct, turbomind_llama3_1_8b_instruct, pytorch_llama3_1_8b_instruct, turbomind_qwen2_7b_instruct, pytorch_qwen2_7b_instruct, turbomind_qwen2_5_7b_instruct, pytorch_qwen2_5_7b_instruct, turbomind_qwen2_5_32b_instruct, pytorch_qwen2_5_32b_instruct, turbomind_llama2_7b_chat, pytorch_qwen1_5_moe_2_7b_chat, pytorch_gemma_2_9b_it, pytorch_gemma_2_27b_it, turbomind_internlm2_5_7b_chat_4bits, turbomind_internlm2_5_7b_chat_kvint4, turbomind_internlm2_5_7b_chat_kvint8, pytorch_internlm2_5_7b_chat_w8a8, turbomind_internlm3_8b_instruct_4bits, turbomind_internlm3_8b_instruct_kvint4, turbomind_internlm3_8b_instruct_kvint8, pytorch_internlm3_8b_instruct_w8a8, turbomind_llama3_8b_instruct_4bits, turbomind_llama3_8b_instruct_kvint4, turbomind_llama3_1_8b_instruct_4bits, turbomind_llama3_1_8b_instruct_kvint4, turbomind_llama3_1_8b_instruct_kvint8,turbomind_llama3_8b_instruct_kvint8, pytorch_llama3_1_8b_instruct_w8a8, turbomind_qwen2_7b_instruct_kvint8, turbomind_qwen2_5_7b_instruct_4bits, turbomind_qwen2_5_7b_instruct_kvint8, pytorch_qwen2_5_7b_instruct_w8a8, turbomind_qwen2_5_32b_instruct_4bits, turbomind_qwen2_5_32b_instruct_kvint8,turbomind_llama2_7b_chat_4bits, turbomind_llama2_7b_chat_kvint4, turbomind_llama2_7b_chat_kvint8]'
2121
chat_datasets:
2222
required: true
2323
description: 'Tested datasets list. eg. [*bbh_datasets,*ceval_datasets,*cmmlu_datasets,*GaokaoBench_datasets,*gpqa_datasets,*gsm8k_datasets,*hellaswag_datasets,*humaneval_datasets,*ifeval_datasets,*math_datasets,*sanitized_mbpp_datasets,*mmlu_datasets,*nq_datasets,*race_datasets,*TheoremQA_datasets,*triviaqa_datasets,*winogrande_datasets,*crowspairs_datasets]'
@@ -132,8 +132,6 @@ jobs:
132132
name: my-artifact-${{ github.run_id }}-py310
133133
- name: Install lmdeploy - dependency
134134
run: |
135-
# manually install flash attn
136-
# the install packeage from. https://github.com/Dao-AILab/flash-attention/releases
137135
python3 -m pip install -r /root/models/offline_pkg/requirements.txt
138136
- name: Install lmdeploy
139137
if: ${{github.event_name == 'schedule' || !inputs.offline_mode}}

.github/workflows/evaluate_new.yml

Lines changed: 0 additions & 183 deletions
This file was deleted.

0 commit comments

Comments
 (0)