@@ -36,8 +36,12 @@ steps:
3636 commands :
3737 - |
3838 timeout 20m bash -c '
39- pytest -s -v tests/e2e/offline_inference/test_diffusion_cpu_offload.py &&
39+ set +e
40+ pytest -s -v tests/e2e/offline_inference/test_diffusion_cpu_offload.py
41+ EXIT1=$$?
4042 pytest -s -v tests/e2e/offline_inference/test_diffusion_layerwise_offload.py
43+ EXIT2=$$?
44+ exit $$((EXIT1 | EXIT2))
4145 '
4246 agents :
4347 queue : " gpu_1_queue" # g6.4xlarge instance on AWS, has 1 L4 GPU
@@ -117,43 +121,33 @@ steps:
117121 volumes :
118122 - " /fsx/hf_cache:/fsx/hf_cache"
119123
120- # - label: "Benchmark & Engine Test with H100"
121- # depends_on: upload-ready-pipeline
122- # commands:
123- # - |
124- # timeout 15m bash -c '
125- # export VLLM_WORKER_MULTIPROC_METHOD=spawn
126- # pytest -s -v tests/benchmarks/test_serve_cli.py
127- # pytest -s -v tests/engine/test_async_omni_engine_abort.py
128- # '
129- # agents:
130- # queue: "mithril-h100-pool"
131- # plugins:
132- # - kubernetes:
133- # podSpec:
134- # containers:
135- # - image: 936637512419.dkr.ecr.us-west-2.amazonaws.com/vllm-ci-pull-through-cache/q9t5s3a7/vllm-ci-test-repo:$BUILDKITE_COMMIT
136- # resources:
137- # limits:
138- # nvidia.com/gpu: 2
139- # volumeMounts:
140- # - name: devshm
141- # mountPath: /dev/shm
142- # - name: hf-cache
143- # mountPath: /root/.cache/huggingface
144- # env:
145- # - name: HF_HOME
146- # value: /root/.cache/huggingface
147- # nodeSelector:
148- # node.kubernetes.io/instance-type: gpu-h100-sxm
149- # volumes:
150- # - name: devshm
151- # emptyDir:
152- # medium: Memory
153- # - name: hf-cache
154- # hostPath:
155- # path: /mnt/hf-cache
156- # type: DirectoryOrCreate
124+
125+ - label : " Benchmark & Engine Test"
126+ depends_on : upload-ready-pipeline
127+ commands :
128+ - |
129+ timeout 15m bash -c '
130+ export VLLM_WORKER_MULTIPROC_METHOD=spawn
131+ set +e
132+ pytest -s -v tests/benchmarks/test_serve_cli.py
133+ EXIT1=$$?
134+ pytest -s -v tests/engine/test_async_omni_engine_abort.py
135+ EXIT2=$$?
136+ exit $$((EXIT1 | EXIT2))
137+ '
138+ agents :
139+ queue : " gpu_4_queue" # g6.12xlarge instance on AWS, has 4 L4 GPU
140+ plugins :
141+ - docker#v5.2.0:
142+ image : public.ecr.aws/q9t5s3a7/vllm-ci-test-repo:$BUILDKITE_COMMIT
143+ always-pull : true
144+ propagate-environment : true
145+ shm-size : " 8gb"
146+ environment :
147+ - " HF_HOME=/fsx/hf_cache"
148+ volumes :
149+ - " /fsx/hf_cache:/fsx/hf_cache"
150+
157151
158152 - label : " Omni Model Test"
159153 depends_on : upload-ready-pipeline
0 commit comments