Skip to content

Commit 1d63610

Browse files
committed
Update L0_multi_gpu_vllm script in order to be able access resources … (#97)
1 parent 2d0f250 commit 1d63610

File tree

2 files changed

+6
-0
lines changed

2 files changed

+6
-0
lines changed

ci/L0_multi_gpu_vllm/multi_lora/test.sh

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -41,6 +41,9 @@ EXPECTED_NUM_TESTS=2
4141
GENERATE_ENDPOINT="localhost:8000/v2/models/vllm_llama_multi_lora/generate"
4242
CHECK_FOR_ERROR=true
4343

44+
export C_INCLUDE_PATH=/usr/local/cuda/include:$C_INCLUDE_PATH
45+
export TRITON_PTXAS_PATH=/usr/local/cuda/bin/ptxas
46+
4447
make_api_call() {
4548
local endpoint="$1"
4649
local data="$2"

ci/L0_multi_gpu_vllm/vllm_backend/test.sh

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -36,6 +36,9 @@ CLIENT_PY="./vllm_multi_gpu_test.py"
3636
SAMPLE_MODELS_REPO="../../../samples/model_repository"
3737
EXPECTED_NUM_TESTS=1
3838

39+
export C_INCLUDE_PATH=/usr/local/cuda/include:$C_INCLUDE_PATH
40+
export TRITON_PTXAS_PATH=/usr/local/cuda/bin/ptxas
41+
3942
### Helpers
4043
function validate_file_contains() {
4144
local KEY="${1}"

0 commit comments

Comments
 (0)