File tree Expand file tree Collapse file tree 2 files changed +11
-3
lines changed Expand file tree Collapse file tree 2 files changed +11
-3
lines changed Original file line number Diff line number Diff line change @@ -62,7 +62,8 @@ model_json=$(cat <<EOF
6262 "enforce_eager": "true",
6363 "enable_lora": "true",
6464 "max_lora_rank": 32,
65- "lora_extra_vocab_size": 256
65+ "lora_extra_vocab_size": 256,
66+ "distributed_executor_backend":"ray"
6667}
6768EOF
6869)
@@ -120,7 +121,8 @@ model_json=$(cat <<EOF
120121 "block_size": 16,
121122 "enforce_eager": "true",
122123 "enable_lora": "false",
123- "lora_extra_vocab_size": 256
124+ "lora_extra_vocab_size": 256,
125+ "distributed_executor_backend":"ray"
124126}
125127EOF
126128)
Original file line number Diff line number Diff line change @@ -63,6 +63,7 @@ function run_multi_gpu_test() {
6363 export KIND=" ${1} "
6464 export TENSOR_PARALLELISM=" ${2} "
6565 export INSTANCE_COUNT=" ${3} "
66+ export DISTRIBUTED_EXECUTOR_BACKEND=" ${4} "
6667
6768 # Setup a clean model repository
6869 export TEST_MODEL=" vllm_opt_${KIND} _tp${TENSOR_PARALLELISM} _count${INSTANCE_COUNT} "
@@ -73,6 +74,10 @@ function run_multi_gpu_test() {
7374 cp -r " ${SAMPLE_MODELS_REPO} /vllm_model" " models/${TEST_MODEL} "
7475 sed -i " s/KIND_MODEL/${KIND} /" " ${TEST_MODEL_TRITON_CONFIG} "
7576 sed -i " 3s/^/ \" tensor_parallel_size\" : ${TENSOR_PARALLELISM} ,\n/" " ${TEST_MODEL_VLLM_CONFIG} "
77+ if [ $TENSOR_PARALLELISM -ne " 1" ]; then
78+ jq --arg backend $DISTRIBUTED_EXECUTOR_BACKEND ' . += {"distributed_executor_backend":$backend}' " ${TEST_MODEL_VLLM_CONFIG} " > " temp.json"
79+ mv temp.json " ${TEST_MODEL_VLLM_CONFIG} "
80+ fi
7681 # Assert the correct kind is set in case the template config changes in the future
7782 validate_file_contains " ${KIND} " " ${TEST_MODEL_TRITON_CONFIG} "
7883
@@ -119,10 +124,11 @@ RET=0
119124KINDS=" KIND_MODEL KIND_GPU"
120125TPS=" 1 2"
121126INSTANCE_COUNTS=" 1 2"
127+ DISTRIBUTED_EXECUTOR_BACKEND=" ray"
122128for kind in ${KINDS} ; do
123129 for tp in ${TPS} ; do
124130 for count in ${INSTANCE_COUNTS} ; do
125- run_multi_gpu_test " ${kind} " " ${tp} " " ${count} "
131+ run_multi_gpu_test " ${kind} " " ${tp} " " ${count} " " ${DISTRIBUTED_EXECUTOR_BACKEND} "
126132 done
127133 done
128134done
You can’t perform that action at this time.
0 commit comments