File tree Expand file tree Collapse file tree 3 files changed +6
-6
lines changed Expand file tree Collapse file tree 3 files changed +6
-6
lines changed Original file line number Diff line number Diff line change @@ -42,7 +42,7 @@ This document defines the process for releasing Gateway API Inference Extension.
4242 ``` shell
4343 export VLLM_GPU=0.9.2
4444 export VLLM_CPU=0.9.3
45- export VLLM_SIM=0.1.2
45+ export VLLM_SIM=0.3.0
4646 ```
4747
4848## Release Process
Original file line number Diff line number Diff line change @@ -14,17 +14,17 @@ spec:
1414 spec :
1515 containers :
1616 - name : vllm-sim
17- image : ghcr.io/llm-d/llm-d-inference-sim:v0.1.2
18- imagePullPolicy : IfNotPresent
17+ image : ghcr.io/llm-d/llm-d-inference-sim:v0.3.0
18+ imagePullPolicy : Always
1919 args :
2020 - --model
2121 - meta-llama/Llama-3.1-8B-Instruct
2222 - --port
2323 - " 8000"
2424 - --max-loras
2525 - " 2"
26- - --lora
27- - food-review-1
26+ - --lora-modules
27+ - ' {"name": " food-review-1"} '
2828 env :
2929 - name : POD_NAME
3030 valueFrom :
Original file line number Diff line number Diff line change @@ -35,7 +35,7 @@ VLLM_GPU="${VLLM_GPU:-0.9.1}"
3535# The CPU image is from https://gallery.ecr.aws/q9t5s3a7/vllm-cpu-release-repo
3636VLLM_CPU=" ${VLLM_CPU:- 0.9.1} "
3737# The sim image is from https://github.com/llm-d/llm-d-inference-sim/pkgs/container/llm-d-inference-sim
38- VLLM_SIM=" ${VLLM_SIM:- 0.1.1 } "
38+ VLLM_SIM=" ${VLLM_SIM:- 0.3.0 } "
3939
4040echo " Using release tag: ${RELEASE_TAG} "
4141echo " Using vLLM GPU image version: ${VLLM_GPU} "
You can’t perform that action at this time.
0 commit comments