Skip to content

Commit fc119ce

Browse files
authored
chore: 🔨 Use the v0.3.0 llm-d-inference-sim image tag. (#1140)
* chore: 🔨 Use the latest llm-d-inference-sim image tag. * feat: 🔨 Compatible with llm-d-inference-sim v0.3.0. * chore: ⬆️ bump llm-d-inference-sim to v0.3.0 (default). * chore: ⬆️ replace with v0.3.0 tag.
1 parent b89ab9b commit fc119ce

File tree

3 files changed

+5
-5
lines changed

3 files changed

+5
-5
lines changed

.github/ISSUE_TEMPLATE/new-release.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ This document defines the process for releasing Gateway API Inference Extension.
4242
```shell
4343
export VLLM_GPU=0.9.2
4444
export VLLM_CPU=0.9.3
45-
export VLLM_SIM=0.1.2
45+
export VLLM_SIM=0.3.0
4646
```
4747

4848
## Release Process

config/manifests/vllm/sim-deployment.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ spec:
1414
spec:
1515
containers:
1616
- name: vllm-sim
17-
image: ghcr.io/llm-d/llm-d-inference-sim:v0.1.2
17+
image: ghcr.io/llm-d/llm-d-inference-sim:v0.3.0
1818
imagePullPolicy: Always
1919
args:
2020
- --model
@@ -23,8 +23,8 @@ spec:
2323
- "8000"
2424
- --max-loras
2525
- "2"
26-
- --lora
27-
- food-review-1
26+
- --lora-modules
27+
- '{"name": "food-review-1"}'
2828
env:
2929
- name: POD_NAME
3030
valueFrom:

hack/release-quickstart.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ VLLM_GPU="${VLLM_GPU:-0.9.1}"
3535
# The CPU image is from https://gallery.ecr.aws/q9t5s3a7/vllm-cpu-release-repo
3636
VLLM_CPU="${VLLM_CPU:-0.9.1}"
3737
# The sim image is from https://github.com/llm-d/llm-d-inference-sim/pkgs/container/llm-d-inference-sim
38-
VLLM_SIM="${VLLM_SIM:-0.1.1}"
38+
VLLM_SIM="${VLLM_SIM:-0.3.0}"
3939

4040
echo "Using release tag: ${RELEASE_TAG}"
4141
echo "Using vLLM GPU image version: ${VLLM_GPU}"

0 commit comments

Comments
 (0)