Skip to content

Commit a9cfe50

Browse files
authored
[Misc] Updating default LMCache docker image for remote KV cache sharing (#384)
* Update helm chart template for newest vllm version Signed-off-by: YuhanLiu11 <yliu738@wisc.edu> * Update shared storage yaml file Signed-off-by: YuhanLiu11 <yliu738@wisc.edu> --------- Signed-off-by: YuhanLiu11 <yliu738@wisc.edu>
1 parent 0d4564c commit a9cfe50

File tree

2 files changed

+10
-4
lines changed

2 files changed

+10
-4
lines changed

helm/templates/deployment-vllm-multi.yaml

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -77,8 +77,11 @@ spec:
7777
{{- end }}
7878
{{- with $modelSpec.vllmConfig }}
7979
{{- if hasKey . "enableChunkedPrefill" }}
80+
{{- if .enableChunkedPrefill }}
8081
- "--enable-chunked-prefill"
81-
- {{ .enableChunkedPrefill | quote }}
82+
{{- else }}
83+
- "--no-enable-chunked-prefill"
84+
{{- end }}
8285
{{- end }}
8386
{{- if .enablePrefixCaching }}
8487
- "--enable-prefix-caching"

tutorials/assets/values-06-shared-storage.yaml

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ servingEngineSpec:
33
modelSpec:
44
- name: "mistral"
55
repository: "lmcache/vllm-openai"
6-
tag: "latest"
6+
tag: "2025-04-18"
77
modelURL: "mistralai/Mistral-7B-Instruct-v0.2"
88
replicaCount: 2
99
requestCPU: 10
@@ -14,11 +14,14 @@ servingEngineSpec:
1414
enableChunkedPrefill: false
1515
enablePrefixCaching: false
1616
maxModelLen: 16384
17+
v1: 0
1718

1819
lmcacheConfig:
1920
enabled: true
2021
cpuOffloadingBufferSize: "20"
21-
22+
env:
23+
- name: LMCACHE_LOG_LEVEL
24+
value: "DEBUG"
2225
hf_token: <YOUR HF TOKEN>
2326

2427
cacheserverSpec:
@@ -36,7 +39,7 @@ cacheserverSpec:
3639

3740
# -- Cache server image (reusing the vllm image)
3841
repository: "lmcache/vllm-openai"
39-
tag: "latest"
42+
tag: "2025-04-18"
4043

4144
# TODO (Jiayi): please adjust this once we have evictor
4245
# -- router resource requests and limits

0 commit comments

Comments
 (0)