Skip to content

Commit 62ddc97

Browse files
committed
feat: add sglanf backend to release workflow
1 parent b295064 commit 62ddc97

File tree

2 files changed

+38
-6
lines changed

2 files changed

+38
-6
lines changed

.github/workflows/promote-to-latest.yml

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ on:
55
workflow_dispatch:
66
inputs:
77
version:
8-
description: 'version'
8+
description: "version"
99
required: true
1010
type: string
1111

@@ -42,6 +42,11 @@ jobs:
4242
echo "Promoting vLLM CUDA images"
4343
crane tag "docker/model-runner:${{ inputs.version }}-vllm-cuda" "latest-vllm-cuda"
4444
45+
- name: Promote SGLang CUDA images
46+
run: |
47+
echo "Promoting SGLang CUDA images"
48+
crane tag "docker/model-runner:${{ inputs.version }}-sglang-cuda" "latest-sglang-cuda"
49+
4550
- name: Promote ROCm images
4651
run: |
4752
echo "Promoting ROCm images"

.github/workflows/release.yml

Lines changed: 32 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -5,28 +5,33 @@ on:
55
workflow_dispatch:
66
inputs:
77
pushLatest:
8-
description: 'Tag images produced by this job as latest'
8+
description: "Tag images produced by this job as latest"
99
required: false
1010
type: boolean
1111
default: false
1212
releaseTag:
13-
description: 'Release tag'
13+
description: "Release tag"
1414
required: false
1515
type: string
1616
default: "test"
1717
llamaServerVersion:
18-
description: 'llama-server version'
18+
description: "llama-server version"
1919
required: false
2020
type: string
2121
default: "latest"
2222
vllmVersion:
23-
description: 'vLLM version'
23+
description: "vLLM version"
2424
required: false
2525
type: string
2626
default: "0.12.0"
27+
sglangVersion:
28+
description: "SGLang version"
29+
required: false
30+
type: string
31+
default: "0.4.0"
2732
# This can be removed once we have llama.cpp built for MUSA and CANN.
2833
buildMusaCann:
29-
description: 'Build MUSA and CANN images'
34+
description: "Build MUSA and CANN images"
3035
required: false
3136
type: boolean
3237
default: false
@@ -76,6 +81,12 @@ jobs:
7681
echo "docker/model-runner:latest-vllm-cuda" >> "$GITHUB_OUTPUT"
7782
fi
7883
echo 'EOF' >> "$GITHUB_OUTPUT"
84+
echo "sglang-cuda<<EOF" >> "$GITHUB_OUTPUT"
85+
echo "docker/model-runner:${{ inputs.releaseTag }}-sglang-cuda" >> "$GITHUB_OUTPUT"
86+
if [ "${{ inputs.pushLatest }}" == "true" ]; then
87+
echo "docker/model-runner:latest-sglang-cuda" >> "$GITHUB_OUTPUT"
88+
fi
89+
echo 'EOF' >> "$GITHUB_OUTPUT"
7990
echo "rocm<<EOF" >> "$GITHUB_OUTPUT"
8091
echo "docker/model-runner:${{ inputs.releaseTag }}-rocm" >> "$GITHUB_OUTPUT"
8192
if [ "${{ inputs.pushLatest }}" == "true" ]; then
@@ -155,6 +166,22 @@ jobs:
155166
provenance: mode=max
156167
tags: ${{ steps.tags.outputs.vllm-cuda }}
157168

169+
- name: Build SGLang CUDA image
170+
uses: docker/build-push-action@263435318d21b8e681c14492fe198d362a7d2c83
171+
with:
172+
file: Dockerfile
173+
target: final-sglang
174+
platforms: linux/amd64, linux/arm64
175+
build-args: |
176+
"LLAMA_SERVER_VERSION=${{ inputs.llamaServerVersion }}"
177+
"LLAMA_SERVER_VARIANT=cuda"
178+
"BASE_IMAGE=nvidia/cuda:12.9.0-runtime-ubuntu24.04"
179+
"SGLANG_VERSION=${{ inputs.sglangVersion }}"
180+
push: true
181+
sbom: true
182+
provenance: mode=max
183+
tags: ${{ steps.tags.outputs.sglang-cuda }}
184+
158185
- name: Build ROCm image
159186
uses: docker/build-push-action@263435318d21b8e681c14492fe198d362a7d2c83
160187
with:

0 commit comments

Comments
 (0)