1919 required : false
2020 type : string
2121 default : " latest"
22+ vllmVersion :
23+ description : ' vLLM version'
24+ required : false
25+ type : string
26+ default : " 0.11.0"
2227
2328jobs :
2429 test :
2530 runs-on : ubuntu-latest
2631 steps :
2732 - name : Checkout code
28- uses : actions/checkout@v4
33+ uses : actions/checkout@08eba0b27e820071cde6df949e0beb9ba4906955
2934
3035 - name : Set up Go
31- uses : actions/setup-go@v5
36+ uses : actions/setup-go@d35c59abb061a4a6fb18e82ac0862c26744d6ab5
3237 with :
3338 go-version : 1.24.2
3439 cache : true
4146 runs-on : ubuntu-latest
4247 steps :
4348 - name : Checkout repo
44- uses : actions/checkout@v4
49+ uses : actions/checkout@08eba0b27e820071cde6df949e0beb9ba4906955
4550
4651 - name : Format tags
4752 id : tags
@@ -59,25 +64,32 @@ jobs:
5964 echo "docker/model-runner:latest-cuda" >> "$GITHUB_OUTPUT"
6065 fi
6166 echo 'EOF' >> "$GITHUB_OUTPUT"
67+ echo "vllm-cuda<<EOF" >> "$GITHUB_OUTPUT"
68+ echo "docker/model-runner:${{ inputs.releaseTag }}-vllm-cuda" >> "$GITHUB_OUTPUT"
69+ if [ "${{ inputs.pushLatest }}" == "true" ]; then
70+ echo "docker/model-runner:latest-vllm-cuda" >> "$GITHUB_OUTPUT"
71+ fi
72+ echo 'EOF' >> "$GITHUB_OUTPUT"
6273
6374 - name : Log in to DockerHub
64- uses : docker/login-action@v3
75+ uses : docker/login-action@5e57cd118135c172c3672efd75eb46360885c0ef
6576 with :
6677 username : " docker"
6778 password : ${{ secrets.ORG_ACCESS_TOKEN }}
6879
6980 - name : Set up Buildx
70- uses : docker/setup-buildx-action@v3
81+ uses : docker/setup-buildx-action@e468171a9de216ec08956ac3ada2f0791b6bd435
7182 with :
7283 version : " lab:latest"
7384 driver : cloud
7485 endpoint : " docker/make-product-smarter"
7586 install : true
7687
7788 - name : Build CPU image
78- uses : docker/build-push-action@v5
89+ uses : docker/build-push-action@ca052bb54ab0790a636c9b5f226502c73d547a25
7990 with :
8091 file : Dockerfile
92+ target : final-llamacpp
8193 platforms : linux/amd64, linux/arm64
8294 build-args : |
8395 "LLAMA_SERVER_VERSION=${{ inputs.llamaServerVersion }}"
8799 tags : ${{ steps.tags.outputs.cpu }}
88100
89101 - name : Build CUDA image
90- uses : docker/build-push-action@v5
102+ uses : docker/build-push-action@ca052bb54ab0790a636c9b5f226502c73d547a25
91103 with :
92104 file : Dockerfile
105+ target : final-llamacpp
93106 platforms : linux/amd64, linux/arm64
94107 build-args : |
95108 "LLAMA_SERVER_VERSION=${{ inputs.llamaServerVersion }}"
@@ -99,3 +112,19 @@ jobs:
99112 sbom : true
100113 provenance : mode=max
101114 tags : ${{ steps.tags.outputs.cuda }}
115+
116+ - name : Build vLLM CUDA image
117+ uses : docker/build-push-action@ca052bb54ab0790a636c9b5f226502c73d547a25
118+ with :
119+ file : Dockerfile
120+ target : final-vllm
121+ platforms : linux/amd64
122+ build-args : |
123+ "LLAMA_SERVER_VERSION=${{ inputs.llamaServerVersion }}"
124+ "LLAMA_SERVER_VARIANT=cuda"
125+ "BASE_IMAGE=nvidia/cuda:12.9.0-runtime-ubuntu24.04"
126+ "VLLM_VERSION=${{ inputs.vllmVersion }}"
127+ push : true
128+ sbom : true
129+ provenance : mode=max
130+ tags : ${{ steps.tags.outputs.vllm-cuda }}
0 commit comments