Skip to content

Commit 49d849b

Browse files
docs: Add tutorial on deploying vLLM model with KServe (#2586)
Signed-off-by: Yuan Tang <[email protected]>
1 parent 27ca23d commit 49d849b

File tree

2 files changed

+9
-0
lines changed

2 files changed

+9
-0
lines changed

docs/source/index.rst

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -70,6 +70,7 @@ Documentation
7070

7171
serving/distributed_serving
7272
serving/run_on_sky
73+
serving/deploying_with_kserve
7374
serving/deploying_with_triton
7475
serving/deploying_with_docker
7576
serving/serving_with_langchain
Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,8 @@
1+
.. _deploying_with_kserve:
2+
3+
Deploying with KServe
4+
============================
5+
6+
vLLM can be deployed with `KServe <https://github.com/kserve/kserve>`_ on Kubernetes for highly scalable distributed model serving.
7+
8+
Please see `this guide <https://kserve.github.io/website/latest/modelserving/v1beta1/llm/vllm/>`_ for more details on using vLLM with KServe.

0 commit comments

Comments
 (0)