We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 27ca23d commit 49d849bCopy full SHA for 49d849b
docs/source/index.rst
@@ -70,6 +70,7 @@ Documentation
70
71
serving/distributed_serving
72
serving/run_on_sky
73
+ serving/deploying_with_kserve
74
serving/deploying_with_triton
75
serving/deploying_with_docker
76
serving/serving_with_langchain
docs/source/serving/deploying_with_kserve.rst
@@ -0,0 +1,8 @@
1
+.. _deploying_with_kserve:
2
+
3
+Deploying with KServe
4
+============================
5
6
+vLLM can be deployed with `KServe <https://github.com/kserve/kserve>`_ on Kubernetes for highly scalable distributed model serving.
7
8
+Please see `this guide <https://kserve.github.io/website/latest/modelserving/v1beta1/llm/vllm/>`_ for more details on using vLLM with KServe.
0 commit comments