Skip to content

Commit 01664a2

Browse files
authored
Merge pull request #19 from terrytangyuan/patch-1
Update 2025-01-27-intro-to-llama-stack-with-vllm.md
2 parents d28e882 + 1146ced commit 01664a2

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

_posts/2025-01-27-intro-to-llama-stack-with-vllm.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ Llama Stack defines and standardizes the set of core building blocks needed to b
1616
Llama Stack focuses on making it easy to build production applications with a variety of models - ranging from the latest Llama 3.3 model to specialized models like Llama Guard for safety and other models. The goal is to provide pre-packaged implementations (aka “distributions”) which can be run in a variety of deployment environments. The Stack can assist you in your entire app development lifecycle - start iterating on local, mobile or desktop and seamlessly transition to on-prem or public cloud deployments. At every point in this transition, the same set of APIs and the same developer experience are available.
1717

1818

19-
Each specific implementation of an API is called a "Provider" in this architecture. Users can swap providers via configuration. `vLLM` is a prominent example of a high-performance API backing the inference API.
19+
Each specific implementation of an API is called a "Provider" in this architecture. Users can swap providers via configuration. vLLM is a prominent example of a high-performance API backing the inference API.
2020

2121
# vLLM Inference Provider
2222

0 commit comments

Comments
 (0)