Skip to content

Commit c23c72b

Browse files
ariG23498hmellor
andauthored
Apply suggestions from code review
Co-authored-by: Harry Mellor <[email protected]>
1 parent 7cb80a6 commit c23c72b

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

_posts/2025-04-11-transformers-backend.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ vLLM will therefore optimize throughput/latency on top of existing transformers
2020
In this post, we’ll explore how vLLM leverages the transformers backend to combine **flexibility**
2121
with **efficiency**, enabling you to deploy state-of-the-art models faster and smarter.
2222

23-
## transformers and vLLM: Inference in Action
23+
## Transformers and vLLM: Inference in Action
2424

2525
Let’s start with a simple text generation task using the `meta-llama/Llama-3.2-1B` model to see how
2626
these libraries stack up.
@@ -92,7 +92,7 @@ print("Completion result:", completion.choices[0].text)
9292

9393
This compatibility slashes costs and boosts control, letting you scale inference locally with vLLM’s optimizations.
9494

95-
## Why need the transformers backend?
95+
## Why do we need the transformers backend?
9696

9797
The transformers library is optimized for contributions and
9898
[addition of new models](https://huggingface.co/docs/transformers/en/add_new_model). Adding a new

0 commit comments

Comments
 (0)