Skip to content

Commit ab0a471

Browse files
committed
vllm korea meetup
Signed-off-by: rebel-jiwonk <[email protected]>
1 parent ec14814 commit ab0a471

File tree

1 file changed

+0
-1
lines changed

1 file changed

+0
-1
lines changed

_posts/2025-09-10-vllm-meetup.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,6 @@ The event brought together local developers, researchers, and AI infrastructure
2727
</picture><br>
2828
</p>
2929

30-
3130
Nicolò Lucchesi, Senior ML Engineer at Red Hat, opened the event by highlighting the original innovation behind vLLM — solving long-standing challenges in KV caching and dynamic batching with a novel paged attention architecture. He emphasized that “modern problems require traditional solutions,” noting that the exact challenges in scheduling and memory management had already been tackled in operating systems, and vLLM simply applies the same proven ideas to AI inference.
3231

3332
He also introduced llm-d, a project enabling distributed inference. llm-d is a Kubernetes-native orchestration layer that coordinates multiple vLLM instances with auto-scaling support — “vLLM meeting Kubernetes.”

0 commit comments

Comments
 (0)