Skip to content

Latest commit

 

History

History
91 lines (58 loc) · 2.65 KB

File metadata and controls

91 lines (58 loc) · 2.65 KB

vllm-cpu-blueprint

Welcome to the vLLM CPU Blueprint!
Use this to quickly get a vLLM up and running in your environment.
To see how it's done, jump straight to installation.

Description

The vLLM CPU Blueprint is a quick-start template for deploying vLLM on CPU-based infrastructure within Red Hat OpenShift. It’s designed for environments where GPUs are not available or necessary, making it ideal for lightweight inference use cases, prototyping, or constrained environments.
In this Blueprint, we are utilizing it to easily get an LLM deployed in most environments.

This blueprint includes a Helm chart for deploying:

  • An OpenShift AI Project.
  • vLLM with CPU support running an instance of TinyLlama.
  • AnythingLLM (a versitile chat interface) running as a workbench and connected to the vLLM.

Use this project to quickly spin up a minimal vLLM instance and start serving models like TinyLlama on CPU—no GPU required. 🚀

Architecture diagrams

architecture.png

References

Minimum hardware requirements

  • No GPU needed! 🤖
  • 2 cores
  • 8 Gi
  • Storage: 5Gi

Required software

  • Red Hat OpenShift
  • Red Hat OpenShift AI
  • Dependencies for Single-model server:
    • Red Hat OpenShift Service Mesh
    • Red Hat OpenShift Serverless

Required permissions

  • Standard user. No elevated cluster permissions required

Install

Please note before you start

This example was tested on Red Hat OpenShift 4.16.24 & Red Hat OpenShift AI v2.16.2.

Clone:

git clone https://github.com/RHRolun/vllm-cpu-blueprint.git && \
    cd vllm-cpu-blueprint/  

Create the project

PROJECT="tinyllama-cpu-demo"

oc new-project ${PROJECT}

Install:

helm install vllm-cpu . \
    --namespace  ${PROJECT} 

wait for pods:

oc -n ${PROJECT}  get pods -w

Uninstall:

helm uninstall vllm-cpu --namespace ${PROJECT}