Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 3 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,9 @@ This is achieved by leveraging Envoy's [External Processing] (ext-proc) to exten

[Inference Gateway]:#concepts-and-definitions


> ***NOTE*** : As we prep for our `v1` release, some of our docs may fall out of scope, we are working hard to get these up to date and they will be ready by the time we launch `v1`. Thanks!

## New!
Inference Gateway has partnered with vLLM to accelerate LLM serving optimizations with [llm-d](https://llm-d.ai/blog/llm-d-announce)!

Expand Down
5 changes: 5 additions & 0 deletions site-src/guides/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,11 @@

This project is still in an alpha state and breaking changes may occur in the future.

???+ warning


This page is out of date with the v1.0.0 release candidate. Updates under active development

This quickstart guide is intended for engineers familiar with k8s and model servers (vLLM in this instance). The goal of this guide is to get an Inference Gateway up and running!

## **Prerequisites**
Expand Down
6 changes: 6 additions & 0 deletions site-src/index.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,11 @@
# Introduction

???+ warning


Some portions of this site may be out of date with the v1.0.0 release candidate.
Updates under active development!

Gateway API Inference Extension is an official Kubernetes project that optimizes self-hosting Generative Models on Kubernetes.

The overall resource model focuses on 2 new inference-focused
Expand Down