Skip to content

Commit ed25ed3

Browse files
authored
fix dead links (#989)
1 parent aa4989e commit ed25ed3

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

site-src/index.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ The following specific terms to this project:
2020
inference workloads.
2121
- **Inference Scheduler**: An extendable component that makes decisions about which endpoint is optimal (best cost /
2222
best performance) for an inference request based on `Metrics and Capabilities`
23-
from [Model Serving](/docs/proposals/003-model-server-protocol/README.md).
23+
from [Model Serving](https://github.com/kubernetes-sigs/gateway-api-inference-extension/tree/main/docs/proposals/003-model-server-protocol/README.md).
2424
- **Metrics and Capabilities**: Data provided by model serving platforms about
2525
performance, availability and capabilities to optimize routing. Includes
2626
things like [Prefix Cache] status or [LoRA Adapters] availability.
@@ -33,8 +33,8 @@ Gateway API Inference Extension optimizes self-hosting Generative AI Models on K
3333
It provides optimized load-balancing for self-hosted Generative AI Models on Kubernetes.
3434
The project’s goal is to improve and standardize routing to inference workloads across the ecosystem.
3535

36-
This is achieved by leveraging Envoy's [External Processing](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_filters/ext_proc_filter) to extend any gateway that supports both ext-proc and [Gateway API](https://github.com/kubernetes-sigs/gateway-api) into an [inference gateway](../index.md#concepts-and-definitions).
37-
This extension extends popular gateways like Envoy Gateway, kgateway, and GKE Gateway - to become [Inference Gateway](../index.md#concepts-and-definitions) -
36+
This is achieved by leveraging Envoy's [External Processing](https://www.envoyproxy.io/docs/envoy/latest/configuration/http/http_filters/ext_proc_filter) to extend any gateway that supports both ext-proc and [Gateway API](https://github.com/kubernetes-sigs/gateway-api) into an [inference gateway](#concepts-and-definitions).
37+
This extension extends popular gateways like Envoy Gateway, kgateway, and GKE Gateway - to become [Inference Gateway](#concepts-and-definitions) -
3838
supporting inference platform teams self-hosting Generative Models (with a current focus on large language models) on Kubernetes.
3939
This integration makes it easy to expose and control access to your local [OpenAI-compatible chat completion endpoints](https://platform.openai.com/docs/api-reference/chat)
4040
to other workloads on or off cluster, or to integrate your self-hosted models alongside model-as-a-service providers

0 commit comments

Comments
 (0)