Skip to content

Commit 6b26360

Browse files
committed
updates
1 parent a13de02 commit 6b26360

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

articles/ai-foundry/openai/how-to/spillover-traffic-management.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ author: aahill # GitHub alias
55
ms.author: aahi
66
ms.service: azure-ai-openai
77
ms.topic: how-to
8-
ms.date: 05/20/2025
8+
ms.date: 08/27/2025
99
---
1010

1111
# Manage traffic with spillover for provisioned deployments
@@ -16,14 +16,14 @@ Spillover manages traffic fluctuations on provisioned deployments by routing ove
1616
> Spillover is currently not available for the `/v1` [API endpoint](../reference-preview-latest.md).
1717
1818
## Prerequisites
19-
- Two separate deployments. The primary deployment can be any (standard, data zone) provisioned type. The spillover deployment can be a provisioned deployment of the same type as the primary deployment, or a pay-as-you-go deployment.
19+
- You need to have a provisioned managed deployment and a standard deployment.
2020

2121
- The provisioned and standard deployments must be in the same Azure OpenAI resource to be eligible for spillover.
2222

2323
- The data processing level of your standard deployment must match your provisioned deployment (for example, a global provisioned deployment must be used with a global standard spillover deployment).
2424

2525
## When to enable spillover on provisioned deployments
26-
To maximize the utilization of your provisioned deployment, you can enable spillover for all global and data zone provisioned deployments. With spillover, bursts or fluctuations in traffic can be automatically managed by the service. This capability reduces the risk of experiencing disruptions when a provisioned deployment is fully utilized. Alternatively, spillover is configurable per-request to provide flexibility across different scenarios and workloads.
26+
To maximize the utilization of your provisioned deployment, you can enable spillover for all global and data zone provisioned deployments. With spillover, bursts or fluctuations in traffic can be automatically managed by the service. This capability reduces the risk of experiencing disruptions when a provisioned deployment is fully utilized. Alternatively, spillover is configurable per-request to provide flexibility across different scenarios and workloads. Spillover can be used for the [Azure AI Foundry Agent Service](../../agents/overview.md) and [responses API](./responses.md).
2727

2828
## When does spillover come into effect?
2929
When spillover is enabled for a deployment or configured for a given inference request, spillover is initiated when a non-200 response code is received for a given inference request. When a request results in a non-200 response code, the Azure OpenAI automatically sends the request from your provisioned deployment to your standard deployment to be processed. Even if a subset of requests is routed to the standard deployment, the service prioritizes sending requests to the provisioned deployment before sending any overage requests to the standard deployment, which may incur additional latency.

0 commit comments

Comments
 (0)