Skip to content

Commit 12cde2b

Browse files
authored
Merge branch 'main' into 361124-translator-january-freshness
2 parents 72f5d50 + af56131 commit 12cde2b

File tree

69 files changed

+438
-155
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

69 files changed

+438
-155
lines changed

articles/ai-foundry/model-inference/concepts/deployment-types.md

Lines changed: 19 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
title: Understanding deployment types in Azure AI model inference
33
titleSuffix: Azure AI Foundry
44
description: Learn how to use deployment types in Azure AI model deployments
5-
author: mrbullwinkle
5+
author: santiagxf
66
manager: nitinme
77
ms.service: azure-ai-model-inference
88
ms.topic: how-to
@@ -13,43 +13,36 @@ ms.custom: ignite-2024, github-universe-2024
1313

1414
# Deployment types in Azure AI model inference
1515

16-
Azure AI model inference in Azure AI services provides customers with choices on the hosting structure that fits their business and usage patterns. The service offers two main types of deployment: **standard** and **provisioned**. Standard is offered with a global deployment option, routing traffic globally to provide higher throughput. Provisioned is also offered with a global deployment option, allowing customers to purchase and deploy provisioned throughput units across Azure global infrastructure.
16+
Azure AI model inference makes models available using the *model deployment* concept in Azure AI Services resources. *Model deployments* are also Azure resources and, when created, they give access to a given model under certain configurations. Such configuration includes the infrastructure require to process the requests.
1717

18-
All deployments can perform the exact same inference operations, however the billing, scale, and performance are substantially different. As part of your solution design, you need to make two key decisions:
18+
Azure AI model inference provides customers with choices on the hosting structure that fits their business and usage patterns. Those options are translated to different deployments types (or SKUs) that are available at model deployment time in the Azure AI Services resource.
1919

20-
- **Data residency needs**: global vs. regional resources
21-
- **Call volume**: standard vs. provisioned
20+
:::image type="content" source="../media/add-model-deployments/models-deploy-deployment-type.png" alt-text="Screenshot showing how to customize the deployment type for a given model deployment." lightbox="../media/add-model-deployments/models-deploy-deployment-type.png":::
2221

23-
Deployment types support varies by model and model provider. You can see which deployment type (SKU) each model supports in the [Models section](models.md).
22+
Different model providers offer different deployments SKUs that you can select from. When selecting a deployment type, consider your **data residency needs** and **call volume/capacity** requirements.
2423

25-
## Global versus regional deployment types
24+
## Deployment types for Azure OpenAI models
2625

27-
For standard and provisioned deployments, you have an option of two types of configurations within your resource – **global** or **regional**. Global standard is the recommended starting point.
26+
The service offers two main types of deployments: **standard** and **provisioned**. For a given deployment type, customers can align their workloads with their data processing requirements by choosing an Azure geography (`Standard` or `Provisioned-Managed`), Microsoft specified data zone (`DataZone-Standard` or `DataZone Provisioned-Managed`), or Global (`Global-Standard` or `Global Provisioned-Managed`) processing options.
2827

29-
Global deployments leverage Azure's global infrastructure, dynamically route customer traffic to the data center with best availability for the customer's inference requests. This means you get the highest initial throughput limits and best model availability with Global while still providing our uptime SLA and low latency. For high volume workloads above the specified usage tiers on standard and global standard, you may experience increased latency variation. For customers that require the lower latency variance at large workload usage, we recommend purchasing provisioned throughput.
28+
To learn more about deployment options for Azure OpenAI models see [Azure OpenAI documentation](../../../ai-services/openai/how-to/deployment-types.md).
3029

31-
Our global deployments are the first location for all new models and features. Customers with large throughput requirements should consider our provisioned deployment offering.
30+
## Deployment types for Models-as-a-Service models
3231

33-
## Standard
32+
Models from third-party model providers with pay-as-you-go billing (collectively called Models-as-a-Service), makes models available in Azure AI model inference under **standard** deployments with a Global processing option (`Global-Standard`).
3433

35-
Standard deployments provide a pay-per-call billing model on the chosen model. Provides the fastest way to get started as you only pay for what you consume. Models available in each region and throughput may be limited.
34+
### Global-Standard
3635

37-
Standard deployments are optimized for low to medium volume workloads with high burstiness. Customers with high consistent volume may experience greater latency variability.
36+
Global deployments leverage Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global standard provides the highest default quota and eliminates the need to load balance across multiple resources. Data stored at rest remains in the designated Azure geography, while data may be processed for inferencing in any Azure location. Learn more about [data residency](https://azure.microsoft.com/explore/global-infrastructure/data-residency/).
3837

39-
Only Azure OpenAI models support this deployment type.
38+
> [!NOTE]
39+
> Models-as-a-Service offers regional deployment options under [Serverless API endpoints](../../../ai-studio/how-to/deploy-models-serverless.md) in Azure AI Foundry. Prompts and outputs are processed within the geography specified during deployment. However, those deployments can't be accessed using the Azure AI model inference endpoint in Azure AI Services.
4040
41-
## Global standard
41+
## Control deployment options
4242

43-
Global deployments are available in the same Azure AI services resources as non-global deployment types but allow you to leverage Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global standard provides the highest default quota and eliminates the need to load balance across multiple resources.
43+
Administrators can control which model deployment types are available to their users by using Azure Policies. Learn more about [How to control AI model deployment with custom policies](../../../ai-studio/how-to/custom-policy-model-deployment.md).
4444

45-
Customers with high consistent volume may experience greater latency variability. The threshold is set per model. For applications that require the lower latency variance at large workload usage, we recommend purchasing provisioned throughput if available.
45+
## Related content
4646

47-
## Global provisioned
48-
49-
Global deployments are available in the same Azure AI services resources as non-global deployment types but allow you to leverage Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global provisioned deployments provide reserved model processing capacity for high and predictable throughput using Azure global infrastructure.
50-
51-
Only Azure OpenAI models support this deployment type.
52-
53-
## Next steps
54-
55-
- [Quotas & limits](../quotas-limits.md)
47+
- [Quotas & limits](../quotas-limits.md)
48+
- [Data privacy, and security for Models-as-a-Service models](../../../ai-studio/how-to/concept-data-privacy.md)

articles/ai-foundry/model-inference/concepts/endpoints.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
title: Model inference endpoint in Azure AI services
33
titleSuffix: Azure AI Foundry
44
description: Learn about the model inference endpoint in Azure AI services
5-
author: mrbullwinkle
5+
author: santiagxf
66
manager: nitinme
77
ms.service: azure-ai-model-inference
88
ms.topic: how-to

articles/ai-foundry/model-inference/faq.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ metadata:
55
description: Get answers to the most popular questions about Azure AI model inference
66
#services: cognitive-services
77
manager: nitinme
8-
ms.service: azure-ai-models
8+
ms.service: azure-ai-model-inference
99
ms.topic: faq
1010
ms.date: 1/21/2025
1111
ms.author: fasantia

articles/ai-foundry/model-inference/how-to/configure-content-filters.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,8 +6,8 @@ manager: nitinme
66
ms.service: azure-ai-model-inference
77
ms.topic: how-to
88
ms.date: 1/21/2025
9-
author: mrbullwinkle
10-
ms.author: mbullwin
9+
author: santiagxf
10+
ms.author: fasantia
1111
recommendations: false
1212
ms.custom: ignite-2024, github-universe-2024
1313
zone_pivot_groups: azure-ai-models-deployment

articles/ai-foundry/model-inference/how-to/configure-entra-id.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ ms.topic: how-to
77
ms.date: 10/01/2024
88
ms.custom: ignite-2024, github-universe-2024
99
manager: nitinme
10-
author: mrbullwinkle
10+
author: santiagxf
1111
ms.author: fasantia
1212
recommendations: false
1313
zone_pivot_groups: azure-ai-models-deployment

articles/ai-foundry/model-inference/how-to/configure-project-connection.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ ms.topic: how-to
77
ms.date: 1/21/2025
88
ms.custom: ignite-2024, github-universe-2024
99
manager: nitinme
10-
author: mrbullwinkle
10+
author: santiagxf
1111
ms.author: fasantia
1212
recommendations: false
1313
zone_pivot_groups: azure-ai-models-deployment

articles/ai-foundry/model-inference/how-to/create-model-deployments.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ ms.topic: how-to
77
ms.date: 1/21/2025
88
ms.custom: ignite-2024, github-universe-2024
99
manager: nitinme
10-
author: mrbullwinkle
10+
author: santiagxf
1111
ms.author: fasantia
1212
recommendations: false
1313
zone_pivot_groups: azure-ai-models-deployment

articles/ai-foundry/model-inference/how-to/github/create-model-deployments.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ ms.topic: how-to
77
ms.date: 1/21/2025
88
ms.custom: ignite-2024, github-universe-2024
99
manager: nitinme
10-
author: mrbullwinkle
10+
author: santiagxf
1111
ms.author: fasantia
1212
recommendations: false
1313
---

articles/ai-foundry/model-inference/how-to/manage-costs.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
11
---
22
title: Plan to manage costs for model inference in Azure AI Services
33
description: Learn how to plan for and manage costs for Azure AI model inference in Azure AI Services by using cost analysis in the Azure portal.
4-
author: mrbullwinkle
5-
ms.author: mbullwin
4+
author: santiagxf
5+
ms.author: fasantia
66
ms.custom: subject-cost-optimization
77
ms.service: azure-ai-model-inference
88
ms.topic: how-to

articles/ai-foundry/model-inference/how-to/quickstart-ai-project.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ ms.topic: how-to
77
ms.date: 1/21/2025
88
ms.custom: ignite-2024, github-universe-2024
99
manager: nitinme
10-
author: mrbullwinkle
10+
author: santiagxf
1111
ms.author: fasantia
1212
recommendations: false
1313
---

0 commit comments

Comments
 (0)