You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/deployments-overview.md
+31-27Lines changed: 31 additions & 27 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
---
2
-
title: Deploy models in Azure AI Foundry portal
2
+
title: Deployment options for Azure AI Foundry Models
3
3
titleSuffix: Azure AI Foundry
4
-
description: Learn about deploying models in Azure AI Foundry portal.
4
+
description: Learn about deployment options for Azure AI Foundry Models.
5
5
manager: scottpolly
6
6
ms.service: azure-ai-foundry
7
7
ms.topic: concept-article
@@ -11,56 +11,62 @@ ms.author: mopeakande
11
11
author: msakande
12
12
---
13
13
14
-
# Overview: Deploy AI models in Azure AI Foundry
14
+
# Deployment overview for Azure AI Foundry Models
15
15
16
-
The model catalog in Azure AI Foundry is the hub to discover and use a wide range of models for building generative AI applications. Models need to be deployed to make them available for receiving inference requests. Azure AI Foundry offers a comprehensive suite of deployment options for models, depending on your needs and model requirements.
16
+
The model catalog in Azure AI Foundry is the hub to discover and use a wide range of Foundry Models for building generative AI applications. Models need to be deployed to make them available for receiving inference requests. Azure AI Foundry offers a comprehensive suite of deployment options for Foundry Models, depending on your needs and model requirements.
17
17
18
18
## Deployment options
19
19
20
-
Azure AI Foundry provides multiple deployment options depending on the type of resources and models that you need to provision. The following 3 deployment options are available:
20
+
Azure AI Foundry provides multiple deployment options depending on the type of models and resources you need to provision. The following deployment options are available:
21
21
22
-
### Standard deployments in Azure AI Foundry resources
22
+
- Standard deployment in Azure AI Foundry resources
23
+
- Deployment to serverless API endpoint
24
+
- Deployment to managed compute
23
25
24
-
Formerly known Azure AI model inference in Azure AI Services, is **the preferred deployment option** in Azure AI Foundry. It offers the biggest range of options including regional, data zone, or global processing; and standard and provisioned (PTU) options. Flagship models in Azure AI Foundry Models support this deployment option.
26
+
### Standard deployment in Azure AI Foundry resources
27
+
28
+
Azure AI Foundry resources (formerly referred to as Azure AI model inference, in Azure AI Services), is **the preferred deployment option** in Azure AI Foundry. It offers the widest range of options including regional, data zone, or global processing, and it offers standard and [provisioned throughput (PTU)](../../ai-services/openai/concepts/provisioned-throughput.md) options. Flagship models in Azure AI Foundry Models support this deployment option.
25
29
26
30
This deployment option is available in:
27
31
28
32
* Azure OpenAI resources<sup>1</sup>
29
-
* Azure AI Foundry resources (formerly known Azure AI Services)
30
-
* Azure AI Hub when connected to an Azure AI Foundry resource (requires the feature [Deploy models to Azure AI Foundry resources](#configure-azure-ai-foundry-portal-for-deployment-options) on).
33
+
* Azure AI Foundry resources
34
+
* Azure AI hub, when connected to an Azure AI Foundry resource (requires the [Deploy models to Azure AI Foundry resources](#configure-azure-ai-foundry-portal-for-deployment-options) feature to be turned on).
35
+
36
+
<sup>1</sup>If you're using Azure OpenAI resources, the model catalog only shows Azure OpenAI in Foundry Models for deployment. You can get the full list of Foundry Models by upgrading to an Azure AI Foundry resource.
31
37
32
-
<sup>1</sup>If you are using Azure OpenAI resources, the model catalog only shows Azure OpenAI models for deployment. You can get the full list of models by upgrading to an Azure AI Foundry resource.
38
+
To get started with standard deployment in Azure AI Foundry resources, see [How-to: Deploy models to Azure AI Foundry Models](../model-inference/how-to/create-model-deployments.md).
33
39
34
-
To get started, see [How-to: Deploy models to Azure AI Foundry Models](../model-inference/how-to/create-model-deployments.md).
40
+
### Serverless API endpoint
35
41
36
-
### Serverless API Endpoint
42
+
This option is available **only in**[Azure AI hub resources](ai-resources.md) and it allows the creation of dedicated endpoints to host the model, accessible via API. Azure AI Foundry Models support serverless API endpoints with pay-as-you-go billing.
37
43
38
-
This option is available **only in Azure AI Hubs resources** and it allows the creation of dedicated endpoints to host the model, accessible via API with pay-as-you-go billing. It's supported by Azure AI Foundry Models with pay-as-you-go billing. Only regional deployments can be created for Serverless API Endpoints. It requires the feature [Deploy models to Azure AI Foundry resources](#configure-azure-ai-foundry-portal-for-deployment-options)**off**.
44
+
Only regional deployments can be created for serverless API endpoints, and to use it, you_must_**turn off**the "Deploy models to Azure AI Foundry resources" option.
39
45
40
-
To get started, see [How-to: Deploy models to Serverless API Endpoints](../model-inference/how-to/create-model-deployments.md)
46
+
To get started with deployment to a serverless API endpoint, see [Deploy models as serverless API deployments](../how-to/deploy-models-serverless.md).
41
47
42
48
### Managed Compute
43
49
44
-
This option is available **only inAzure AI Hubs resources** and it allows the creation of dedicated endpoint to host the model in **dedicated compute**. You need to have compute quota in your subscription to host the model and you are billed per compute up-time.
50
+
This option is available **only in**[Azure AI hub resources](ai-resources.md) and it allows the creation of a dedicated endpoint to host the model in a **dedicated compute**. You need to have compute quota in your subscription to host the model and you're billed per compute uptime.
45
51
46
-
This option is required for the following model collections:
52
+
This deployment option is required for model collections such as these:
To get started, see [How-to: Deploy to Managed compute](../how-to/deploy-models-managed.md).
60
+
To get started, see [How to deploy and inference a managed compute deployment](../how-to/deploy-models-managed.md) and [Deploy Azure AI Foundry Models to managed compute with pay-as-you-go billing](../how-to/deploy-models-managed-pay-go.md).
55
61
56
-
## Features
62
+
## Capabilities for the deployment options
57
63
58
-
We recommend using Standard deployments in Azure AI Foundry resources (formerly known Azure AI model inference in Azure AI Services) whenever possible as it offers the larger set of features. The following table shows details about specific features available on each deployment option:
64
+
We recommend using [Standard deployments in Azure AI Foundry resources](#standard-deployment-in-azure-ai-foundry-resources) whenever possible, as it offers the largest set of capabilities among the available deployment options. The following table lists details about specific capabilities available for each deployment option:
59
65
60
-
|Feature | Azure OpenAI | Azure AI Foundry | Serverless API Endpoint | Managed compute |
66
+
|Capability| Azure OpenAI | Azure AI Foundry | Serverless API Endpoint | Managed compute |
| Which models can be deployed? |[Azure OpenAI models](../../ai-services/openai/concepts/models.md)|[Azure OpenAI models and Foundry Models with pay-as-you-go billing](../../ai-foundry/model-inference/concepts/models.md)|[Foundry Models with pay-as-you-go billing](../how-to/model-catalog-overview.md)|[Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute)|
63
-
| Deployment resource | Azure OpenAI resource | Azure AI Foundry resource (formerly known Azure AI Services)| AI project (in AI Hub resource) | AI project (in AI Hub resource) |
69
+
| Deployment resource | Azure OpenAI resource | Azure AI Foundry resource | AI project (in AI Hub resource) | AI project (in AI Hub resource) |
64
70
| Requires AI Hubs | No | No | Yes | Yes |
65
71
| Data processing options | Regional <br /> Data-zone <br /> Global | Regional <br /> Data-zone <br /> Global | Regional | Regional |
66
72
| Private networking | Yes | Yes | Yes | Yes |
@@ -75,13 +81,11 @@ We recommend using Standard deployments in Azure AI Foundry resources (formerly
75
81
76
82
## Configure Azure AI Foundry portal for deployment options
77
83
78
-
Azure AI Foundry portal may automatically pick up a deployment option based on your environment and configuration. When possible, we default to the most convenient deployment option available to you.
79
-
80
-
We recommend using Azure AI Foundry resources (formerly known Azure AI Services) for deployment whenever possible. To do that, ensure you have the feature **Deploy models to Azure AI Foundry resources** on.
84
+
Azure AI Foundry portal might automatically pick up a deployment option based on your environment and configuration. We recommend using Azure AI Foundry resources for deployment whenever possible. To do that, ensure that the **Deploy models to Azure AI Foundry resources** feature is **turned on**.
81
85
82
-
:::image type="content" source="../model-inference/media/models/docs-flag-enable-foundry.gif" alt-text="An animation showing how to enable deployment to Azure AI Foundry resources (formerly known Azure AI Services)." lightbox="../model-inference/media/models/docs-flag-enable-foundry.gif":::
86
+
:::image type="content" source="../media/concepts/deployments-overview/docs-flag-enable-foundry.png" alt-text="A screenshot showing the steps to enable deployment to Azure AI Foundry resources in the Azure AI Foundry portal." lightbox="../media/concepts/deployments-overview/docs-flag-enable-foundry.png":::
83
87
84
-
Notice that once enabled, models that support multiple deployment options will default to deploy to Azure AI Foundry resources for deployment. To access other deployment options, either disable the feature or use the Azure CLI or Azure Machine Learning SDK for deployment. You can disable and enable the feature as many times as needed. Existing deployments won't be affected.
88
+
Once the **Deploy models to Azure AI Foundry resources** feature is enabled, models that support multiple deployment options default to deploy to Azure AI Foundry resources for deployment. To access other deployment options, either disable the feature or use the Azure CLI or Azure Machine Learning SDK for deployment. You can disable and enable the feature as many times as needed without affecting existing deployments.
0 commit comments