Skip to content

Commit 9738d98

Browse files
committed
review of PM's edits
1 parent 275a8d5 commit 9738d98

File tree

2 files changed

+31
-27
lines changed

2 files changed

+31
-27
lines changed

articles/ai-foundry/concepts/deployments-overview.md

Lines changed: 31 additions & 27 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
2-
title: Deploy models in Azure AI Foundry portal
2+
title: Deployment options for Azure AI Foundry Models
33
titleSuffix: Azure AI Foundry
4-
description: Learn about deploying models in Azure AI Foundry portal.
4+
description: Learn about deployment options for Azure AI Foundry Models.
55
manager: scottpolly
66
ms.service: azure-ai-foundry
77
ms.topic: concept-article
@@ -11,56 +11,62 @@ ms.author: mopeakande
1111
author: msakande
1212
---
1313

14-
# Overview: Deploy AI models in Azure AI Foundry
14+
# Deployment overview for Azure AI Foundry Models
1515

16-
The model catalog in Azure AI Foundry is the hub to discover and use a wide range of models for building generative AI applications. Models need to be deployed to make them available for receiving inference requests. Azure AI Foundry offers a comprehensive suite of deployment options for models, depending on your needs and model requirements.
16+
The model catalog in Azure AI Foundry is the hub to discover and use a wide range of Foundry Models for building generative AI applications. Models need to be deployed to make them available for receiving inference requests. Azure AI Foundry offers a comprehensive suite of deployment options for Foundry Models, depending on your needs and model requirements.
1717

1818
## Deployment options
1919

20-
Azure AI Foundry provides multiple deployment options depending on the type of resources and models that you need to provision. The following 3 deployment options are available:
20+
Azure AI Foundry provides multiple deployment options depending on the type of models and resources you need to provision. The following deployment options are available:
2121

22-
### Standard deployments in Azure AI Foundry resources
22+
- Standard deployment in Azure AI Foundry resources
23+
- Deployment to serverless API endpoint
24+
- Deployment to managed compute
2325

24-
Formerly known Azure AI model inference in Azure AI Services, is **the preferred deployment option** in Azure AI Foundry. It offers the biggest range of options including regional, data zone, or global processing; and standard and provisioned (PTU) options. Flagship models in Azure AI Foundry Models support this deployment option.
26+
### Standard deployment in Azure AI Foundry resources
27+
28+
Azure AI Foundry resources (formerly referred to as Azure AI model inference, in Azure AI Services), is **the preferred deployment option** in Azure AI Foundry. It offers the widest range of options including regional, data zone, or global processing, and it offers standard and [provisioned throughput (PTU)](../../ai-services/openai/concepts/provisioned-throughput.md) options. Flagship models in Azure AI Foundry Models support this deployment option.
2529

2630
This deployment option is available in:
2731

2832
* Azure OpenAI resources<sup>1</sup>
29-
* Azure AI Foundry resources (formerly known Azure AI Services)
30-
* Azure AI Hub when connected to an Azure AI Foundry resource (requires the feature [Deploy models to Azure AI Foundry resources](#configure-azure-ai-foundry-portal-for-deployment-options) on).
33+
* Azure AI Foundry resources
34+
* Azure AI hub, when connected to an Azure AI Foundry resource (requires the [Deploy models to Azure AI Foundry resources](#configure-azure-ai-foundry-portal-for-deployment-options) feature to be turned on).
35+
36+
<sup>1</sup>If you're using Azure OpenAI resources, the model catalog only shows Azure OpenAI in Foundry Models for deployment. You can get the full list of Foundry Models by upgrading to an Azure AI Foundry resource.
3137

32-
<sup>1</sup>If you are using Azure OpenAI resources, the model catalog only shows Azure OpenAI models for deployment. You can get the full list of models by upgrading to an Azure AI Foundry resource.
38+
To get started with standard deployment in Azure AI Foundry resources, see [How-to: Deploy models to Azure AI Foundry Models](../model-inference/how-to/create-model-deployments.md).
3339

34-
To get started, see [How-to: Deploy models to Azure AI Foundry Models](../model-inference/how-to/create-model-deployments.md).
40+
### Serverless API endpoint
3541

36-
### Serverless API Endpoint
42+
This option is available **only in** [Azure AI hub resources](ai-resources.md) and it allows the creation of dedicated endpoints to host the model, accessible via API. Azure AI Foundry Models support serverless API endpoints with pay-as-you-go billing.
3743

38-
This option is available **only in Azure AI Hubs resources** and it allows the creation of dedicated endpoints to host the model, accessible via API with pay-as-you-go billing. It's supported by Azure AI Foundry Models with pay-as-you-go billing. Only regional deployments can be created for Serverless API Endpoints. It requires the feature [Deploy models to Azure AI Foundry resources](#configure-azure-ai-foundry-portal-for-deployment-options) **off**.
44+
Only regional deployments can be created for serverless API endpoints, and to use it, you _must_ **turn off** the "Deploy models to Azure AI Foundry resources" option.
3945

40-
To get started, see [How-to: Deploy models to Serverless API Endpoints](../model-inference/how-to/create-model-deployments.md)
46+
To get started with deployment to a serverless API endpoint, see [Deploy models as serverless API deployments](../how-to/deploy-models-serverless.md).
4147

4248
### Managed Compute
4349

44-
This option is available **only in Azure AI Hubs resources** and it allows the creation of dedicated endpoint to host the model in **dedicated compute**. You need to have compute quota in your subscription to host the model and you are billed per compute up-time.
50+
This option is available **only in** [Azure AI hub resources](ai-resources.md) and it allows the creation of a dedicated endpoint to host the model in a **dedicated compute**. You need to have compute quota in your subscription to host the model and you're billed per compute uptime.
4551

46-
This option is required for the following model collections:
52+
This deployment option is required for model collections such as these:
4753

4854
* Hugging Face
49-
* NVIDIA NIMs
55+
* NVIDIA inference microservices (NIMs)
5056
* Industry models (Saifr, Rockwell, Bayer, Cerence, Sight Machine, Page AI, SDAIA)
5157
* Databricks
5258
* Custom models
5359

54-
To get started, see [How-to: Deploy to Managed compute](../how-to/deploy-models-managed.md).
60+
To get started, see [How to deploy and inference a managed compute deployment](../how-to/deploy-models-managed.md) and [Deploy Azure AI Foundry Models to managed compute with pay-as-you-go billing](../how-to/deploy-models-managed-pay-go.md).
5561

56-
## Features
62+
## Capabilities for the deployment options
5763

58-
We recommend using Standard deployments in Azure AI Foundry resources (formerly known Azure AI model inference in Azure AI Services) whenever possible as it offers the larger set of features. The following table shows details about specific features available on each deployment option:
64+
We recommend using [Standard deployments in Azure AI Foundry resources](#standard-deployment-in-azure-ai-foundry-resources) whenever possible, as it offers the largest set of capabilities among the available deployment options. The following table lists details about specific capabilities available for each deployment option:
5965

60-
| Feature | Azure OpenAI | Azure AI Foundry | Serverless API Endpoint | Managed compute |
66+
| Capability | Azure OpenAI | Azure AI Foundry | Serverless API Endpoint | Managed compute |
6167
|-------------------------------|----------------------|-------------------|----------------|-----------------|
6268
| Which models can be deployed? | [Azure OpenAI models](../../ai-services/openai/concepts/models.md) | [Azure OpenAI models and Foundry Models with pay-as-you-go billing](../../ai-foundry/model-inference/concepts/models.md) | [Foundry Models with pay-as-you-go billing](../how-to/model-catalog-overview.md) | [Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute) |
63-
| Deployment resource | Azure OpenAI resource | Azure AI Foundry resource (formerly known Azure AI Services) | AI project (in AI Hub resource) | AI project (in AI Hub resource) |
69+
| Deployment resource | Azure OpenAI resource | Azure AI Foundry resource | AI project (in AI Hub resource) | AI project (in AI Hub resource) |
6470
| Requires AI Hubs | No | No | Yes | Yes |
6571
| Data processing options | Regional <br /> Data-zone <br /> Global | Regional <br /> Data-zone <br /> Global | Regional | Regional |
6672
| Private networking | Yes | Yes | Yes | Yes |
@@ -75,13 +81,11 @@ We recommend using Standard deployments in Azure AI Foundry resources (formerly
7581

7682
## Configure Azure AI Foundry portal for deployment options
7783

78-
Azure AI Foundry portal may automatically pick up a deployment option based on your environment and configuration. When possible, we default to the most convenient deployment option available to you.
79-
80-
We recommend using Azure AI Foundry resources (formerly known Azure AI Services) for deployment whenever possible. To do that, ensure you have the feature **Deploy models to Azure AI Foundry resources** on.
84+
Azure AI Foundry portal might automatically pick up a deployment option based on your environment and configuration. We recommend using Azure AI Foundry resources for deployment whenever possible. To do that, ensure that the **Deploy models to Azure AI Foundry resources** feature is **turned on**.
8185

82-
:::image type="content" source="../model-inference/media/models/docs-flag-enable-foundry.gif" alt-text="An animation showing how to enable deployment to Azure AI Foundry resources (formerly known Azure AI Services)." lightbox="../model-inference/media/models/docs-flag-enable-foundry.gif":::
86+
:::image type="content" source="../media/concepts/deployments-overview/docs-flag-enable-foundry.png" alt-text="A screenshot showing the steps to enable deployment to Azure AI Foundry resources in the Azure AI Foundry portal." lightbox="../media/concepts/deployments-overview/docs-flag-enable-foundry.png":::
8387

84-
Notice that once enabled, models that support multiple deployment options will default to deploy to Azure AI Foundry resources for deployment. To access other deployment options, either disable the feature or use the Azure CLI or Azure Machine Learning SDK for deployment. You can disable and enable the feature as many times as needed. Existing deployments won't be affected.
88+
Once the **Deploy models to Azure AI Foundry resources** feature is enabled, models that support multiple deployment options default to deploy to Azure AI Foundry resources for deployment. To access other deployment options, either disable the feature or use the Azure CLI or Azure Machine Learning SDK for deployment. You can disable and enable the feature as many times as needed without affecting existing deployments.
8589

8690
## Related content
8791

135 KB
Loading

0 commit comments

Comments
 (0)