You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/deployments-overview.md
+21-21Lines changed: 21 additions & 21 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,37 +4,39 @@ titleSuffix: Azure AI Foundry
4
4
description: Learn about deploying models in Azure AI Foundry portal.
5
5
manager: scottpolly
6
6
ms.service: azure-ai-foundry
7
-
ms.custom:
8
-
- ignite-2023
9
-
- build-2024
10
-
- ignite-2024
11
7
ms.topic: concept-article
12
-
ms.date: 10/21/2024
8
+
ms.date: 03/24/2025
13
9
ms.reviewer: fasantia
14
10
ms.author: mopeakande
15
11
author: msakande
16
12
---
17
13
18
14
# Overview: Deploy AI models in Azure AI Foundry portal
19
15
20
-
The model catalog in Azure AI Foundry portal is the hub to discover and use a wide range of models for building generative AI applications. Models need to be deployed to make them available for receiving inference requests. The process of interacting with a deployed model is called *inferencing*. Azure AI Foundry offer a comprehensive suite of deployment options for those models depending on your needs and model requirements.
16
+
The model catalog in Azure AI Foundry portal is the hub to discover and use a wide range of models for building generative AI applications. Models need to be deployed to make them available for receiving inference requests. Azure AI Foundry offers a comprehensive suite of deployment options for models, depending on your needs and model requirements.
21
17
22
18
## Deploying models
23
19
24
-
Deployment options vary depending on the model type:
20
+
Deployment options vary depending on the model offering:
25
21
26
-
***Azure OpenAI models:** The latest OpenAI models that have enterprise features from Azure.
27
-
***Models as a Service models:** These models don't require compute quota from your subscription. This option allows you to deploy your Model as a Service (MaaS). You use a serverless API deployment and are billed per token in a pay-as-you-go fashion.
28
-
***Open and custom models:** The model catalog offers access to a large variety of models across modalities that are of open access. You can host open models in your own subscription with a managed infrastructure, virtual machines, and the number of instances for capacity management. There's a wide range of models from Azure OpenAI, Hugging Face, and NVIDIA.
22
+
***Azure OpenAI models:** The latest OpenAI models that have enterprise features from Azure with flexible billing options.
23
+
***Models-as-a-Service models:** These models don't require compute quota from your subscriptionand are billed per token in a pay-as-you-go fashion.
24
+
***Open and custom models:** The model catalog offers access to a large variety of models across modalities, including models of open access. You can host open models in your own subscription with a managed infrastructure, virtual machines, and the number of instances for capacity management.
29
25
30
26
Azure AI Foundry offers four different deployment options:
31
27
32
28
|Name | Azure OpenAI service | Azure AI model inference | Serverless API | Managed compute |
| Which models can be deployed? |[Azure OpenAI models](../../ai-services/openai/concepts/models.md)|[Azure OpenAI models and Models as a Service](../../ai-foundry/model-inference/concepts/models.md)|[Models as a Service](../how-to/model-catalog-overview.md#content-safety-for-models-deployed-via-serverless-apis)|[Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute)|
30
+
| Which models can be deployed? |[Azure OpenAI models](../../ai-services/openai/concepts/models.md)|[Azure OpenAI models and Models-as-a-Service](../../ai-foundry/model-inference/concepts/models.md)|[Models-as-a-Service](../how-to/model-catalog-overview.md#content-safety-for-models-deployed-via-serverless-apis)|[Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute)|
35
31
| Deployment resource | Azure OpenAI resource | Azure AI services resource | AI project resource | AI project resource |
36
-
| Best suited when | You are planning to use only OpenAI models | You are planning to take advantage of the flagship models in Azure AI catalog, including OpenAI. | You are planning to use a single model from a specific provider (excluding OpenAI). | If you plan to use open models and you have enough compute quota available in your subscription. |
| Data processing options | Regional <br /> Data-zone <br /> Global | Global | Regional | Regional |
34
+
| Private networking | Yes | Yes | Yes | Yes |
35
+
| Content filtering | Yes | Yes | Yes | No |
36
+
| Custom content filtering | Yes | Yes | No | No |
37
+
| Key-less authentication | Yes | Yes | No | No |
38
+
| Best suited when | You're planning to use only OpenAI models | You're planning to take advantage of the flagship models in Azure AI catalog, including OpenAI. | You're planning to use a single model from a specific provider (excluding OpenAI). | If you plan to use open models and you have enough compute quota available in your subscription. |
| Deployment instructions |[Deploy to Azure OpenAI Service](../how-to/deploy-models-openai.md)|[Deploy to Azure AI model inference](../model-inference/how-to/create-model-deployments.md)|[Deploy to Serverless API](../how-to/deploy-models-serverless.md)|[Deploy to Managed compute](../how-to/deploy-models-managed.md)|
39
41
40
42
<sup>1</sup> A minimal endpoint infrastructure is billed per minute. You aren't billed for the infrastructure that hosts the model in pay-as-you-go. After you delete the endpoint, no further charges accrue.
@@ -46,20 +48,18 @@ Azure AI Foundry offers four different deployment options:
46
48
47
49
### How should I think about deployment options?
48
50
49
-
Azure AI Foundry encourages customers to explore the deployment options and pick the one that best suites their business and technical needs. In general you can use the following thinking process:
51
+
Azure AI Foundry encourages you to explore various deployment options and choose the one that best suites your business and technical needs. In general, Consider using the following approach to select a deployment option:
50
52
51
-
1. Start with the deployment options that have the bigger scopes. This allows you to iterate and prototype faster in your application without having to rebuild your architecture each time you decide to change something. [Azure AI model inference](../../ai-foundry/model-inference/overview.md) is a deployment target that supports all the flagship models in the Azure AI catalog, including latest innovation from Azure OpenAI. To get started, follow [Configure your AI project to use Azure AI model inference](../../ai-foundry/model-inference/how-to/quickstart-ai-project.md).
53
+
* Start with [Azure AI model inference](../../ai-foundry/model-inference/overview.md), which is the option with the largest scope. This option allows you to iterate and prototype faster in your application without having to rebuild your architecture each time you decide to change something. If you're using Azure AI Foundry hubs or projects, enable this option by [turning on the Azure AI model inference feature](../model-inference/how-to/quickstart-ai-project.md#configure-the-project-to-use-azure-ai-model-inference).
52
54
53
-
2. When you are looking to use a specific model:
55
+
* When you're looking to use a specific model:
54
56
55
-
1. When you are interested in Azure OpenAI models, use the Azure OpenAI Service which offers a wide range of capabilities for them and it's designed for them.
57
+
* If you're interested in Azure OpenAI models, use the Azure OpenAI Service. This option is designed for Azure OpenAI models and offers a wide range of capabilities for them.
56
58
57
-
2. When you are interested in a particular model from Models as a Service, and you don't expect to use any other type of model, use [Serverless API endpoints](../how-to/deploy-models-serverless.md). They allow deployment of a single model under a unique set of endpoint URL and keys.
59
+
* If you're interested in a particular model from Models-as-a-Service, and you don't expect to use any other type of model, use [Serverless API endpoints](../how-to/deploy-models-serverless.md). Serverless endpoints allow deployment of a single model under a unique set of endpoint URL and keys.
58
60
59
-
3. When your model is not available in Models as a Service and you have compute quota available in your subscription, use [Managed Compute](../how-to/deploy-models-managed.md) which support deployment of open and custom models. It also allows high level of customization of the deployment inference server, protocols, and detailed configuration.
61
+
* When your model isn't available in Models-as-a-Service and you have compute quota available in your subscription, use [Managed Compute](../how-to/deploy-models-managed.md), which supports deployment of open and custom models. It also allows a high level of customization of the deployment inference server, protocols, and detailed configuration.
60
62
61
-
> [!TIP]
62
-
> Each deployment option may offer different capabilities in terms of networking, security, and additional features like content safety. Review the documentation for each of them to understand their limitations.
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/encryption-keys-portal.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -87,7 +87,7 @@ Customer-managed key encryption is configured via Azure portal in a similar way
87
87
:::image type="content" source="../../machine-learning/media/concept-customer-managed-keys/cmk-service-side-encryption.png" alt-text="Screenshot of the encryption tab with the option for service side encryption selected." lightbox="../../machine-learning/media/concept-customer-managed-keys/cmk-service-side-encryption.png":::
88
88
89
89
Alternatively, use infrastructure-as-code options for automation. Example Bicep templates for Azure AI Foundry are available on the Azure Quickstart repo:
90
-
1.[CMK encryption for hub](https://github.com/Azure/azure-quickstart-templates/tree/master/quickstarts/microsoft.machinelearningservices/aistudio-cmk).
90
+
1.[CMK encryption for hub](https://github.com/Azure/azure-quickstart-templates/tree/master/quickstarts/microsoft.machinelearningservices/aifoundry-cmk).
91
91
1.[Service-side CMK encryption preview for hub](https://github.com/azure/azure-quickstart-templates/tree/master/quickstarts/microsoft.machinelearningservices/aistudio-cmk-service-side-encryption).
0 commit comments