Skip to content

Commit 5265172

Browse files
committed
Merge branch 'main' into release-2025-openai-march-latest
2 parents e45c54d + 4a7a2f8 commit 5265172

File tree

8 files changed

+92
-79
lines changed

8 files changed

+92
-79
lines changed

articles/ai-foundry/concepts/deployments-overview.md

Lines changed: 17 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -4,10 +4,6 @@ titleSuffix: Azure AI Foundry
44
description: Learn about deploying models in Azure AI Foundry portal.
55
manager: scottpolly
66
ms.service: azure-ai-foundry
7-
ms.custom:
8-
- ignite-2023
9-
- build-2024
10-
- ignite-2024
117
ms.topic: concept-article
128
ms.date: 10/21/2024
139
ms.reviewer: fasantia
@@ -17,22 +13,28 @@ author: msakande
1713

1814
# Overview: Deploy AI models in Azure AI Foundry portal
1915

20-
The model catalog in Azure AI Foundry portal is the hub to discover and use a wide range of models for building generative AI applications. Models need to be deployed to make them available for receiving inference requests. The process of interacting with a deployed model is called *inferencing*. Azure AI Foundry offer a comprehensive suite of deployment options for those models depending on your needs and model requirements.
16+
The model catalog in Azure AI Foundry portal is the hub to discover and use a wide range of models for building generative AI applications. Models need to be deployed to make them available for receiving inference requests. Azure AI Foundry offers a comprehensive suite of deployment options for those models depending on your needs and model requirements.
2117

2218
## Deploying models
2319

24-
Deployment options vary depending on the model type:
20+
Deployment options vary depending on the model offering:
2521

26-
* **Azure OpenAI models:** The latest OpenAI models that have enterprise features from Azure.
27-
* **Models as a Service models:** These models don't require compute quota from your subscription. This option allows you to deploy your Model as a Service (MaaS). You use a serverless API deployment and are billed per token in a pay-as-you-go fashion.
28-
* **Open and custom models:** The model catalog offers access to a large variety of models across modalities that are of open access. You can host open models in your own subscription with a managed infrastructure, virtual machines, and the number of instances for capacity management. There's a wide range of models from Azure OpenAI, Hugging Face, and NVIDIA.
22+
* **Azure OpenAI models:** The latest OpenAI models that have enterprise features from Azure with flexible billing options.
23+
* **Models-as-a-Service models:** These models don't require compute quota from your subscription and are billed per token in a pay-as-you-go fashion.
24+
* **Open and custom models:** The model catalog offers access to a large variety of models across modalities, including models of open access. You can host open models in your own subscription with a managed infrastructure, virtual machines, and the number of instances for capacity management.
2925

3026
Azure AI Foundry offers four different deployment options:
3127

3228
|Name | Azure OpenAI service | Azure AI model inference | Serverless API | Managed compute |
3329
|-------------------------------|----------------------|-------------------|----------------|-----------------|
34-
| Which models can be deployed? | [Azure OpenAI models](../../ai-services/openai/concepts/models.md) | [Azure OpenAI models and Models as a Service](../../ai-foundry/model-inference/concepts/models.md) | [Models as a Service](../how-to/model-catalog-overview.md#content-safety-for-models-deployed-via-serverless-apis) | [Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute) |
30+
| Which models can be deployed? | [Azure OpenAI models](../../ai-services/openai/concepts/models.md) | [Azure OpenAI models and Models-as-a-Service](../../ai-foundry/model-inference/concepts/models.md) | [Models-as-a-Service](../how-to/model-catalog-overview.md#content-safety-for-models-deployed-via-serverless-apis) | [Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute) |
3531
| Deployment resource | Azure OpenAI resource | Azure AI services resource | AI project resource | AI project resource |
32+
| Requires Hubs/Projects | No | No | Yes | Yes |
33+
| Data processing options | Regional <br /> Data-zone <br /> Global | Global | Regional | Regional |
34+
| Private networking | Yes | Yes | Yes | Yes |
35+
| Content filtering | Yes | Yes | Yes | No |
36+
| Custom content filtering | Yes | Yes | No | No |
37+
| Key-less authentication | Yes | Yes | No | No |
3638
| Best suited when | You are planning to use only OpenAI models | You are planning to take advantage of the flagship models in Azure AI catalog, including OpenAI. | You are planning to use a single model from a specific provider (excluding OpenAI). | If you plan to use open models and you have enough compute quota available in your subscription. |
3739
| Billing bases | Token usage & PTU | Token usage | Token usage<sup>1</sup> | Compute core hours<sup>2</sup> |
3840
| Deployment instructions | [Deploy to Azure OpenAI Service](../how-to/deploy-models-openai.md) | [Deploy to Azure AI model inference](../model-inference/how-to/create-model-deployments.md) | [Deploy to Serverless API](../how-to/deploy-models-serverless.md) | [Deploy to Managed compute](../how-to/deploy-models-managed.md) |
@@ -48,18 +50,16 @@ Azure AI Foundry offers four different deployment options:
4850

4951
Azure AI Foundry encourages customers to explore the deployment options and pick the one that best suites their business and technical needs. In general you can use the following thinking process:
5052

51-
1. Start with the deployment options that have the bigger scopes. This allows you to iterate and prototype faster in your application without having to rebuild your architecture each time you decide to change something. [Azure AI model inference](../../ai-foundry/model-inference/overview.md) is a deployment target that supports all the flagship models in the Azure AI catalog, including latest innovation from Azure OpenAI. To get started, follow [Configure your AI project to use Azure AI model inference](../../ai-foundry/model-inference/how-to/quickstart-ai-project.md).
53+
* Start with [Azure AI model inference](../../ai-foundry/model-inference/overview.md) which is the option with the bigger scope. This allows you to iterate and prototype faster in your application without having to rebuild your architecture each time you decide to change something. If you are using Azure AI Foundry Hubs or Projects, enable it by [turning on Azure AI model inference](../../ai-foundry/model-inference/how-to/quickstart-ai-project.md).
5254

53-
2. When you are looking to use a specific model:
55+
* When you are looking to use a specific model:
5456

55-
1. When you are interested in Azure OpenAI models, use the Azure OpenAI Service which offers a wide range of capabilities for them and it's designed for them.
57+
* When you are interested in Azure OpenAI models, use the Azure OpenAI Service which offers a wide range of capabilities for them and it's designed for them.
5658

57-
2. When you are interested in a particular model from Models as a Service, and you don't expect to use any other type of model, use [Serverless API endpoints](../how-to/deploy-models-serverless.md). They allow deployment of a single model under a unique set of endpoint URL and keys.
59+
* When you are interested in a particular model from Models-as-a-Service, and you don't expect to use any other type of model, use [Serverless API endpoints](../how-to/deploy-models-serverless.md). They allow deployment of a single model under a unique set of endpoint URL and keys.
5860

59-
3. When your model is not available in Models as a Service and you have compute quota available in your subscription, use [Managed Compute](../how-to/deploy-models-managed.md) which support deployment of open and custom models. It also allows high level of customization of the deployment inference server, protocols, and detailed configuration.
61+
* When your model is not available in Models-as-a-Service and you have compute quota available in your subscription, use [Managed Compute](../how-to/deploy-models-managed.md) which support deployment of open and custom models. It also allows high level of customization of the deployment inference server, protocols, and detailed configuration.
6062

61-
> [!TIP]
62-
> Each deployment option may offer different capabilities in terms of networking, security, and additional features like content safety. Review the documentation for each of them to understand their limitations.
6363

6464
## Related content
6565

articles/ai-foundry/how-to/fine-tune-serverless.md

Lines changed: 8 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -145,7 +145,7 @@ After you select and upload the training dataset, select **Next** to continue.
145145
The next step provides options to configure the model to use validation data in the training process. If you don't want to use validation data, you can choose **Next** to continue to the advanced options for the model. Otherwise, if you have a validation dataset, you can either choose existing prepared validation data or upload new prepared validation data to use when customizing your model.
146146
The **Validation data** pane displays any existing, previously uploaded training and validation datasets and provides options by which you can upload new validation data.
147147

148-
### Automatic Split of Training Data
148+
### Split training data
149149
You can automatically divide your training data to generate a validation dataset.
150150
After you select Automatic split of training data, select **Next** to continue.
151151

@@ -213,7 +213,7 @@ Here are some of the tasks you can do on the **Models** tab:
213213
Several enterprise scenarios are supported for MaaS finetuning. The table below outlines the supported configurations for user storage networking and authentication to ensure smooth operation within enterprise scenarios:
214214

215215
>[!Note]
216-
>- Data connections auth can be changed via AI Studio by clicking on the datastore connection which your dataset is stored in, and navigating to the **Access details** > **Authentication Method** setting.
216+
>- Data connections auth can be changed via AI Foundry by clicking on the datastore connection which your dataset is stored in, and navigating to the **Access details** > **Authentication Method** setting.
217217
>- Storage auth can be changed in Azure Storage > **Settings** > **Configurations** page > **Allow storage account key access**.
218218
>- Storage networking can be changed in Azure Storage > **Networking** page.
219219
@@ -228,7 +228,7 @@ Several enterprise scenarios are supported for MaaS finetuning. The table below
228228
| Public Network Access = Disabled | Account key disabled | Entra-Based Auth (Credentialless) | Yes, UX and SDK. <br><br> *Note:* for UX data upload and submission to work, the workspace _needs to be accessed from within the Vnet_ that has appropriate access to the storage |
229229

230230

231-
The scenarios above should work in a Managed Vnet workspace as well. See setup of Managed Vnet AI Studio hub here: [How to configure a managed network for Azure AI Foundry hubs](./configure-managed-network.md)
231+
The scenarios above should work in a Managed Vnet workspace as well. See setup of Managed Vnet AI Foundry hub here: [How to configure a managed network for Azure AI Foundry hubs](./configure-managed-network.md)
232232

233233
Customer-Managed Keys (CMKs) is **not** a supported enterprise scenario with MaaS finetuning.
234234

@@ -240,12 +240,10 @@ Issues finetuning with unique network setups on the workspace and storage usuall
240240
When the fine-tuning job succeeds, you can deploy the custom model from the **Fine-tune** tab. You must deploy your custom model to make it available for use with completion calls.
241241

242242
> [!IMPORTANT]
243-
> After you deploy a customized model, if at any time the deployment remains inactive for greater than fifteen (15) days, the deployment is deleted. The deployment of a
244-
> customized model is inactive if the model was deployed more than fifteen (15) days ago and no completions or chat completions calls were made to it during a continuous 15
245-
> day period.
246-
> The deletion of an inactive deployment doesn't delete or affect the underlying customized model, and the customized model can be redeployed at any time. As described in
247-
> Azure AI Foundry pricing, each customized (fine-tuned) model that's deployed incurs an hourly hosting cost regardless of whether completions or chat completions calls are
248-
> being made to the model. To learn more about planning and managing costs with Azure AI Foundry, refer to the guidance in [Plan to manage costs for Azure AI Foundry Service](./costs-plan-manage.md).
243+
> After you deploy a customized model and finishing with the endpoint, please remember to clean up any inactive endpoints. The deletion of an inactive deployment doesn't
244+
> delete or affect the underlying customized model, and the customized model can be redeployed at any time. As described in Azure AI Foundry pricing, each customized (fine-
245+
> tuned) model that's deployed incurs an hourly hosting cost regardless of whether completions or chat completions calls are being made to the model. To learn more about
246+
> planning and managing costs with Azure AI Foundry, refer to the guidance in [Plan to manage costs for Azure AI Foundry Service](./costs-plan-manage.md).
249247
250248
> [!NOTE]
251249
> Only one deployment is permitted for a custom model. An error message is displayed if you select an already-deployed custom model.
@@ -297,7 +295,7 @@ For more information on how to track costs, see [Monitor costs for models offere
297295

298296
:::image type="content" source="../media/deploy-monitor/serverless/costs-model-as-service-cost-details.png" alt-text="A screenshot showing different resources corresponding to different model offers and their associated meters." lightbox="../media/deploy-monitor/serverless/costs-model-as-service-cost-details.png":::
299297

300-
## Sample Notebook
298+
## Sample notebook
301299

302300
You can use this [sample notebook](https://github.com/Azure/azureml-examples/blob/main/sdk/python/jobs/finetuning/standalone/model-as-a-service/chat-completion/chat_completion_with_model_as_service.ipynb) to create a standalone fine-tuning job to enhance a model's ability to summarize dialogues between two people using the Samsum dataset. The training data utilized is the ultrachat_200k dataset, which is divided into four splits suitable for supervised fine-tuning (sft) and generation ranking (gen). The notebook employs the available Azure AI models for the chat-completion task (If you would like to use a different model than what's used in the notebook, you can replace the model name). The notebook includes setting up prerequisites, selecting a model to fine-tune, creating training and validation datasets, configuring and submitting the fine-tuning job, and finally, creating a serverless deployment using the fine-tuned model for sample inference.
303301

articles/ai-services/document-intelligence/authentication/encrypt-data-at-rest.md

Lines changed: 10 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -6,27 +6,30 @@ author: erindormier
66
manager: venkyv
77
ms.service: azure-ai-document-intelligence
88
ms.topic: conceptual
9-
ms.date: 11/19/2024
9+
ms.date: 03/19/2025
1010
monikerRange: '<=doc-intel-4.0.0'
1111
---
1212

1313

14-
# Document Intelligence encryption of data at rest
14+
# Encrypt data at rest
1515

1616
[!INCLUDE [applies to v4.0, v3.1, v3.0, and v2.1](../includes/applies-to-v40-v31-v30-v21.md)]
1717

1818
> [!IMPORTANT]
1919
>
20-
> * Earlier versions of customer managed keys only encrypted your models.
21-
> *Starting with the ```07/31/2023``` release, all new resources use customer managed keys to encrypt both the models and document results.
22-
> To upgrade an existing service to encrypt both the models and the data, simply disable and reenable the customer managed key.
20+
> * Earlier versions of customer managed keys (`CMK`) only encrypted your models.
21+
> * Beginning with the ```07/31/2023``` release, all new resources utilize customer-managed keys to encrypt both models and document results.
22+
> * [Delete analyze response](/rest/api/aiservices/document-models/delete-analyze-result?view=rest-aiservices-v4.0%20(2024-11-30)&preserve-view=true&tabs=HTTP). the `analyze response` is stored for 24 hours from when the operation completes for retrieval. For scenarios where you want to delete the response sooner, use the delete analyze response API to delete the response.
23+
> * To upgrade an existing service to encrypt both the models and the data, disable and reenable the customer managed key.
2324
24-
Azure AI Document Intelligence automatically encrypts your data when persisting it to the cloud. Document Intelligence encryption protects your data to help you to meet your organizational security and compliance commitments.
25+
Azure AI Document Intelligence automatically encrypts your data when persisting it to the cloud. Document Intelligence encryption protects your data to help you to meet your organizational security and compliance commitments.
2526

2627
[!INCLUDE [cognitive-services-about-encryption](../../../ai-services/includes/cognitive-services-about-encryption.md)]
2728

2829
> [!IMPORTANT]
29-
> Customer-managed keys are only available resources created after 11 May, 2020. To use CMK with Document Intelligence, you will need to create a new Document Intelligence resource. Once the resource is created, you can use Azure Key Vault to set up your managed identity.
30+
> * Customer-managed keys are only available resources created after May 11, 2020. To use customer-managed keys with Document Intelligence, you need to create a new Document Intelligence resource. Once the resource is created, you can use Azure Key Vault to set up your managed identity.
31+
> * The scope for data encrypted with customer-managed keys includes the `analysis response` stored for 24 hours, allowing the operation results to be retrieved during that 24-hour time period.
32+
3033

3134
[!INCLUDE [cognitive-services-cmk](../../../ai-services/includes/configure-customer-managed-keys.md)]
3235

articles/ai-services/document-intelligence/tutorial/azure-function.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ manager: nitinme
77

88
ms.service: azure-ai-document-intelligence
99
ms.topic: tutorial
10-
ms.date: 11/19/2024
10+
ms.date: 03/19/2025
1111
ms.author: lajanuar
1212
ms.custom: VS Code-azure-extension-update-completed, devx-track-python
1313
---
-67 KB
Loading

0 commit comments

Comments
 (0)