You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/content-filtering.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -20,7 +20,7 @@ author: PatrickFarley
20
20
[Azure AI Foundry](https://ai.azure.com) includes a content filtering system that works alongside core models and image generation models.
21
21
22
22
> [!IMPORTANT]
23
-
> The content filtering system isn't applied to prompts and completions processed by the Whisper model in Azure OpenAI Service. Learn more about the [Whisper model in Azure OpenAI](../../ai-services/openai/concepts/models.md).
23
+
> The content filtering system isn't applied to prompts and completions processed by the Whisper model in Azure OpenAI in Azure AI Foundry Models. Learn more about the [Whisper model in Azure OpenAI](../../ai-services/openai/concepts/models.md).
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/deployments-overview.md
+8-8Lines changed: 8 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -19,15 +19,15 @@ The model catalog in Azure AI Foundry portal is the hub to discover and use a wi
19
19
20
20
Deployment options vary depending on the model offering:
21
21
22
-
***Azure OpenAI models:** The latest OpenAI models that have enterprise features from Azure with flexible billing options.
23
-
***Models-as-a-Service models:** These models don't require compute quota from your subscription and are billed per token in a pay-as-you-go fashion.
22
+
***Azure OpenAI in Azure AI Foundry Models:** The latest OpenAI models that have enterprise features from Azure with flexible billing options.
23
+
***Standard deployment:** These models don't require compute quota from your subscription and are billed per token in a pay-as-you-go fashion.
24
24
***Open and custom models:** The model catalog offers access to a large variety of models across modalities, including models of open access. You can host open models in your own subscription with a managed infrastructure, virtual machines, and the number of instances for capacity management.
25
25
26
26
Azure AI Foundry offers four different deployment options:
27
27
28
-
|Name | Azure OpenAI service | Azure AI model inference |Serverless API| Managed compute |
28
+
|Name | Azure OpenAI | Azure AI model inference |Standard deployment| Managed compute |
| Which models can be deployed? |[Azure OpenAI models](../../ai-services/openai/concepts/models.md)|[Azure OpenAI models and Models-as-a-Service](../../ai-foundry/model-inference/concepts/models.md)|[Models-as-a-Service](../how-to/model-catalog-overview.md#content-safety-for-models-deployed-via-serverless-apis)|[Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute)|
30
+
| Which models can be deployed? |[Azure OpenAI models](../../ai-services/openai/concepts/models.md)|[Azure OpenAI models and Standard deployment](../../ai-foundry/model-inference/concepts/models.md)|[Standard deployment](../how-to/model-catalog-overview.md#content-safety-for-models-deployed-via-serverless-apis)|[Open and custom models](../how-to/model-catalog-overview.md#availability-of-models-for-deployment-as-managed-compute)|
31
31
| Deployment resource | Azure OpenAI resource | Azure AI services resource | AI project resource | AI project resource |
32
32
| Requires Hubs/Projects | No | No | Yes | Yes |
33
33
| Data processing options | Regional <br /> Data-zone <br /> Global | Global | Regional | Regional |
@@ -37,7 +37,7 @@ Azure AI Foundry offers four different deployment options:
37
37
| Key-less authentication | Yes | Yes | No | No |
38
38
| Best suited when | You're planning to use only OpenAI models | You're planning to take advantage of the flagship models in Azure AI catalog, including OpenAI. | You're planning to use a single model from a specific provider (excluding OpenAI). | If you plan to use open models and you have enough compute quota available in your subscription. |
| Deployment instructions |[Deploy to Azure OpenAI Service](../how-to/deploy-models-openai.md)|[Deploy to Azure AI model inference](../model-inference/how-to/create-model-deployments.md)|[Deploy to Serverless API](../how-to/deploy-models-serverless.md)|[Deploy to Managed compute](../how-to/deploy-models-managed.md)|
40
+
| Deployment instructions |[Deploy to Azure OpenAI](../how-to/deploy-models-openai.md)|[Deploy to Azure AI model inference](../model-inference/how-to/create-model-deployments.md)|[Deploy to Standard deployment](../how-to/deploy-models-serverless.md)|[Deploy to Managed compute](../how-to/deploy-models-managed.md)|
41
41
42
42
<sup>1</sup> A minimal endpoint infrastructure is billed per minute. You aren't billed for the infrastructure that hosts the model in pay-as-you-go. After you delete the endpoint, no further charges accrue.
43
43
@@ -54,11 +54,11 @@ Azure AI Foundry encourages you to explore various deployment options and choose
54
54
55
55
* When you're looking to use a specific model:
56
56
57
-
* If you're interested in Azure OpenAI models, use the Azure OpenAI Service. This option is designed for Azure OpenAI models and offers a wide range of capabilities for them.
57
+
* If you're interested in Azure OpenAI models, use Azure OpenAI in Foundry Models. This option is designed for Azure OpenAI models and offers a wide range of capabilities for them.
58
58
59
-
* If you're interested in a particular model from Models-as-a-Service, and you don't expect to use any other type of model, use [Serverless API endpoints](../how-to/deploy-models-serverless.md). Serverless endpoints allow deployment of a single model under a unique set of endpoint URL and keys.
59
+
* If you're interested in a particular model from serverless pay per token offer, and you don't expect to use any other type of model, use [Standard deployment](../how-to/deploy-models-serverless.md). Standard deployments allow deployment of a single model under a unique set of endpoint URL and keys.
60
60
61
-
* When your model isn't available in Models-as-a-Service and you have compute quota available in your subscription, use [Managed Compute](../how-to/deploy-models-managed.md), which supports deployment of open and custom models. It also allows a high level of customization of the deployment inference server, protocols, and detailed configuration.
61
+
* When your model isn't available in standard deployment and you have compute quota available in your subscription, use [Managed Compute](../how-to/deploy-models-managed.md), which supports deployment of open and custom models. It also allows a high level of customization of the deployment inference server, protocols, and detailed configuration.
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/fine-tuning-overview.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -90,7 +90,7 @@ Fine-tuning is available in specific Azure regions for some models that are depl
90
90
91
91
For more information on fine-tuning using a managed compute (preview), see [Fine-tune models using managed compute (preview)](../how-to/fine-tune-managed-compute.md).
92
92
93
-
For details about Azure OpenAI models that are available for fine-tuning, see the [Azure OpenAI Service models documentation](../../ai-services/openai/concepts/models.md#fine-tuning-models) or the [Azure OpenAI models table](#fine-tuning-azure-openai-models) later in this guide.
93
+
For details about Azure OpenAI in Azure AI Foundry Models that are available for fine-tuning, see the [Azure OpenAI in Foundry Models documentation](../../ai-services/openai/concepts/models.md#fine-tuning-models) or the [Azure OpenAI models table](#fine-tuning-azure-openai-models) later in this guide.
94
94
95
95
For the Azure OpenAI Service models that you can fine tune, supported regions for fine-tuning include North Central US, Sweden Central, and more.
Copy file name to clipboardExpand all lines: articles/ai-foundry/concepts/model-lifecycle-retirement.md
+6-9Lines changed: 6 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,11 +1,11 @@
1
1
---
2
-
title: Deprecation for models in Azure AI model catalog
2
+
title: Deprecation for Foundry Models
3
3
titleSuffix: Azure AI Foundry
4
-
description: Learn about the lifecycle stages, deprecation, and retirement for models in the Azure AI model catalog.
4
+
description: Learn about the lifecycle stages, deprecation, and retirement for Azure AI Foundry Models.
5
5
manager: scottpolly
6
6
ms.service: azure-ai-foundry
7
7
ms.topic: concept-article
8
-
ms.date: 03/20/2025
8
+
ms.date: 05/05/2025
9
9
ms.author: mopeakande
10
10
author: msakande
11
11
ms.reviewer: kritifaujdar
@@ -14,15 +14,12 @@ reviewer: fkriti
14
14
#Customer intent: As a data scientist, I want to learn about the lifecycle of models that are available in the model catalog.
15
15
---
16
16
17
-
# Model deprecation and retirement in Azure AI model catalog
17
+
# Model deprecation and retirement for Azure AI Foundry Models
18
18
19
-
Models in the model catalog are continually refreshed with newer and more capable models. As part of this process, model providers might deprecate and retire their older models, and you might need to update your applications to use a newer model. This document communicates information about the model lifecycle and deprecation timelines and explains how you're informed of model lifecycle stages.
19
+
Azure AI Foundry Models in the model catalog are continually refreshed with newer and more capable models. As part of this process, model providers might deprecate and retire their older models, and you might need to update your applications to use a newer model. This document communicates information about the model lifecycle and deprecation timelines and explains how you're informed of model lifecycle stages.
20
20
21
21
> [!IMPORTANT]
22
-
> This article describes deprecation and retirement only for models that can be deployed to [serverless APIs](../how-to/model-catalog-overview.md#model-deployment-managed-compute-and-serverless-apis) or the [Azure AI model Inference](../../ai-foundry/model-inference/overview.md). This article doesn't cover deprecation information for models that can be deployed only to [managed computes](../how-to/model-catalog-overview.md#managed-compute).
23
-
24
-
> [!NOTE]
25
-
> Azure OpenAI models in the model catalog are provided through Azure OpenAI Service. For information about Azure OpenAI model deprecation and retirement, see the [Azure OpenAI service product documentation](/azure/ai-services/openai/concepts/model-retirements).
22
+
> This article describes deprecation and retirement only for Azure Direct models and Azure Ecosystem models models in Foundry Models. For information about deprecation and retirement for Azure OpenAI in Foundry Models, see the Azure OpenAI models lifecycle documentation.
The Azure AI model catalog offers a large selection of models from a wide range of providers. You have various options for deploying models from the model catalog. This article lists featured models in the model catalog that can be deployed and hosted on Microsoft's servers via serverless APIs. For some of these models, you can also host them on your infrastructure for deployment via managed compute. See [Available models for supported deployment options](../how-to/model-catalog-overview.md#available-models-for-supported-deployment-options) to find models in the catalog that are available for deployment via managed compute or serverless API.
18
+
The Azure AI model catalog offers a large selection of Azure AI Foundry Models from a wide range of providers. You have various options for deploying models from the model catalog. This article lists featured models in the model catalog that can be deployed and hosted on Microsoft's servers via standard deployment. For some of these models, you can also host them on your infrastructure for deployment via managed compute. See [Available models for supported deployment options](../how-to/model-catalog-overview.md#available-models-for-supported-deployment-options) to find models in the catalog that are available for deployment via managed compute or standard deployment.
Copy file name to clipboardExpand all lines: articles/ai-foundry/faq.yml
+2-2Lines changed: 2 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -57,11 +57,11 @@ sections:
57
57
- question: |
58
58
Can all models be secured with content filtering?
59
59
answer: |
60
-
Azure AI Content Safety can be used for AI-generated content from Azure OpenAI Service, open-source, and frontier models. For more information, see [How Azure AI Content Safety helps protect users from the classroom to the chatroom](https://aka.ms/contentsafety_GA_blog).
60
+
Azure AI Content Safety can be used for AI-generated content from Azure OpenAI in Azure AI Foundry Models, open-source, and frontier models. For more information, see [How Azure AI Content Safety helps protect users from the classroom to the chatroom](https://aka.ms/contentsafety_GA_blog).
61
61
- question: |
62
62
Do you use my company data to train any of the models?
63
63
answer: |
64
-
Azure OpenAI Service doesn't use customer data to retrain models. For more information, see the [Azure OpenAI data, privacy, and security guide](/legal/cognitive-services/openai/data-privacy?context=/azure/ai-services/openai/context/context).
64
+
Azure OpenAI doesn't use customer data to retrain models. For more information, see the [Azure OpenAI data, privacy, and security guide](/legal/cognitive-services/openai/data-privacy?context=/azure/ai-services/openai/context/context).
0 commit comments