Skip to content

Commit f3cd750

Browse files
committed
final updates
1 parent 84cc704 commit f3cd750

File tree

18 files changed

+61
-61
lines changed

18 files changed

+61
-61
lines changed

articles/ai-foundry/how-to/develop/langchain.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -126,7 +126,7 @@ model = AzureAIChatCompletionsModel(
126126
)
127127
```
128128
129-
If your endpoint is serving one model, like with the standard deployment, you don't have to indicate `model` parameter:
129+
If your endpoint is serving one model, like with the serverless API deployment, you don't have to indicate `model` parameter:
130130

131131
```python
132132
import os

articles/ai-foundry/how-to/develop/llama-index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ To run this tutorial, you need:
3333
* An Azure AI project as explained at [Create a project in Azure AI Foundry portal](../create-projects.md).
3434
* A model supporting the [Model Inference API](https://aka.ms/azureai/modelinference) deployed. In this example, we use a `Mistral-Large` deployment, but use any model of your preference. For using embeddings capabilities in LlamaIndex, you need an embedding model like `cohere-embed-v3-multilingual`.
3535

36-
* You can follow the instructions at [Deploy models as standard deployments](../deploy-models-serverless.md).
36+
* You can follow the instructions at [Deploy models as serverless API deployments](../deploy-models-serverless.md).
3737

3838
* Python 3.8 or later installed, including pip.
3939
* LlamaIndex installed. You can do it with:

articles/ai-foundry/how-to/develop/semantic-kernel.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ In this article, you learn how to use [Semantic Kernel](/semantic-kernel/overvie
2121
- An Azure AI project as explained at [Create a project in Azure AI Foundry portal](../create-projects.md).
2222
- A model supporting the [Azure AI Model Inference API](../../../ai-foundry/model-inference/reference/reference-model-inference-api.md?tabs=python) deployed. In this example, we use a `Mistral-Large` deployment, but use any model of your preference. For using embeddings capabilities in LlamaIndex, you need an embedding model like `cohere-embed-v3-multilingual`.
2323

24-
- You can follow the instructions at [Deploy models as standard deployments](../deploy-models-serverless.md).
24+
- You can follow the instructions at [Deploy models as serverless API deployments](../deploy-models-serverless.md).
2525

2626
- Python **3.10** or later installed, including pip.
2727
- Semantic Kernel installed. You can do it with:

articles/ai-foundry/how-to/fine-tune-serverless.md

Lines changed: 18 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
2-
title: Fine-tune models using standard deployments in Azure AI Foundry portal
2+
title: Fine-tune models using serverless API deployments in Azure AI Foundry portal
33
titleSuffix: Azure AI Foundry
4-
description: Learn how to fine-tune models deployed via standard deployments in Azure AI Foundry.
4+
description: Learn how to fine-tune models deployed via serverless API deployments in Azure AI Foundry.
55
manager: scottpolly
66
ms.service: azure-ai-foundry
77
ms.topic: how-to
@@ -14,19 +14,19 @@ ms.custom: references_regions
1414
zone_pivot_groups: azure-ai-model-fine-tune
1515
---
1616

17-
# Fine-tune models using standard deployments in Azure AI Foundry
17+
# Fine-tune models using serverless API deployments in Azure AI Foundry
1818

1919
[!INCLUDE [Feature preview](~/reusable-content/ce-skilling/azure/includes/ai-studio/includes/feature-preview.md)]
2020

2121
Azure AI Foundry enables you to customize large language models to your specific datasets through a process called fine-tuning. This process offers significant benefits by allowing for customization and optimization tailored to specific tasks and applications. The advantages include improved performance, cost efficiency, reduced latency, and tailored outputs.
2222

23-
**Cost Efficiency**: Azure AI Foundry's fine-tuning can be more cost-effective, especially for large-scale deployments, thanks to Standard pricing.
23+
**Cost Efficiency**: Azure AI Foundry's fine-tuning can be more cost-effective, especially for large-scale deployments, thanks to pay-as-you-go pricing.
2424

25-
**Model Variety**: Azure AI Foundry's standard deployment fine-tuning offers support for both proprietary and open-source models, providing users with the flexibility to select the models that best suit their needs without being restricted to a single type.
25+
**Model Variety**: Azure AI Foundry's serverless API deployment fine-tuning offers support for both proprietary and open-source models, providing users with the flexibility to select the models that best suit their needs without being restricted to a single type.
2626

2727
**Customization and Control**: Azure AI Foundry provides greater customization and control over the fine-tuning process, enabling users to tailor models more precisely to their specific requirements.
2828

29-
In this article, you will discover how to fine-tune models that are deployed using standard deployments in [Azure AI Foundry](https://ai.azure.com/?cid=learnDocs).
29+
In this article, you will discover how to fine-tune models that are deployed using serverless API deployments in [Azure AI Foundry](https://ai.azure.com/?cid=learnDocs).
3030

3131

3232
## Prerequisites
@@ -51,7 +51,7 @@ Verify the subscription is registered to the Microsoft.Network resource provider
5151

5252
## Find models with fine-tuning support
5353

54-
The AI Foundry model catalog offers fine-tuning support for multiple types of models, including chat completions and text generation. For a list of models that support fine-tuning and the Azure regions of support for fine-tuning, see [region availability for models as standard deployment.](deploy-models-serverless-availability.md) Fine-tuning tasks are available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available. If the offer is available in the relevant region, the user then must have a project resource in the Azure region where the model is available for deployment or fine-tuning, as applicable.
54+
The AI Foundry model catalog offers fine-tuning support for multiple types of models, including chat completions and text generation. For a list of models that support fine-tuning and the Azure regions of support for fine-tuning, see [region availability for models as serverless API deployment.](deploy-models-serverless-availability.md) Fine-tuning tasks are available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available. If the offer is available in the relevant region, the user then must have a project resource in the Azure region where the model is available for deployment or fine-tuning, as applicable.
5555

5656

5757
You can also go to the Azure AI Foundry portal to view all models that contain fine-tuning support:
@@ -121,8 +121,8 @@ Azure AI Foundry portal provides the Create custom model wizard, so you can inte
121121
### Select the base model
122122

123123
1. Choose the model you want to fine-tune from the Azure AI Foundry [model catalog](https://ai.azure.com/explore/models).
124-
2. On the model's **Details page**, select **fine-tune**. Some foundation models support both **standard deployment** and **Managed compute**, while others support one or the other.
125-
3. If you're presented the options for **standard deployment** and [**Managed compute**](./fine-tune-managed-compute.md), select **standard deployment** for fine-tuning. This action opens up a wizard that shows information about **Standard** fine-tuning for your model.
124+
2. On the model's **Details page**, select **fine-tune**. Some foundation models support both **serverless API deployment** and **Managed compute**, while others support one or the other.
125+
3. If you're presented the options for **serverless API deployment** and [**Managed compute**](./fine-tune-managed-compute.md), select **serverless API deployment** for fine-tuning. This action opens up a wizard that shows information about **Serverless API** fine-tuning for your model.
126126

127127
### Choose your training data
128128
The next step is to either choose existing prepared training data or upload new prepared training data to use when customizing your model. The **Training data** pane displays any existing, previously uploaded datasets and also provides options to upload new training data.
@@ -210,7 +210,7 @@ Here are some of the tasks you can do on the **Models** tab:
210210

211211
### Supported enterprise scenarios for fine-tuning
212212

213-
Several enterprise scenarios are supported for standard deployment fine-tuning. The table below outlines the supported configurations for user storage networking and authentication to ensure smooth operation within enterprise scenarios:
213+
Several enterprise scenarios are supported for serverless API deployment fine-tuning. The table below outlines the supported configurations for user storage networking and authentication to ensure smooth operation within enterprise scenarios:
214214

215215
>[!Note]
216216
>- Data connections auth can be changed via AI Foundry by clicking on the datastore connection which your dataset is stored in, and navigating to the **Access details** > **Authentication Method** setting.
@@ -230,7 +230,7 @@ Several enterprise scenarios are supported for standard deployment fine-tuning.
230230

231231
The scenarios above should work in a Managed Vnet workspace as well. See setup of Managed Vnet AI Foundry hub here: [How to configure a managed network for Azure AI Foundry hubs](./configure-managed-network.md)
232232

233-
Customer-Managed Keys (CMKs) is **not** a supported enterprise scenario with standard deployment fine-tuning.
233+
Customer-Managed Keys (CMKs) is **not** a supported enterprise scenario with serverless API deployment fine-tuning.
234234

235235
Issues fine-tuning with unique network setups on the workspace and storage usually points to a networking setup issue.
236236

@@ -331,7 +331,7 @@ workspace.id
331331

332332
### Find models with fine-tuning support
333333

334-
The AI Foundry model catalog offers fine-tuning support for multiple types of models, including chat completions and text generation. For a list of models that support fine-tuning and the Azure regions of support for fine-tuning, see [region availability for models in a standard deployment.](deploy-models-serverless-availability.md) Fine-tuning tasks are available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available. If the offer is available in the relevant region, the user then must have a project resource in the Azure region where the model is available for deployment or fine-tuning, as applicable.
334+
The AI Foundry model catalog offers fine-tuning support for multiple types of models, including chat completions and text generation. For a list of models that support fine-tuning and the Azure regions of support for fine-tuning, see [region availability for models in a serverless API deployment.](deploy-models-serverless-availability.md) Fine-tuning tasks are available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available. If the offer is available in the relevant region, the user then must have a project resource in the Azure region where the model is available for deployment or fine-tuning, as applicable.
335335

336336
For this example, we use a Phi-4-mini-instruct model. In this code snippet, the model ID property of the model will be passed as input to the fine tuning job. This is also available as the Asset ID field in model details page in Azure AI Foundry Model Catalog.
337337

@@ -577,7 +577,7 @@ model_id = f"azureml://locations/{workspace.location}/workspaces/{workspace._wor
577577

578578
### Supported enterprise scenarios for fine-tuning
579579

580-
Several enterprise scenarios are supported for standard deployment fine-tuning. The table below outlines the supported configurations for user storage networking and authentication to ensure smooth operation within enterprise scenarios:
580+
Several enterprise scenarios are supported for serverless API deployment fine-tuning. The table below outlines the supported configurations for user storage networking and authentication to ensure smooth operation within enterprise scenarios:
581581

582582
>[!Note]
583583
>- Data connections auth can be changed via AI Foundry by clicking on the datastore connection which your dataset is stored in, and navigating to the **Access details** > **Authentication Method** setting.
@@ -597,7 +597,7 @@ Several enterprise scenarios are supported for standard deployment fine-tuning.
597597

598598
The scenarios above should work in a Managed Vnet workspace as well. See setup of Managed Vnet AI Foundry hub here: [How to configure a managed network for Azure AI Foundry hubs](./configure-managed-network.md)
599599

600-
Customer-Managed Keys (CMKs) is **not** a supported enterprise scenario with standard deployment fine-tuning.
600+
Customer-Managed Keys (CMKs) is **not** a supported enterprise scenario with serverless API deployment fine-tuning.
601601

602602
Issues fine-tuning with unique network setups on the workspace and storage usually points to a networking setup issue.
603603

@@ -695,17 +695,17 @@ workspace_ml_client.serverless_endpoints.begin_delete(endpoint_name).result()
695695

696696
::: zone-end
697697

698-
## Cost and quota considerations for models deployed as a standard deployment
698+
## Cost and quota considerations for models deployed as a serverless API deployment
699699

700700
Quota is managed per deployment. Each deployment has a rate limit of 200,000 tokens per minute and 1,000 API requests per minute. However, we currently limit one deployment per model per project. Contact Microsoft Azure Support if the current rate limits aren't sufficient for your scenarios.
701701

702702
#### Cost for Microsoft models
703703

704-
You can find the pricing information on the __Pricing and terms__ tab of the deployment wizard when deploying Microsoft models (such as Phi-3 models) as a standard deployment.
704+
You can find the pricing information on the __Pricing and terms__ tab of the deployment wizard when deploying Microsoft models (such as Phi-3 models) as a serverless API deployment.
705705

706706
#### Cost for non-Microsoft models
707707

708-
Non-Microsoft models deployed as a standard deployment are offered through Azure Marketplace and integrated with Azure AI Foundry for use. You can find Azure Marketplace pricing when deploying or fine-tuning these models.
708+
Non-Microsoft models deployed as a serverless API deployment are offered through Azure Marketplace and integrated with Azure AI Foundry for use. You can find Azure Marketplace pricing when deploying or fine-tuning these models.
709709

710710
Each time a project subscribes to a given offer from Azure Marketplace, a new resource is created to track the costs associated with its consumption. The same resource is used to track costs associated with inference and fine-tuning; however, multiple meters are available to track each scenario independently.
711711

@@ -754,7 +754,7 @@ The training data used is the same as demonstrated in the SDK notebook. The CLI
754754

755755
## Content filtering
756756

757-
Standard deployment models are protected by Azure AI Content Safety. When deployed to real-time endpoints, you can opt out of this capability. With Azure AI Content Safety enabled, both the prompt and completion pass through an ensemble of classification models aimed at detecting and preventing the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions. Learn more about [Azure AI Content Safety](../concepts/content-filtering.md).
757+
serverless API deployment models are protected by Azure AI Content Safety. When deployed to real-time endpoints, you can opt out of this capability. With Azure AI Content Safety enabled, both the prompt and completion pass through an ensemble of classification models aimed at detecting and preventing the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions. Learn more about [Azure AI Content Safety](../concepts/content-filtering.md).
758758

759759

760760
## Next steps

articles/ai-foundry/how-to/healthcare-ai/deploy-cxrreportgen.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ To __deploy the model through the UI__:
5050
1. Go to the model catalog.
5151
1. Search for the _CxrReportGen_ model and select its model card.
5252
1. On the model's overview page, select __Deploy__.
53-
1. If given the option to choose between standard deployment and deployment using a managed compute, select **Managed Compute**.
53+
1. If given the option to choose between serverless API deployment and deployment using a managed compute, select **Managed Compute**.
5454
1. Fill out the details in the deployment window.
5555

5656
> [!NOTE]

articles/ai-foundry/how-to/healthcare-ai/deploy-medimageinsight.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ To __deploy the model through the UI__:
4949
1. Go to the model catalog.
5050
1. Search for the _MedImageInsight_ model and select its model card.
5151
1. On the model's overview page, select __Deploy__.
52-
1. If given the option to choose between standard deployment and deployment using a managed compute, select **Managed Compute**.
52+
1. If given the option to choose between serverless API deployment and deployment using a managed compute, select **Managed Compute**.
5353
1. Fill out the details in the deployment window.
5454

5555
> [!NOTE]

articles/ai-foundry/how-to/healthcare-ai/deploy-medimageparse.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@ To __deploy the model through the UI__:
6060
1. Go to the model catalog.
6161
1. Search for the model and select its model card.
6262
1. On the model's overview page, select __Deploy__.
63-
1. If given the option to choose between standard deployment and deployment using a managed compute, select **Managed Compute**.
63+
1. If given the option to choose between serverless API deployment and deployment using a managed compute, select **Managed Compute**.
6464
1. Fill out the details in the deployment window.
6565

6666
> [!NOTE]

articles/ai-foundry/how-to/quota.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,7 @@ A quota is a credit limit on Azure resources, not a capacity guarantee. If
5858
> [!NOTE]
5959
> Azure AI Foundry compute has a separate quota from the core compute quota.
6060
61-
Default limits vary by offer category type, such as free trial, standard deployment, and virtual machine (VM) series (such as Dv2, F, and G).
61+
Default limits vary by offer category type, such as free trial, serverless API deployment, and virtual machine (VM) series (such as Dv2, F, and G).
6262

6363
## Azure AI Foundry quota
6464

articles/ai-foundry/includes/content-safety-serverless-apis-note.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,4 +10,4 @@ ms.custom: include file
1010
---
1111

1212
> [!NOTE]
13-
> Azure AI Content Safety is currently available for models deployed as standard deployment, but not to models deployed via managed compute. To learn more about Azure AI Content Safety for models deployed as standard deployment, see [Guardrails & controls for Models Sold Directly by Azure ](../concepts/model-catalog-content-safety.md).
13+
> Azure AI Content Safety is currently available for models deployed as a serverless API, but not to models deployed via managed compute. To learn more about Azure AI Content Safety for models deployed as a serverless API, see [Guardrails & controls for Models Sold Directly by Azure ](../concepts/model-catalog-content-safety.md).

0 commit comments

Comments
 (0)