You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/how-to/develop/llama-index.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -33,7 +33,7 @@ To run this tutorial, you need:
33
33
* An Azure AI project as explained at [Create a project in Azure AI Foundry portal](../create-projects.md).
34
34
* A model supporting the [Model Inference API](https://aka.ms/azureai/modelinference) deployed. In this example, we use a `Mistral-Large` deployment, but use any model of your preference. For using embeddings capabilities in LlamaIndex, you need an embedding model like `cohere-embed-v3-multilingual`.
35
35
36
-
* You can follow the instructions at [Deploy models as standard deployments](../deploy-models-serverless.md).
36
+
* You can follow the instructions at [Deploy models as serverless API deployments](../deploy-models-serverless.md).
Copy file name to clipboardExpand all lines: articles/ai-foundry/how-to/develop/semantic-kernel.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -21,7 +21,7 @@ In this article, you learn how to use [Semantic Kernel](/semantic-kernel/overvie
21
21
- An Azure AI project as explained at [Create a project in Azure AI Foundry portal](../create-projects.md).
22
22
- A model supporting the [Azure AI Model Inference API](../../../ai-foundry/model-inference/reference/reference-model-inference-api.md?tabs=python) deployed. In this example, we use a `Mistral-Large` deployment, but use any model of your preference. For using embeddings capabilities in LlamaIndex, you need an embedding model like `cohere-embed-v3-multilingual`.
23
23
24
-
- You can follow the instructions at [Deploy models as standard deployments](../deploy-models-serverless.md).
24
+
- You can follow the instructions at [Deploy models as serverless API deployments](../deploy-models-serverless.md).
25
25
26
26
- Python **3.10** or later installed, including pip.
Azure AI Foundry enables you to customize large language models to your specific datasets through a process called fine-tuning. This process offers significant benefits by allowing for customization and optimization tailored to specific tasks and applications. The advantages include improved performance, cost efficiency, reduced latency, and tailored outputs.
22
22
23
-
**Cost Efficiency**: Azure AI Foundry's fine-tuning can be more cost-effective, especially for large-scale deployments, thanks to Standard pricing.
23
+
**Cost Efficiency**: Azure AI Foundry's fine-tuning can be more cost-effective, especially for large-scale deployments, thanks to pay-as-you-go pricing.
24
24
25
-
**Model Variety**: Azure AI Foundry's standard deployment fine-tuning offers support for both proprietary and open-source models, providing users with the flexibility to select the models that best suit their needs without being restricted to a single type.
25
+
**Model Variety**: Azure AI Foundry's serverless API deployment fine-tuning offers support for both proprietary and open-source models, providing users with the flexibility to select the models that best suit their needs without being restricted to a single type.
26
26
27
27
**Customization and Control**: Azure AI Foundry provides greater customization and control over the fine-tuning process, enabling users to tailor models more precisely to their specific requirements.
28
28
29
-
In this article, you will discover how to fine-tune models that are deployed using standard deployments in [Azure AI Foundry](https://ai.azure.com/?cid=learnDocs).
29
+
In this article, you will discover how to fine-tune models that are deployed using serverless API deployments in [Azure AI Foundry](https://ai.azure.com/?cid=learnDocs).
30
30
31
31
32
32
## Prerequisites
@@ -51,7 +51,7 @@ Verify the subscription is registered to the Microsoft.Network resource provider
51
51
52
52
## Find models with fine-tuning support
53
53
54
-
The AI Foundry model catalog offers fine-tuning support for multiple types of models, including chat completions and text generation. For a list of models that support fine-tuning and the Azure regions of support for fine-tuning, see [region availability for models as standard deployment.](deploy-models-serverless-availability.md) Fine-tuning tasks are available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available. If the offer is available in the relevant region, the user then must have a project resource in the Azure region where the model is available for deployment or fine-tuning, as applicable.
54
+
The AI Foundry model catalog offers fine-tuning support for multiple types of models, including chat completions and text generation. For a list of models that support fine-tuning and the Azure regions of support for fine-tuning, see [region availability for models as serverless API deployment.](deploy-models-serverless-availability.md) Fine-tuning tasks are available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available. If the offer is available in the relevant region, the user then must have a project resource in the Azure region where the model is available for deployment or fine-tuning, as applicable.
55
55
56
56
57
57
You can also go to the Azure AI Foundry portal to view all models that contain fine-tuning support:
@@ -121,8 +121,8 @@ Azure AI Foundry portal provides the Create custom model wizard, so you can inte
121
121
### Select the base model
122
122
123
123
1. Choose the model you want to fine-tune from the Azure AI Foundry [model catalog](https://ai.azure.com/explore/models).
124
-
2. On the model's **Details page**, select **fine-tune**. Some foundation models support both **standard deployment** and **Managed compute**, while others support one or the other.
125
-
3. If you're presented the options for **standard deployment** and [**Managed compute**](./fine-tune-managed-compute.md), select **standard deployment** for fine-tuning. This action opens up a wizard that shows information about **Standard** fine-tuning for your model.
124
+
2. On the model's **Details page**, select **fine-tune**. Some foundation models support both **serverless API deployment** and **Managed compute**, while others support one or the other.
125
+
3. If you're presented the options for **serverless API deployment** and [**Managed compute**](./fine-tune-managed-compute.md), select **serverless API deployment** for fine-tuning. This action opens up a wizard that shows information about **Serverless API** fine-tuning for your model.
126
126
127
127
### Choose your training data
128
128
The next step is to either choose existing prepared training data or upload new prepared training data to use when customizing your model. The **Training data** pane displays any existing, previously uploaded datasets and also provides options to upload new training data.
@@ -210,7 +210,7 @@ Here are some of the tasks you can do on the **Models** tab:
210
210
211
211
### Supported enterprise scenarios for fine-tuning
212
212
213
-
Several enterprise scenarios are supported for standard deployment fine-tuning. The table below outlines the supported configurations for user storage networking and authentication to ensure smooth operation within enterprise scenarios:
213
+
Several enterprise scenarios are supported for serverless API deployment fine-tuning. The table below outlines the supported configurations for user storage networking and authentication to ensure smooth operation within enterprise scenarios:
214
214
215
215
>[!Note]
216
216
>- Data connections auth can be changed via AI Foundry by clicking on the datastore connection which your dataset is stored in, and navigating to the **Access details** > **Authentication Method** setting.
@@ -230,7 +230,7 @@ Several enterprise scenarios are supported for standard deployment fine-tuning.
230
230
231
231
The scenarios above should work in a Managed Vnet workspace as well. See setup of Managed Vnet AI Foundry hub here: [How to configure a managed network for Azure AI Foundry hubs](./configure-managed-network.md)
232
232
233
-
Customer-Managed Keys (CMKs) is **not** a supported enterprise scenario with standard deployment fine-tuning.
233
+
Customer-Managed Keys (CMKs) is **not** a supported enterprise scenario with serverless API deployment fine-tuning.
234
234
235
235
Issues fine-tuning with unique network setups on the workspace and storage usually points to a networking setup issue.
236
236
@@ -331,7 +331,7 @@ workspace.id
331
331
332
332
### Find models with fine-tuning support
333
333
334
-
The AI Foundry model catalog offers fine-tuning support for multiple types of models, including chat completions and text generation. For a list of models that support fine-tuning and the Azure regions of support for fine-tuning, see [region availability for models in a standard deployment.](deploy-models-serverless-availability.md) Fine-tuning tasks are available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available. If the offer is available in the relevant region, the user then must have a project resource in the Azure region where the model is available for deployment or fine-tuning, as applicable.
334
+
The AI Foundry model catalog offers fine-tuning support for multiple types of models, including chat completions and text generation. For a list of models that support fine-tuning and the Azure regions of support for fine-tuning, see [region availability for models in a serverless API deployment.](deploy-models-serverless-availability.md) Fine-tuning tasks are available only to users whose Azure subscription belongs to a billing account in a country/region where the model provider has made the offer available. If the offer is available in the relevant region, the user then must have a project resource in the Azure region where the model is available for deployment or fine-tuning, as applicable.
335
335
336
336
For this example, we use a Phi-4-mini-instruct model. In this code snippet, the model ID property of the model will be passed as input to the fine tuning job. This is also available as the Asset ID field in model details page in Azure AI Foundry Model Catalog.
### Supported enterprise scenarios for fine-tuning
579
579
580
-
Several enterprise scenarios are supported forstandard deployment fine-tuning. The table below outlines the supported configurations for user storage networking and authentication to ensure smooth operation within enterprise scenarios:
580
+
Several enterprise scenarios are supported forserverless API deployment fine-tuning. The table below outlines the supported configurations for user storage networking and authentication to ensure smooth operation within enterprise scenarios:
581
581
582
582
>[!Note]
583
583
>- Data connections auth can be changed via AI Foundry by clicking on the datastore connection which your dataset is stored in, and navigating to the **Access details**>**Authentication Method** setting.
@@ -597,7 +597,7 @@ Several enterprise scenarios are supported for standard deployment fine-tuning.
597
597
598
598
The scenarios above should work in a Managed Vnet workspace as well. See setup of Managed Vnet AI Foundry hub here: [How to configure a managed network for Azure AI Foundry hubs](./configure-managed-network.md)
## Cost and quota considerations for models deployed as a standard deployment
698
+
## Cost and quota considerations for models deployed as a serverless API deployment
699
699
700
700
Quota is managed per deployment. Each deployment has a rate limit of 200,000 tokens per minute and1,000API requests per minute. However, we currently limit one deployment per model per project. Contact Microsoft Azure Support if the current rate limits aren't sufficient for your scenarios.
701
701
702
702
#### Cost for Microsoft models
703
703
704
-
You can find the pricing information on the __Pricing and terms__ tab of the deployment wizard when deploying Microsoft models (such as Phi-3 models) as a standard deployment.
704
+
You can find the pricing information on the __Pricing and terms__ tab of the deployment wizard when deploying Microsoft models (such as Phi-3 models) as a serverless API deployment.
705
705
706
706
#### Cost for non-Microsoft models
707
707
708
-
Non-Microsoft models deployed as a standard deployment are offered through Azure Marketplace and integrated with Azure AI Foundry for use. You can find Azure Marketplace pricing when deploying or fine-tuning these models.
708
+
Non-Microsoft models deployed as a serverless API deployment are offered through Azure Marketplace and integrated with Azure AI Foundry for use. You can find Azure Marketplace pricing when deploying or fine-tuning these models.
709
709
710
710
Each time a project subscribes to a given offer from Azure Marketplace, a new resource is created to track the costs associated with its consumption. The same resource is used to track costs associated with inference and fine-tuning; however, multiple meters are available to track each scenario independently.
711
711
@@ -754,7 +754,7 @@ The training data used is the same as demonstrated in the SDK notebook. The CLI
754
754
755
755
## Content filtering
756
756
757
-
Standard deployment models are protected by Azure AI Content Safety. When deployed to real-time endpoints, you can opt out of this capability. With Azure AI Content Safety enabled, both the prompt and completion pass through an ensemble of classification models aimed at detecting and preventing the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions. Learn more about [Azure AI Content Safety](../concepts/content-filtering.md).
757
+
serverless API deployment models are protected by Azure AI Content Safety. When deployed to real-time endpoints, you can opt out of this capability. With Azure AI Content Safety enabled, both the prompt and completion pass through an ensemble of classification models aimed at detecting and preventing the output of harmful content. The content filtering system detects and takes action on specific categories of potentially harmful content in both input prompts and output completions. Learn more about [Azure AI Content Safety](../concepts/content-filtering.md).
Copy file name to clipboardExpand all lines: articles/ai-foundry/includes/content-safety-serverless-apis-note.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -10,4 +10,4 @@ ms.custom: include file
10
10
---
11
11
12
12
> [!NOTE]
13
-
> Azure AI Content Safety is currently available for models deployed as standard deployment, but not to models deployed via managed compute. To learn more about Azure AI Content Safety for models deployed as standard deployment, see [Guardrails & controls for Models Sold Directly by Azure ](../concepts/model-catalog-content-safety.md).
13
+
> Azure AI Content Safety is currently available for models deployed as a serverless API, but not to models deployed via managed compute. To learn more about Azure AI Content Safety for models deployed as a serverless API, see [Guardrails & controls for Models Sold Directly by Azure ](../concepts/model-catalog-content-safety.md).
0 commit comments