Skip to content

Commit 72c5e37

Browse files
authored
Merge branch 'main' into rework-how-to-fine-tune
2 parents d045a22 + 07a20c8 commit 72c5e37

File tree

260 files changed

+5016
-3969
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

260 files changed

+5016
-3969
lines changed

.openpublishing.redirection.json

Lines changed: 20 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,15 @@
11
{
22
"redirections": [
3+
{
4+
"source_path": "articles/search/performance-benchmarks.md",
5+
"redirect_url": "/previous-versions/azure/search/performance-benchmarks",
6+
"redirect_document_id": false
7+
},
8+
{
9+
"source_path": "articles/search/search-traffic-analytics.md",
10+
"redirect_url": "/previous-versions/azure/search/search-traffic-analytics",
11+
"redirect_document_id": false
12+
},
313
{
414
"source_path": "articles/genomics/frequently-asked-questions-genomics.yml",
515
"redirect_url": "/previous-versions/azure/genomics/frequently-asked-questions-genomics",
@@ -235,6 +245,16 @@
235245
"redirect_url": "/azure/open-datasets/dataset-catalog",
236246
"redirect_document_id": false
237247
},
248+
{
249+
"source_path_from_root": "/articles/ai-services/language-service/language-studio.md",
250+
"redirect_url": "/azure/ai-services/language-service/overview",
251+
"redirect_document_id": false
252+
},
253+
{
254+
"source_path_from_root": "/articles/ai-services/language-service/summarization/region-support.md",
255+
"redirect_url": "/azure/ai-services/language-service/concepts/regional-support",
256+
"redirect_document_id": false
257+
},
238258
{
239259
"source_path_from_root": "/articles/open-datasets/dataset-genomics-data-lake.md",
240260
"redirect_url": "/azure/open-datasets/dataset-catalog",

articles/ai-foundry/model-inference/faq.yml

Lines changed: 27 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -24,18 +24,19 @@ sections:
2424
Both Azure OpenAI Service and Azure AI model inference are part of the Azure AI services family and build on top of the same security and enterprise promise of Azure.
2525
2626
While Azure AI model inference focus on inference, Azure OpenAI Service can be used with more advanced APIs like batch, fine-tuning, assistants, and files.
27-
- question: |
28-
What's the difference between OpenAI and Azure OpenAI?
29-
answer: |
30-
Azure AI Models and Azure OpenAI Service give customers access to advanced language models from OpenAI with the security and enterprise promise of Azure. Azure OpenAI codevelops the APIs with OpenAI, ensuring compatibility and a smooth transition from one to the other.
31-
32-
Customers get the security capabilities of Microsoft Azure while running the same models as OpenAI. It offers private networking, regional availability, and responsible AI content filtering.
33-
34-
Learn more about the [Azure OpenAI service](../../ai-services/openai/overview.md).
3527
- question: |
3628
What's the difference between Azure AI services and Azure AI Foundry?
3729
answer: |
3830
Azure AI services are a suite of AI services that provide prebuilt APIs for common AI scenarios. Azure AI Services is part of the Azure AI Foundry platform. Azure AI services can be used in Azure AI Foundry portal to enhance your models with prebuilt AI capabilities.
31+
- question: |
32+
What's the difference between Serverless API Endpoints and Azure AI model inference?
33+
answer: |
34+
Both features allow you to deploy Models-as-a-Service models in Azure AI Foundry. However, there are some differences between them:
35+
- *Resource involved*: Serverless API Endpoints are deployed within an AI project resource, while Azure AI model inference is part of the Azure AI services resource.
36+
- *Deployment options*: Serverless API Endpoints allow regional deployments, while Azure AI model inference allows deployments under a global capacity.
37+
- *Models*: Azure AI model inference also supports deploying Azure OpenAI models.
38+
- *Endpoint*: Serverless API Endpoints creates one endpoint and credential per deployment, while Azure AI model inference creates one endpoint and credential per resource.
39+
- *Model router*: Azure AI model inference allows you to switch between models without changing your code using a model router.
3940
- name: Models
4041
questions:
4142
- question: |
@@ -44,6 +45,10 @@ sections:
4445
Azure AI model inference in AI services supports all the models in the Azure AI catalog having pay-as-you-go billing. For more information, see [the Models article](concepts/models.md).
4546
4647
The Azure AI model catalog contains a wider list of models, however, those models require compute quota from your subscription. They also need to have a project or AI hub where to host the deployment. For more information, see [deployment options in Azure AI Foundry](../../ai-studio/concepts/deployments-overview.md).
48+
- question: |
49+
My company hasn't approved specific models for use. How can I prevent users from deploying them?
50+
answer: |
51+
You can restrict the models available for deployment in Azure AI services by using the Azure policies. Models are listed in the catalog but any attempt to deploy them is blocked. Read [Control AI model deployment with custom policies](how-to/configure-deployment-policies.md).
4752
- name: SDKs and programming languages
4853
questions:
4954
- question: |
@@ -94,10 +99,23 @@ sections:
9499
You can set up a spending limit in the [Azure portal](https://portal.azure.com) under **Azure Cost Management + Billing**. This limit prevents you from spending more than the limit you set. Once spending limit is reached, the subscription will be disabled and you won't be able to use the endpoint until the next billing cycle.
95100
- name: Data and Privacy
96101
questions:
102+
- question: |
103+
How are third-party models available?
104+
answer: |
105+
Third-party models available for deployment in Azure AI Services with pay-as-you-go billing (for example, Meta AI models or Mistral models) are offered by the model provider but hosted in Microsoft-managed Azure infrastructure and accessed via API in the Azure AI model inference endpoint. Model providers define the license terms and set the price for use of their models, while Azure AI Services service manages the hosting infrastructure, makes the inference APIs available, and acts as the data processor for prompts submitted and content output by models deployed. Read about [Data privacy, and security for third-party models](../../ai-studio/how-to/concept-data-privacy.md).
106+
- question: |
107+
How is data processed by the Global-Standard deployment type?
108+
answer: |
109+
For model deployments under Azure AI Services resources, prompts and outputs are processed using Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global standard provides the highest default quota and eliminates the need to load balance across multiple resources. Data stored at rest remains in the designated Azure geography, while data may be processed for inferencing in any Azure location. Learn more about [data residency](https://azure.microsoft.com/explore/global-infrastructure/data-residency/).
97110
- question: |
98111
Do you use my company data to train any of the models?
99112
answer: |
100-
Azure AI model inference doesn't use customer data to retrain models, and customer data is never shared with model providers.
113+
Azure AI model inference doesn't use customer data to retrain models, and customer data is never shared with model providers.
114+
- question: |
115+
Is data shared with model providers?
116+
answer: |
117+
Microsoft acts as the data processor for prompts and outputs sent to, and generated by, a model deployment under Azure AI services resources. Microsoft doesn't share these prompts and outputs with the model provider. Also, Microsoft doesn't use these prompts and outputs to train or improve Microsoft models, the model provider's models, or any third party's models.
118+
As explained during the deployment process for Models-as-a-Service models, Microsoft might share customer contact information and transaction details (including the usage volume associated with the offering) with the model publisher so that the publisher can contact customers regarding the model. Learn more about information available to model publishers in [Access insights for the Microsoft commercial marketplace in Partner Center](/partner-center/analytics).
101119
- name: Customer Copyright Commitment
102120
questions:
103121
- question: |

articles/ai-foundry/model-inference/includes/configure-entra-id/intro.md

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -47,3 +47,17 @@ To complete this article, you need:
4747
* Security principal: e.g. your user account.
4848
* Role definition: the *Cognitive Services User* role.
4949
* Scope: the Azure AI Services resource.
50+
51+
* If you want to create a custom role definition instead of using *Cognitive Services User* role, ensure the role has the following permissions:
52+
53+
```json
54+
{
55+
"permissions": [
56+
{
57+
"dataActions": [
58+
"Microsoft.CognitiveServices/accounts/MaaS/*"
59+
]
60+
}
61+
]
62+
}
63+
```

articles/ai-foundry/model-inference/includes/configure-entra-id/portal.md

Lines changed: 17 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -14,34 +14,42 @@ zone_pivot_groups: azure-ai-models-deployment
1414

1515
Follow these steps to configure Microsoft Entra ID for inference:
1616

17-
1. Go to the [Azure portal](https://portal.azure.com) and locate the Azure AI Services resource you're using. If you're using Azure AI Foundry with projects or hubs, you can navigate to it by:
17+
1. Go to the [Azure portal](https://portal.azure.com) and locate the **Azure AI Services** resource you're using. If you're using Azure AI Foundry with projects or hubs, you can navigate to it by:
1818

1919
1. Go to [Azure AI Foundry portal](https://ai.azure.com).
2020

2121
2. On the landing page, select **Open management center**.
2222

2323
3. Go to the section **Connected resources** and select the connection to the Azure AI Services resource that you want to configure. If it isn't listed, select **View all** to see the full list.
2424

25+
:::image type="content" source="../../media/configure-entra-id/resource-behind-select.png" alt-text="Screenshot showing how to navigate to the details of the connection in Azure AI Foundry in the management center." lightbox="../../media/configure-entra-id/resource-behind-select.png":::
26+
2527
4. On the **Connection details** section, under **Resource**, select the name of the Azure resource. A new page opens.
2628

2729
5. You're now in [Azure portal](https://portal.azure.com) where you can manage all the aspects of the resource itself.
2830

29-
2. On the left navigation bar, select **Access control (IAM)**.
31+
:::image type="content" source="../../media/configure-entra-id/locate-resource-ai-services.png" alt-text="Screenshot showing the resource to which we configure Microsoft Entra ID." lightbox="../../media/configure-entra-id/locate-resource-ai-services.png":::
32+
33+
2. On the left navigation bar, select **Access control (IAM)** and then select **Add** > **Add role assignment**.
34+
35+
:::image type="content" source="../../media/configure-entra-id/resource-aim.png" alt-text="Screenshot showing how to add a role assignment in the Access control section of the resource in the Azure portal." lightbox="../../media/configure-entra-id/resource-aim.png":::
3036

3137
> [!TIP]
3238
> Use the **View my access** option to verify which roles are already assigned to you.
3339
34-
3. Select **Role assignments** and then select **Add** > **Add role assignment**.
40+
3. On **Job function roles**, type **Cognitive Services User**. The list of roles is filtered out.
3541

36-
4. On **Job function roles**, type **Cognitive Services User**. The list of roles is filtered out.
42+
:::image type="content" source="../../media/configure-entra-id/cognitive-services-user.png" alt-text="Screenshot showing how to select the Cognitive Services User role assignment." lightbox="../../media/configure-entra-id/cognitive-services-user.png":::
3743

38-
5. Select the role and select **Next**.
44+
4. Select the role and select **Next**.
3945

40-
6. On **Members**, select the user or group you want to grant access to. We recommend using security groups whenever possible as they are easier to manage and maintain.
46+
5. On **Members**, select the user or group you want to grant access to. We recommend using security groups whenever possible as they are easier to manage and maintain.
4147

42-
7. Select **Next** and finish the wizard.
48+
:::image type="content" source="../../media/configure-entra-id/select-user.png" alt-text="Screenshot showing how to select the user to whom assign the role." lightbox="../../media/configure-entra-id/select-user.png":::
4349

44-
8. The selected user can now use Microsoft Entra ID for inference.
50+
6. Select **Next** and finish the wizard.
51+
52+
7. The selected user can now use Microsoft Entra ID for inference.
4553

4654
> [!TIP]
4755
> Keep in mind that Azure role assignments may take up to five minutes to propagate. When working with security groups, adding or removing users from the security group propagates immediately.
@@ -84,6 +92,4 @@ To change this behavior, you have to update the connections from your projects t
8492

8593
## Disable key-based authentication in the resource
8694

87-
Disabling key-based authentication is advisable when you implemented Microsoft Entra ID and fully addressed compatibility or fallback concerns in all the applications that consume the service.
88-
89-
95+
Disabling key-based authentication is advisable when you implemented Microsoft Entra ID and fully addressed compatibility or fallback concerns in all the applications that consume the service. Disabling key-based authentication is only available when deploying using Bicep/ARM.

articles/ai-foundry/model-inference/includes/configure-entra-id/troubleshooting.md

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,21 @@ ms.date: 01/23/2025
77
ms.topic: include
88
---
99

10+
Before troubleshooting, verify that you have the right permissions assigned:
11+
12+
1. Go to the [Azure portal](https://portal.azure.com) and locate the **Azure AI Services** resource you're using.
13+
14+
2. On the left navigation bar, select **Access control (IAM)** and then select **Check access**.
15+
16+
3. Type the name of the user or identity you are using to connect to the service.
17+
18+
4. Verify that the role **Cognitive Services User** is listed (or a role that contains the required permissions as explained in [Prerequisites](#prerequisites)).
19+
20+
> [!IMPORTANT]
21+
> Roles like **Owner** or **Contributor** don't provide access via Microsoft Entra ID.
22+
23+
5. If not listed, follow the steps in this guide before continuing.
24+
1025
The following table contains multiple scenarios that can help troubleshooting Microsoft Entra ID:
1126

1227
| Error / Scenario | Root cause | Solution |

articles/ai-foundry/model-inference/includes/create-resources/bicep.md

Lines changed: 2 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -30,19 +30,9 @@ The files for this example are in:
3030
cd azureai-model-inference-bicep/infra
3131
```
3232

33-
## Understand the resources
34-
35-
The tutorial helps you create:
36-
37-
> [!div class="checklist"]
38-
> * An Azure AI Services resource.
39-
> * A model deployment in the Global standard SKU for each of the models supporting pay-as-you-go.
40-
> * (Optionally) An Azure AI project and hub.
41-
> * (Optionally) A connection between the hub and the models in Azure AI Services.
42-
43-
Notice that **you have to deploy an Azure AI project and hub** if you plan to use the Azure AI Foundry portal for managing the resource, using playground, or any other feature from the portal.
33+
## Create the resources
4434

45-
You are using the following assets to create those resources:
35+
Follow these steps:
4636

4737
1. Use the template `modules/ai-services-template.bicep` to describe your Azure AI Services resource:
4838

@@ -72,10 +62,6 @@ You are using the following assets to create those resources:
7262

7363
:::code language="bicep" source="~/azureai-model-inference-bicep/infra/modules/ai-services-connection-template.bicep":::
7464

75-
## Create the resources
76-
77-
In your console, follow these steps:
78-
7965
1. Define the main deployment:
8066

8167
__deploy-with-project.bicep__

articles/ai-foundry/model-inference/includes/create-resources/intro.md

Lines changed: 16 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,22 @@ ms.topic: include
1111

1212
In this article, you learn how to create the resources required to use Azure AI model inference and consume flagship models from Azure AI model catalog.
1313

14+
## Understand the resources
15+
16+
Azure AI model inference is a capability in Azure AI Services resources in Azure. You can create model deployments under the resource to consume their predictions. You can also connect the resource to Azure AI Hubs and Projects in Azure AI Foundry to create intelligent applications if needed. The following picture shows the high level architecture.
17+
18+
:::image type="content" source="../../media/create-resources/resources-architecture.png" alt-text="A diagram showing the high level architecture of the resources created in the tutorial." lightbox="../../media/create-resources/resources-architecture.png":::
19+
20+
Azure AI Services resources don't require AI projects or AI hubs to operate and you can create them to consume flagship models from your applications. However, additional capabilities are available if you **deploy an Azure AI project and hub**, including playground, or agents.
21+
22+
The tutorial helps you create:
23+
24+
> [!div class="checklist"]
25+
> * An Azure AI Services resource.
26+
> * A model deployment for each of the models supported with pay-as-you-go.
27+
> * (Optionally) An Azure AI project and hub.
28+
> * (Optionally) A connection between the hub and the models in Azure AI Services.
29+
1430
## Prerequisites
1531

1632
To complete this article, you need:
1.72 MB
Loading
1.14 MB
Loading
1.43 MB
Loading

0 commit comments

Comments
 (0)