You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/model-inference/faq.yml
+27-9Lines changed: 27 additions & 9 deletions
Original file line number
Diff line number
Diff line change
@@ -24,18 +24,19 @@ sections:
24
24
Both Azure OpenAI Service and Azure AI model inference are part of the Azure AI services family and build on top of the same security and enterprise promise of Azure.
25
25
26
26
While Azure AI model inference focus on inference, Azure OpenAI Service can be used with more advanced APIs like batch, fine-tuning, assistants, and files.
27
-
- question: |
28
-
What's the difference between OpenAI and Azure OpenAI?
29
-
answer: |
30
-
Azure AI Models and Azure OpenAI Service give customers access to advanced language models from OpenAI with the security and enterprise promise of Azure. Azure OpenAI codevelops the APIs with OpenAI, ensuring compatibility and a smooth transition from one to the other.
31
-
32
-
Customers get the security capabilities of Microsoft Azure while running the same models as OpenAI. It offers private networking, regional availability, and responsible AI content filtering.
33
-
34
-
Learn more about the [Azure OpenAI service](../../ai-services/openai/overview.md).
35
27
- question: |
36
28
What's the difference between Azure AI services and Azure AI Foundry?
37
29
answer: |
38
30
Azure AI services are a suite of AI services that provide prebuilt APIs for common AI scenarios. Azure AI Services is part of the Azure AI Foundry platform. Azure AI services can be used in Azure AI Foundry portal to enhance your models with prebuilt AI capabilities.
31
+
- question: |
32
+
What's the difference between Serverless API Endpoints and Azure AI model inference?
33
+
answer: |
34
+
Both features allow you to deploy Models-as-a-Service models in Azure AI Foundry. However, there are some differences between them:
35
+
- *Resource involved*: Serverless API Endpoints are deployed within an AI project resource, while Azure AI model inference is part of the Azure AI services resource.
36
+
- *Deployment options*: Serverless API Endpoints allow regional deployments, while Azure AI model inference allows deployments under a global capacity.
37
+
- *Models*: Azure AI model inference also supports deploying Azure OpenAI models.
38
+
- *Endpoint*: Serverless API Endpoints creates one endpoint and credential per deployment, while Azure AI model inference creates one endpoint and credential per resource.
39
+
- *Model router*: Azure AI model inference allows you to switch between models without changing your code using a model router.
39
40
- name: Models
40
41
questions:
41
42
- question: |
@@ -44,6 +45,10 @@ sections:
44
45
Azure AI model inference in AI services supports all the models in the Azure AI catalog having pay-as-you-go billing. For more information, see [the Models article](concepts/models.md).
45
46
46
47
The Azure AI model catalog contains a wider list of models, however, those models require compute quota from your subscription. They also need to have a project or AI hub where to host the deployment. For more information, see [deployment options in Azure AI Foundry](../../ai-studio/concepts/deployments-overview.md).
48
+
- question: |
49
+
My company hasn't approved specific models for use. How can I prevent users from deploying them?
50
+
answer: |
51
+
You can restrict the models available for deployment in Azure AI services by using the Azure policies. Models are listed in the catalog but any attempt to deploy them is blocked. Read [Control AI model deployment with custom policies](how-to/configure-deployment-policies.md).
47
52
- name: SDKs and programming languages
48
53
questions:
49
54
- question: |
@@ -94,10 +99,23 @@ sections:
94
99
You can set up a spending limit in the [Azure portal](https://portal.azure.com) under **Azure Cost Management + Billing**. This limit prevents you from spending more than the limit you set. Once spending limit is reached, the subscription will be disabled and you won't be able to use the endpoint until the next billing cycle.
95
100
- name: Data and Privacy
96
101
questions:
102
+
- question: |
103
+
How are third-party models available?
104
+
answer: |
105
+
Third-party models available for deployment in Azure AI Services with pay-as-you-go billing (for example, Meta AI models or Mistral models) are offered by the model provider but hosted in Microsoft-managed Azure infrastructure and accessed via API in the Azure AI model inference endpoint. Model providers define the license terms and set the price for use of their models, while Azure AI Services service manages the hosting infrastructure, makes the inference APIs available, and acts as the data processor for prompts submitted and content output by models deployed. Read about [Data privacy, and security for third-party models](../../ai-studio/how-to/concept-data-privacy.md).
106
+
- question: |
107
+
How is data processed by the Global-Standard deployment type?
108
+
answer: |
109
+
For model deployments under Azure AI Services resources, prompts and outputs are processed using Azure's global infrastructure to dynamically route traffic to the data center with best availability for each request. Global standard provides the highest default quota and eliminates the need to load balance across multiple resources. Data stored at rest remains in the designated Azure geography, while data may be processed for inferencing in any Azure location. Learn more about [data residency](https://azure.microsoft.com/explore/global-infrastructure/data-residency/).
97
110
- question: |
98
111
Do you use my company data to train any of the models?
99
112
answer: |
100
-
Azure AI model inference doesn't use customer data to retrain models, and customer data is never shared with model providers.
113
+
Azure AI model inference doesn't use customer data to retrain models, and customer data is never shared with model providers.
114
+
- question: |
115
+
Is data shared with model providers?
116
+
answer: |
117
+
Microsoft acts as the data processor for prompts and outputs sent to, and generated by, a model deployment under Azure AI services resources. Microsoft doesn't share these prompts and outputs with the model provider. Also, Microsoft doesn't use these prompts and outputs to train or improve Microsoft models, the model provider's models, or any third party's models.
118
+
As explained during the deployment process for Models-as-a-Service models, Microsoft might share customer contact information and transaction details (including the usage volume associated with the offering) with the model publisher so that the publisher can contact customers regarding the model. Learn more about information available to model publishers in [Access insights for the Microsoft commercial marketplace in Partner Center](/partner-center/analytics).
Follow these steps to configure Microsoft Entra ID for inference:
16
16
17
-
1. Go to the [Azure portal](https://portal.azure.com) and locate the Azure AI Services resource you're using. If you're using Azure AI Foundry with projects or hubs, you can navigate to it by:
17
+
1. Go to the [Azure portal](https://portal.azure.com) and locate the **Azure AI Services** resource you're using. If you're using Azure AI Foundry with projects or hubs, you can navigate to it by:
18
18
19
19
1. Go to [Azure AI Foundry portal](https://ai.azure.com).
20
20
21
21
2. On the landing page, select **Open management center**.
22
22
23
23
3. Go to the section **Connected resources** and select the connection to the Azure AI Services resource that you want to configure. If it isn't listed, select **View all** to see the full list.
24
24
25
+
:::image type="content" source="../../media/configure-entra-id/resource-behind-select.png" alt-text="Screenshot showing how to navigate to the details of the connection in Azure AI Foundry in the management center." lightbox="../../media/configure-entra-id/resource-behind-select.png":::
26
+
25
27
4. On the **Connection details** section, under **Resource**, select the name of the Azure resource. A new page opens.
26
28
27
29
5. You're now in [Azure portal](https://portal.azure.com) where you can manage all the aspects of the resource itself.
28
30
29
-
2. On the left navigation bar, select **Access control (IAM)**.
31
+
:::image type="content" source="../../media/configure-entra-id/locate-resource-ai-services.png" alt-text="Screenshot showing the resource to which we configure Microsoft Entra ID." lightbox="../../media/configure-entra-id/locate-resource-ai-services.png":::
32
+
33
+
2. On the left navigation bar, select **Access control (IAM)** and then select **Add** > **Add role assignment**.
34
+
35
+
:::image type="content" source="../../media/configure-entra-id/resource-aim.png" alt-text="Screenshot showing how to add a role assignment in the Access control section of the resource in the Azure portal." lightbox="../../media/configure-entra-id/resource-aim.png":::
30
36
31
37
> [!TIP]
32
38
> Use the **View my access** option to verify which roles are already assigned to you.
33
39
34
-
3.Select**Role assignments** and then select **Add** > **Add role assignment**.
40
+
3.On**Job function roles**, type **Cognitive Services User**. The list of roles is filtered out.
35
41
36
-
4. On **Job function roles**, type**Cognitive Services User**. The list of roles is filtered out.
42
+
:::image type="content" source="../../media/configure-entra-id/cognitive-services-user.png" alt-text="Screenshot showing how to select the Cognitive Services User role assignment." lightbox="../../media/configure-entra-id/cognitive-services-user.png":::
37
43
38
-
5. Select the role and select **Next**.
44
+
4. Select the role and select **Next**.
39
45
40
-
6. On **Members**, select the user or group you want to grant access to. We recommend using security groups whenever possible as they are easier to manage and maintain.
46
+
5. On **Members**, select the user or group you want to grant access to. We recommend using security groups whenever possible as they are easier to manage and maintain.
41
47
42
-
7. Select **Next** and finish the wizard.
48
+
:::image type="content" source="../../media/configure-entra-id/select-user.png" alt-text="Screenshot showing how to select the user to whom assign the role." lightbox="../../media/configure-entra-id/select-user.png":::
43
49
44
-
8. The selected user can now use Microsoft Entra ID for inference.
50
+
6. Select **Next** and finish the wizard.
51
+
52
+
7. The selected user can now use Microsoft Entra ID for inference.
45
53
46
54
> [!TIP]
47
55
> Keep in mind that Azure role assignments may take up to five minutes to propagate. When working with security groups, adding or removing users from the security group propagates immediately.
@@ -84,6 +92,4 @@ To change this behavior, you have to update the connections from your projects t
84
92
85
93
## Disable key-based authentication in the resource
86
94
87
-
Disabling key-based authentication is advisable when you implemented Microsoft Entra ID and fully addressed compatibility or fallback concerns in all the applications that consume the service.
88
-
89
-
95
+
Disabling key-based authentication is advisable when you implemented Microsoft Entra ID and fully addressed compatibility or fallback concerns in all the applications that consume the service. Disabling key-based authentication is only available when deploying using Bicep/ARM.
Copy file name to clipboardExpand all lines: articles/ai-foundry/model-inference/includes/configure-entra-id/troubleshooting.md
+15Lines changed: 15 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,6 +7,21 @@ ms.date: 01/23/2025
7
7
ms.topic: include
8
8
---
9
9
10
+
Before troubleshooting, verify that you have the right permissions assigned:
11
+
12
+
1. Go to the [Azure portal](https://portal.azure.com) and locate the **Azure AI Services** resource you're using.
13
+
14
+
2. On the left navigation bar, select **Access control (IAM)** and then select **Check access**.
15
+
16
+
3. Type the name of the user or identity you are using to connect to the service.
17
+
18
+
4. Verify that the role **Cognitive Services User** is listed (or a role that contains the required permissions as explained in [Prerequisites](#prerequisites)).
19
+
20
+
> [!IMPORTANT]
21
+
> Roles like **Owner** or **Contributor** don't provide access via Microsoft Entra ID.
22
+
23
+
5. If not listed, follow the steps in this guide before continuing.
24
+
10
25
The following table contains multiple scenarios that can help troubleshooting Microsoft Entra ID:
Copy file name to clipboardExpand all lines: articles/ai-foundry/model-inference/includes/create-resources/bicep.md
+2-16Lines changed: 2 additions & 16 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -30,19 +30,9 @@ The files for this example are in:
30
30
cd azureai-model-inference-bicep/infra
31
31
```
32
32
33
-
## Understand the resources
34
-
35
-
The tutorial helps you create:
36
-
37
-
> [!div class="checklist"]
38
-
> * An Azure AI Services resource.
39
-
> * A model deployment in the Global standard SKU for each of the models supporting pay-as-you-go.
40
-
> * (Optionally) An Azure AI project and hub.
41
-
> * (Optionally) A connection between the hub and the models in Azure AI Services.
42
-
43
-
Notice that **you have to deploy an Azure AI project and hub** if you plan to use the Azure AI Foundry portal for managing the resource, using playground, or any other feature from the portal.
33
+
## Create the resources
44
34
45
-
You are using the following assets to create those resources:
35
+
Follow these steps:
46
36
47
37
1. Use the template `modules/ai-services-template.bicep` to describe your Azure AI Services resource:
48
38
@@ -72,10 +62,6 @@ You are using the following assets to create those resources:
Copy file name to clipboardExpand all lines: articles/ai-foundry/model-inference/includes/create-resources/intro.md
+16Lines changed: 16 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,6 +11,22 @@ ms.topic: include
11
11
12
12
In this article, you learn how to create the resources required to use Azure AI model inference and consume flagship models from Azure AI model catalog.
13
13
14
+
## Understand the resources
15
+
16
+
Azure AI model inference is a capability in Azure AI Services resources in Azure. You can create model deployments under the resource to consume their predictions. You can also connect the resource to Azure AI Hubs and Projects in Azure AI Foundry to create intelligent applications if needed. The following picture shows the high level architecture.
17
+
18
+
:::image type="content" source="../../media/create-resources/resources-architecture.png" alt-text="A diagram showing the high level architecture of the resources created in the tutorial." lightbox="../../media/create-resources/resources-architecture.png":::
19
+
20
+
Azure AI Services resources don't require AI projects or AI hubs to operate and you can create them to consume flagship models from your applications. However, additional capabilities are available if you **deploy an Azure AI project and hub**, including playground, or agents.
21
+
22
+
The tutorial helps you create:
23
+
24
+
> [!div class="checklist"]
25
+
> * An Azure AI Services resource.
26
+
> * A model deployment for each of the models supported with pay-as-you-go.
27
+
> * (Optionally) An Azure AI project and hub.
28
+
> * (Optionally) A connection between the hub and the models in Azure AI Services.
0 commit comments