You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/ai-foundry/model-inference/how-to/quickstart-ai-project.md
+10-9Lines changed: 10 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -14,17 +14,17 @@ recommendations: false
14
14
15
15
# Configure your AI project to use Azure AI model inference
16
16
17
-
If you already have an AI project in an existing AI Hub, models via "Models as a Service" are by default deployed inside of your project as stand-alone endpoints. Each model deployment has its own set of URI and credentials to access it. Azure OpenAI models are deployed to Azure AI Services resource or to the Azure OpenAI Service resource.
17
+
If you already have an AI project in an existing AI Hub, models in the model catalog from third-party model providers are by default deployed inside of your project as stand-alone endpoints. Each model deployment has its own set of URI and credentials to access it. On the other hand, Azure OpenAI models are deployed to Azure AI Services resource or to the Azure OpenAI Service resource.
18
18
19
-
You can configure the AI project to connect with the Azure AI model inference in Azure AI services. Once configured, **deployments of Models as a Service models happen to the connected Azure AI Services resource** instead to the project itself, giving you a single set of endpoint and credential to access all the models deployed in Azure AI Foundry.
19
+
You can configure the AI project to connect with the Azure AI model inference in Azure AI services. Once configured, **deployments of Models as a Service models supporting pay-as-you-go billing happen to the connected Azure AI Services resource** instead to the project itself, giving you a single set of endpoint and credential to access all the models deployed in Azure AI Foundry. Hence, you can manage Azure OpenAI and third-party model providers models in the same way.
20
20
21
21
Additionally, deploying models to Azure AI model inference brings the extra benefits of:
> *[Key-less authentication](configure-entra-id.md) with role-based access control.
28
28
29
29
In this article, you learn how to configure your project to use models deployed in Azure AI model inference in Azure AI services.
30
30
@@ -186,10 +186,11 @@ For each model deployed as Serverless API Endpoints, follow these steps:
186
186
187
187
## Limitations
188
188
189
-
Azure AI model inference in Azure AI Services gives users access to flagship models in the Azure AI model catalog. However, only models supporting pay-as-you-go billing (Models as a Service) are available for deployment.
189
+
Consider the following limitaitons when configuring your project to use Azure AI model inference:
190
190
191
-
Models requiring compute quota from your subscription (Managed Compute), including custom models, can only be deployed within a given project as Managed Online Endpoints and continue to be accessible using their own set of endpoint URI and credentials.
191
+
* Only models supporting pay-as-you-go billing (Models as a Service) are available for deployment to Azure AI model inference. Models requiring compute quota from your subscription (Managed Compute), including custom models, can only be deployed within a given project as Managed Online Endpoints and continue to be accessible using their own set of endpoint URI and credentials.
192
+
* Models available as both pay-as-you-go billing and managed compute offerings are, by default, deployed to Azure AI model inference in Azure AI services resources. Azure AI Foundry portal doesn't offer a way to deploy them to Managed Online Endpoints. You have to turn off the feature mentioned at [Configure the project to use Azure AI model inference](#configure-the-project-to-use-azure-ai-model-inference) or use the Azure CLI/Azure ML SDK/ARM templates to perform the deployment.
192
193
193
194
## Next steps
194
195
195
-
*[Add more models](create-model-deployments.md) to your endpoint.
196
+
*[Add more models](create-model-deployments.md) to your endpoint.
0 commit comments