Skip to content

Commit cc2f5bd

Browse files
authored
Update quickstart-ai-project.md
1 parent b74668b commit cc2f5bd

File tree

1 file changed

+10
-9
lines changed

1 file changed

+10
-9
lines changed

articles/ai-foundry/model-inference/how-to/quickstart-ai-project.md

Lines changed: 10 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -14,17 +14,17 @@ recommendations: false
1414

1515
# Configure your AI project to use Azure AI model inference
1616

17-
If you already have an AI project in an existing AI Hub, models via "Models as a Service" are by default deployed inside of your project as stand-alone endpoints. Each model deployment has its own set of URI and credentials to access it. Azure OpenAI models are deployed to Azure AI Services resource or to the Azure OpenAI Service resource.
17+
If you already have an AI project in an existing AI Hub, models in the model catalog from third-party model providers are by default deployed inside of your project as stand-alone endpoints. Each model deployment has its own set of URI and credentials to access it. On the other hand, Azure OpenAI models are deployed to Azure AI Services resource or to the Azure OpenAI Service resource.
1818

19-
You can configure the AI project to connect with the Azure AI model inference in Azure AI services. Once configured, **deployments of Models as a Service models happen to the connected Azure AI Services resource** instead to the project itself, giving you a single set of endpoint and credential to access all the models deployed in Azure AI Foundry.
19+
You can configure the AI project to connect with the Azure AI model inference in Azure AI services. Once configured, **deployments of Models as a Service models supporting pay-as-you-go billing happen to the connected Azure AI Services resource** instead to the project itself, giving you a single set of endpoint and credential to access all the models deployed in Azure AI Foundry. Hence, you can manage Azure OpenAI and third-party model providers models in the same way.
2020

2121
Additionally, deploying models to Azure AI model inference brings the extra benefits of:
2222

2323
> [!div class="checklist"]
24-
> * [Routing capability](../concepts/endpoints.md#routing)
25-
> * [Custom content filters](../concepts/content-filter.md)
26-
> * Global capacity deployment
27-
> * Entra ID support and role-based access control
24+
> * [Routing capability](../concepts/endpoints.md#routing).
25+
> * [Custom content filters](../concepts/content-filter.md).
26+
> * Global capacity deployment type.
27+
> * [Key-less authentication](configure-entra-id.md) with role-based access control.
2828
2929
In this article, you learn how to configure your project to use models deployed in Azure AI model inference in Azure AI services.
3030

@@ -186,10 +186,11 @@ For each model deployed as Serverless API Endpoints, follow these steps:
186186

187187
## Limitations
188188

189-
Azure AI model inference in Azure AI Services gives users access to flagship models in the Azure AI model catalog. However, only models supporting pay-as-you-go billing (Models as a Service) are available for deployment.
189+
Consider the following limitaitons when configuring your project to use Azure AI model inference:
190190

191-
Models requiring compute quota from your subscription (Managed Compute), including custom models, can only be deployed within a given project as Managed Online Endpoints and continue to be accessible using their own set of endpoint URI and credentials.
191+
* Only models supporting pay-as-you-go billing (Models as a Service) are available for deployment to Azure AI model inference. Models requiring compute quota from your subscription (Managed Compute), including custom models, can only be deployed within a given project as Managed Online Endpoints and continue to be accessible using their own set of endpoint URI and credentials.
192+
* Models available as both pay-as-you-go billing and managed compute offerings are, by default, deployed to Azure AI model inference in Azure AI services resources. Azure AI Foundry portal doesn't offer a way to deploy them to Managed Online Endpoints. You have to turn off the feature mentioned at [Configure the project to use Azure AI model inference](#configure-the-project-to-use-azure-ai-model-inference) or use the Azure CLI/Azure ML SDK/ARM templates to perform the deployment.
192193

193194
## Next steps
194195

195-
* [Add more models](create-model-deployments.md) to your endpoint.
196+
* [Add more models](create-model-deployments.md) to your endpoint.

0 commit comments

Comments
 (0)