You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
To run inference over large amounts of data, you can use batch endpoints to deploy models, including Azure OpenAI models. In this article, you see how to create a batch endpoint to deploy the text-embedding-ada-002 model from Azure OpenAI to compute embeddings at scale. You can use the same approach for completions and chat completions models.
21
+
To run inference over large amounts of data, you can use batch endpoints to deploy models, including Azure OpenAI models. In this article, you see how to create a batch endpoint to deploy the `text-embedding-ada-002` model from Azure OpenAI to compute embeddings at scale. You can use the same approach for completions and chat completions models.
22
22
23
-
The example in this article uses Microsoft Entra authentication to grant access to an Azure OpenAI Service resource, but you can also use an access key. The model is registered in MLflow format. It uses the Azure OpenAI flavor, which provides support for calling Azure OpenAI at scale.
23
+
The example in this article uses Microsoft Entra authentication to grant access to an Azure OpenAI Service resource, but you can also use an access key. The model is registered in MLflow format. It uses the Azure OpenAI flavor, which provides support for calling the Azure OpenAI service at scale.
24
24
25
25
To follow along with the example steps, see the Jupyter notebook [Score OpenAI models in batch using Batch Endpoints](https://github.com/Azure/azureml-examples/blob/main/sdk/python/endpoints/batch/deploy-models/openai-embeddings/deploy-and-test.ipynb).
In this article, you see how to create a deployment for an Azure OpenAI model. The following image shows a deployed Azure OpenAI model and the Azure OpenAI resource that it's deployed to:
71
+
In this article, you see how to create a deployment for an Azure OpenAI model. The following image shows a deployed Azure OpenAI model and highlights the Azure OpenAI resource that it's deployed to:
72
72
73
73
:::image type="content" source="./media/how-to-use-batch-model-openai-embeddings/azure-openai-deployments.png" alt-text="Screenshot of the Azure OpenAI page within Azure AI Foundry. A model deployment that's available in a particular Azure OpenAI resource is visible." lightbox="media/how-to-use-batch-model-openai-embeddings/azure-openai-deployments.png":::
74
74
@@ -140,7 +140,7 @@ You can configure the batch deployment to use the access key of your Azure OpenA
140
140
141
141
Model deployments in batch endpoints can deploy only registered models. You can use MLflow models with the Azure OpenAI flavor to create a model in your workspace that references a deployment in Azure OpenAI.
142
142
143
-
In the cloned repository, the model folder contains an MLflow model that generates embeddings based on the text-embedding-ada-002 model.
143
+
In the cloned repository, the model folder contains an MLflow model that generates embeddings based on the `text-embedding-ada-002` model.
0 commit comments