Skip to content

Commit 00f2017

Browse files
Freshness, in process.
1 parent aadd1f6 commit 00f2017

File tree

1 file changed

+7
-7
lines changed

1 file changed

+7
-7
lines changed

articles/machine-learning/how-to-use-batch-azure-data-factory.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ Azure Data Factory allows the creation of pipelines that can orchestrate multipl
2727
## Prerequisites
2828

2929
- This example assumes that you have a model correctly deployed as a batch endpoint. Particularly, use the *heart condition classifier* created in the tutorial [Using MLflow models in batch deployments](how-to-mlflow-batch.md).
30-
- An Azure Data Factory resource created and configured. If you haven't created your data factory yet, follow the steps in [Quickstart: Create a data factory by using the Azure portal and Azure Data Factory Studio](../data-factory/quickstart-create-data-factory-portal.md) to create one.
30+
- An Azure Data Factory resource created and configured. If you haven't created your data factory yet, follow the steps in [Quickstart: Create a data factory by using the Azure portal](../data-factory/quickstart-create-data-factory-portal.md) to create one.
3131
- After creating it, browse to the data factory in the Azure portal:
3232

3333
:::image type="content" source="~/reusable-content/ce-skilling/azure/media/data-factory/data-factory-home-page.png" alt-text="Screenshot of the home page for the Azure Data Factory, with the Open Azure Data Factory Studio tile.":::
@@ -36,14 +36,14 @@ Azure Data Factory allows the creation of pipelines that can orchestrate multipl
3636

3737
## Authenticating against batch endpoints
3838

39-
Azure Data Factory can invoke the REST APIs of batch endpoints by using the [Web Invoke](../data-factory/control-flow-web-activity.md) activity. Batch endpoints support Microsoft Entra ID for authorization and hence the request made to the APIs require a proper authentication handling.
39+
Azure Data Factory can invoke the REST APIs of batch endpoints by using the *Web Invoke* activity. Batch endpoints support Microsoft Entra ID for authorization and hence the request made to the APIs require a proper authentication handling. For more information, see [Web activity in Azure Data Factory and Azure Synapse Analytics](../data-factory/control-flow-web-activity.md).
4040

4141
You can use a service principal or a [managed identity](../active-directory/managed-identities-azure-resources/overview.md) to authenticate against Batch Endpoints. We recommend using a managed identity as it simplifies the use of secrets.
4242

4343
# [Using a Managed Identity](#tab/mi)
4444

4545
1. You can use Azure Data Factory managed identity to communicate with Batch Endpoints. In this case, you only need to make sure that your Azure Data Factory resource was deployed with a managed identity.
46-
1. If you don't have an Azure Data Factory resource or it was already deployed without a managed identity, follow the following steps to create it: [Managed identity for Azure Data Factory](../data-factory/data-factory-service-identity.md#system-assigned-managed-identity).
46+
1. If you don't have an Azure Data Factory resource or it was already deployed without a managed identity, follow the following steps to create it: [System-assigned managed identity](../data-factory/data-factory-service-identity.md#system-assigned-managed-identity).
4747

4848
> [!WARNING]
4949
> Changing the resource identity once deployed is not possible in Azure Data Factory. After the resource is created, you need to recreate it if you need to change the identity of it.
@@ -56,7 +56,7 @@ You can use a service principal or a [managed identity](../active-directory/mana
5656

5757
# [Using a Service Principal](#tab/sp)
5858

59-
1. Create a service principal following the steps at [Register an application with Microsoft Entra ID and create a service principal](../active-directory/develop/howto-create-service-principal-portal.md#register-an-application-with-azure-ad-and-create-a-service-principal).
59+
1. Create a service principal following the steps at [Register a Microsoft Entra app and create a service principal](../active-directory/develop/howto-create-service-principal-portal.md#register-an-application-with-azure-ad-and-create-a-service-principal).
6060
1. Create a secret to use for authentication as explained at [Option 3: Create a new client secret](../active-directory/develop/howto-create-service-principal-portal.md#option-3-create-a-new-client-secret).
6161
1. Take note of the client secret **Value** that is generated. This value is only displayed once.
6262
1. Take note of the `client ID` and the `tenant id` in the **Overview** pane of the application.
@@ -134,7 +134,7 @@ To create this pipeline in your existing Azure Data Factory and invoke batch end
134134

135135
1. Select **Pipeline** > **Import from pipeline template**
136136

137-
1. You're prompted to select a `zip` file. Uses [the following template if using managed identities](https://azuremlexampledata.blob.core.windows.net/data/templates/batch-inference/Run-BatchEndpoint-MI.zip) or [the following one if using a service principal](https://azuremlexampledata.blob.core.windows.net/data/templates/batch-inference/Run-BatchEndpoint-SP.zip).
137+
1. You're prompted to select a `zip` file. Uses [this file if using managed identities](https://azuremlexampledata.blob.core.windows.net/data/templates/batch-inference/Run-BatchEndpoint-MI.zip) or [this file if using a service principal](https://azuremlexampledata.blob.core.windows.net/data/templates/batch-inference/Run-BatchEndpoint-SP.zip).
138138

139139
1. A preview of the pipeline shows up in the portal. Select **Use this template**.
140140

@@ -170,7 +170,7 @@ When you call Azure Machine Learning batch deployments, consider the following l
170170

171171
- Only Azure Machine Learning data stores or Azure Storage Accounts (Azure Blob Storage, Azure Data Lake Storage Gen1, Azure Data Lake Storage Gen2) are supported as inputs. If your input data is in another source, use the Azure Data Factory Copy activity before the execution of the batch job to sink the data to a compatible store.
172172
- Batch endpoint jobs don't explore nested folders and hence can't work with nested folder structures. If your data is distributed in multiple folders, you have to flatten the structure.
173-
- Make sure that your scoring script provided in the deployment can handle the data as it is expected to be fed into the job. If the model is MLflow, read the limitation in terms of the file type supported by the moment at [Using MLflow models in batch deployments](how-to-mlflow-batch.md).
173+
- Make sure that your scoring script provided in the deployment can handle the data as it is expected to be fed into the job. If the model is MLflow, for the limitations on supported file types, see [Deploy MLflow models in batch deployments](how-to-mlflow-batch.md).
174174

175175
### Data outputs
176176

@@ -179,6 +179,6 @@ When you call Azure Machine Learning batch deployments, consider the following l
179179

180180
## Related content
181181

182-
- [Use low priority VMs in batch deployments](how-to-use-low-priority-batch.md)
182+
- [Using low priority VMs in batch deployments](how-to-use-low-priority-batch.md)
183183
- [Authorization on batch endpoints](how-to-authenticate-batch-endpoint.md)
184184
- [Network isolation in batch endpoints](how-to-secure-batch-endpoint.md)

0 commit comments

Comments
 (0)