Skip to content

Commit 3071117

Browse files
Merge pull request #281385 from santiagxf/santiagxf-patch-1
Update how-to-use-batch-azure-data-factory.md
2 parents 934fdaa + 0630ab5 commit 3071117

File tree

1 file changed

+5
-2
lines changed

1 file changed

+5
-2
lines changed

articles/machine-learning/how-to-use-batch-azure-data-factory.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,6 +21,9 @@ Big data requires a service that can orchestrate and operationalize processes to
2121

2222
Azure Data Factory allows the creation of pipelines that can orchestrate multiple data transformations and manage them as a single unit. Batch endpoints are an excellent candidate to become a step in such processing workflow. In this example, learn how to use batch endpoints in Azure Data Factory activities by relying on the Web Invoke activity and the REST API.
2323

24+
> [!TIP]
25+
> When using data pipelines in Fabric, you can invoke batch endpoint directly using the Azure Machine Learning activity. We recommend using Fabric for data orchestration whenever possible to take advantage of the newest capabilities. The Azure Machine Learning activity in Azure Data Factory can only work with assets from Azure Machine Learning V1. Learn more at [Run Azure Machine Learning models from Fabric, using batch endpoints (preview)](how-to-use-batch-fabric.md).
26+
2427
## Prerequisites
2528

2629
* This example assumes that you have a model correctly deployed as a batch endpoint. Particularly, we are using the *heart condition classifier* created in the tutorial [Using MLflow models in batch deployments](how-to-mlflow-batch.md).
@@ -86,7 +89,7 @@ The pipeline requires the following parameters to be configured:
8689
| --------------------- | -------------|------------- |
8790
| `endpoint_uri` | The endpoint scoring URI | `https://<endpoint_name>.<region>.inference.ml.azure.com/jobs` |
8891
| `poll_interval` | The number of seconds to wait before checking the job status for completion. Defaults to `120`. | `120` |
89-
| `endpoint_input_uri` | The endpoint's input data. Multiple data input types are supported. Ensure that the manage identity you are using for executing the job has access to the underlying location. Alternative, if using Data Stores, ensure the credentials are indicated there. | `azureml://datastores/.../paths/.../data/` |
92+
| `endpoint_input_uri` | The endpoint's input data. Multiple data input types are supported. Ensure that the managed identity you are using for executing the job has access to the underlying location. Alternative, if using Data Stores, ensure the credentials are indicated there. | `azureml://datastores/.../paths/.../data/` |
9093
| `endpoint_input_type` | The type of the input data you are providing. Currently batch endpoints support folders (`UriFolder`) and File (`UriFile`). Defaults to `UriFolder`. | `UriFolder` |
9194
| `endpoint_output_uri` | The endpoint's output data file. It must be a path to an output file in a Data Store attached to the Machine Learning workspace. Not other type of URIs is supported. You can use the default Azure Machine Learning data store, named `workspaceblobstore`. | `azureml://datastores/workspaceblobstore/paths/batch/predictions.csv` |
9295

@@ -111,7 +114,7 @@ The pipeline requires the following parameters to be configured:
111114
| `client_secret` | The client secret of the service principal used to invoke the endpoint | `ABCDEFGhijkLMNOPQRstUVwz` |
112115
| `endpoint_uri` | The endpoint scoring URI | `https://<endpoint_name>.<region>.inference.ml.azure.com/jobs` |
113116
| `poll_interval` | The number of seconds to wait before checking the job status for completion. Defaults to `120`. | `120` |
114-
| `endpoint_input_uri` | The endpoint's input data. Multiple data input types are supported. Ensure that the manage identity you are using for executing the job has access to the underlying location. Alternative, if using Data Stores, ensure the credentials are indicated there. | `azureml://datastores/.../paths/.../data/` |
117+
| `endpoint_input_uri` | The endpoint's input data. Multiple data input types are supported. Ensure that the managed identity you are using for executing the job has access to the underlying location. Alternative, if using Data Stores, ensure the credentials are indicated there. | `azureml://datastores/.../paths/.../data/` |
115118
| `endpoint_input_type` | The type of the input data you are providing. Currently batch endpoints support folders (`UriFolder`) and File (`UriFile`). Defaults to `UriFolder`. | `UriFolder` |
116119
| `endpoint_output_uri` | The endpoint's output data file. It must be a path to an output file in a Data Store attached to the Machine Learning workspace. Not other type of URIs is supported. You can use the default Azure Machine Learning data store, named `workspaceblobstore`. | `azureml://datastores/workspaceblobstore/paths/batch/predictions.csv` |
117120

0 commit comments

Comments
 (0)