You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/machine-learning/how-to-use-batch-azure-data-factory.md
+5-2Lines changed: 5 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -21,6 +21,9 @@ Big data requires a service that can orchestrate and operationalize processes to
21
21
22
22
Azure Data Factory allows the creation of pipelines that can orchestrate multiple data transformations and manage them as a single unit. Batch endpoints are an excellent candidate to become a step in such processing workflow. In this example, learn how to use batch endpoints in Azure Data Factory activities by relying on the Web Invoke activity and the REST API.
23
23
24
+
> [!TIP]
25
+
> When using data pipelines in Fabric, you can invoke batch endpoint directly using the Azure Machine Learning activity. We recommend using Fabric for data orchestration whenever possible to take advantage of the newest capabilities. The Azure Machine Learning activity in Azure Data Factory can only work with assets from Azure Machine Learning V1. Learn more at [Run Azure Machine Learning models from Fabric, using batch endpoints (preview)](how-to-use-batch-fabric.md).
26
+
24
27
## Prerequisites
25
28
26
29
* This example assumes that you have a model correctly deployed as a batch endpoint. Particularly, we are using the *heart condition classifier* created in the tutorial [Using MLflow models in batch deployments](how-to-mlflow-batch.md).
@@ -86,7 +89,7 @@ The pipeline requires the following parameters to be configured:
|`endpoint_uri`| The endpoint scoring URI |`https://<endpoint_name>.<region>.inference.ml.azure.com/jobs`|
88
91
|`poll_interval`| The number of seconds to wait before checking the job status for completion. Defaults to `120`. |`120`|
89
-
|`endpoint_input_uri`| The endpoint's input data. Multiple data input types are supported. Ensure that the manage identity you are using for executing the job has access to the underlying location. Alternative, if using Data Stores, ensure the credentials are indicated there. |`azureml://datastores/.../paths/.../data/`|
92
+
|`endpoint_input_uri`| The endpoint's input data. Multiple data input types are supported. Ensure that the managed identity you are using for executing the job has access to the underlying location. Alternative, if using Data Stores, ensure the credentials are indicated there. |`azureml://datastores/.../paths/.../data/`|
90
93
|`endpoint_input_type`| The type of the input data you are providing. Currently batch endpoints support folders (`UriFolder`) and File (`UriFile`). Defaults to `UriFolder`. |`UriFolder`|
91
94
|`endpoint_output_uri`| The endpoint's output data file. It must be a path to an output file in a Data Store attached to the Machine Learning workspace. Not other type of URIs is supported. You can use the default Azure Machine Learning data store, named `workspaceblobstore`. |`azureml://datastores/workspaceblobstore/paths/batch/predictions.csv`|
92
95
@@ -111,7 +114,7 @@ The pipeline requires the following parameters to be configured:
111
114
|`client_secret`| The client secret of the service principal used to invoke the endpoint |`ABCDEFGhijkLMNOPQRstUVwz`|
112
115
|`endpoint_uri`| The endpoint scoring URI |`https://<endpoint_name>.<region>.inference.ml.azure.com/jobs`|
113
116
|`poll_interval`| The number of seconds to wait before checking the job status for completion. Defaults to `120`. |`120`|
114
-
|`endpoint_input_uri`| The endpoint's input data. Multiple data input types are supported. Ensure that the manage identity you are using for executing the job has access to the underlying location. Alternative, if using Data Stores, ensure the credentials are indicated there. |`azureml://datastores/.../paths/.../data/`|
117
+
|`endpoint_input_uri`| The endpoint's input data. Multiple data input types are supported. Ensure that the managed identity you are using for executing the job has access to the underlying location. Alternative, if using Data Stores, ensure the credentials are indicated there. |`azureml://datastores/.../paths/.../data/`|
115
118
|`endpoint_input_type`| The type of the input data you are providing. Currently batch endpoints support folders (`UriFolder`) and File (`UriFile`). Defaults to `UriFolder`. |`UriFolder`|
116
119
|`endpoint_output_uri`| The endpoint's output data file. It must be a path to an output file in a Data Store attached to the Machine Learning workspace. Not other type of URIs is supported. You can use the default Azure Machine Learning data store, named `workspaceblobstore`. |`azureml://datastores/workspaceblobstore/paths/batch/predictions.csv`|
0 commit comments