You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In this article, you learn about the available options for building a data ingestion pipeline with [Azure Data Factory](../../data-factory/introduction.md). This Azure Data Factory pipeline is used to ingest data for use with [Azure Machine Learning](../overview-what-is-azure-machine-learning.md). Data Factory allows you to easily extract, transform, and load (ETL) data. Once the data has been transformed and loaded into storage, it can be used to train your machine learning models in Azure Machine Learning.
21
+
In this article, you learn about the available options for building a data ingestion pipeline with [Azure Data Factory](../../data-factory/introduction.md). This Azure Data Factory pipeline is used to ingest data for use with [Azure Machine Learning](../overview-what-is-azure-machine-learning.md). Data Factory allows you to easily extract, transform, and load (ETL) data. Once the data is transformed and loaded into storage, it can be used to train your machine learning models in Azure Machine Learning.
22
22
23
23
Simple data transformation can be handled with native Data Factory activities and instruments such as [data flow](../../data-factory/control-flow-execute-data-flow-activity.md). When it comes to more complicated scenarios, the data can be processed with some custom code. For example, Python or R code.
24
24
@@ -43,7 +43,7 @@ The function is invoked with the [Azure Data Factory Azure Function activity](..
43
43
44
44
* Advantages:
45
45
* The data is processed on a serverless compute with a relatively low latency
46
-
* Data Factory pipeline can invoke a [Durable Azure Function](../../azure-functions/durable/durable-functions-overview.md) that may implement a sophisticated data transformation flow
46
+
* Data Factory pipeline can invoke a [Durable Azure Function](../../azure-functions/durable/durable-functions-overview.md) that can implement a sophisticated data transformation flow
47
47
* The details of the data transformation are abstracted away by the Azure Function that can be reused and invoked from other places
48
48
* Disadvantages:
49
49
* The Azure Functions must be created before use with ADF
@@ -95,7 +95,7 @@ This method is recommended for [Machine Learning Operations (MLOps) workflows](c
95
95
Each time the Data Factory pipeline runs,
96
96
97
97
1. The data is saved to a different location in storage.
98
-
1. To pass the location to Azure Machine Learning, the Data Factory pipeline calls an [Azure Machine Learning pipeline](../concept-ml-pipelines.md). When calling the ML pipeline, the data location and job ID are sent as parameters.
98
+
1. To pass the location to Azure Machine Learning, the Data Factory pipeline calls an [Azure Machine Learning pipeline](../concept-ml-pipelines.md). When the Data Factory pipeline calls the Azure Machine Learning pipeline, the data location and job ID are sent as parameters.
99
99
1. The ML pipeline can then create an Azure Machine Learning datastore and dataset with the data location. Learn more in [Execute Azure Machine Learning pipelines in Data Factory](../../data-factory/transform-data-machine-learning-service.md).
100
100
101
101

client_id=client_id, # client id of service principal
137
137
```
138
138
139
-
Next, create a dataset to reference the file(s) you want to use in your machine learning task.
139
+
Next, create a dataset to reference the files you want to use in your machine learning task.
140
140
141
141
The following code creates a TabularDataset from a csv file, `prepared-data.csv`. Learn more about [dataset types and accepted file formats](how-to-create-register-datasets.md#dataset-types).
0 commit comments