You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/data-factory/tutorial-data-flow.md
+13-13Lines changed: 13 additions & 13 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -7,18 +7,18 @@ ms.reviewer: makromer
7
7
ms.subservice: data-flows
8
8
ms.topic: tutorial
9
9
ms.date: 04/18/2025
10
+
11
+
#customer intent: As a data factory user, I want to know the basics of creating a data flow so that I can apply these principals to my own data and begin using Data Factory for our workflows.
12
+
10
13
---
11
14
12
-
# Transform data using mapping data flows
15
+
# Tutorial: Transform data using mapping data flows
If you're new to Azure Data Factory, see [Introduction to Azure Data Factory](introduction.md).
17
-
18
-
In this tutorial, you'll use the Azure Data Factory user interface (UX) to create a pipeline that copies and transforms data from an Azure Data Lake Storage (ADLS) Gen2 source to an ADLS Gen2 sink using mapping data flow. The configuration pattern in this tutorial can be expanded upon when transforming data using mapping data flow
19
+
In this tutorial, you use the Azure Data Factory user interface (UX) to create a pipeline that copies and transforms data from an Azure Data Lake Storage (ADLS) Gen2 source to an ADLS Gen2 sink using mapping data flow. The configuration pattern in this tutorial can be expanded upon when transforming data using mapping data flow
19
20
20
-
>[!NOTE]
21
-
>This tutorial is meant for mapping data flows in general. Data flows are available both in Azure Data Factory and Synapse Pipelines. If you are new to data flows in Azure Synapse Pipelines, please follow [Data Flow using Azure Synapse Pipelines](../synapse-analytics/concepts-data-flow-overview.md)
21
+
This tutorial is meant for mapping data flows in general. Data flows are available both in Azure Data Factory and Synapse Pipelines. If you're new to data flows in Azure Synapse Pipelines, follow [Data Flow using Azure Synapse Pipelines](../synapse-analytics/concepts-data-flow-overview.md).
22
22
23
23
In this tutorial, you do the following steps:
24
24
@@ -65,11 +65,11 @@ In this step, you create a data factory and open the Data Factory UX to create a
65
65
66
66
## Create a pipeline with a Data Flow activity
67
67
68
-
In this step, you'll create a pipeline that contains a Data Flow activity.
68
+
In this step, you create a pipeline that contains a Data Flow activity.
69
69
70
70
1. On the home page of Azure Data Factory, select **Orchestrate**.
71
71
72
-
:::image type="content" source="./media/tutorial-data-flow/orchestrate.png" alt-text="Screenshot that shows the ADF home page.":::
72
+
:::image type="content" source="./media/tutorial-data-flow/orchestrate.png" alt-text="Screenshot that shows the Azure Data Factory home page.":::
73
73
74
74
1. Now a window is open for a new pipeline. In the **General** tab for the pipeline properties, enter **TransformMovies** for **Name** of the pipeline.
75
75
1. In the **Activities** pane, expand the **Move and Transform** accordion. Drag and drop the **Data Flow** activity from the pane to the pipeline canvas.
@@ -84,9 +84,9 @@ In this step, you'll create a pipeline that contains a Data Flow activity.
84
84
85
85
## Build transformation logic in the data flow canvas
86
86
87
-
In this step, you'll build a data flow that takes the moviesDB.csv in ADLS storage and aggregates the average rating of comedies from 1910 to 2000. You'll then write this file back to the ADLS storage.
87
+
In this step, you build a data flow that takes the moviesDB.csv in ADLS storage and aggregates the average rating of comedies from 1910 to 2000. You then write this file back to the ADLS storage.
88
88
89
-
1. In the panel below the canvas, go to the **Settings** of your data flow activity and select **New**, located beside data flow field. This will open the dataflow canvas.
89
+
1. In the panel below the canvas, go to the **Settings** of your data flow activity and select **New**, located beside data flow field. This opens the dataflow canvas.
90
90
91
91
:::image type="content" source="media/tutorial-data-flow/open-data-flow-from-pipeline.png" alt-text="Screenshot showing how to open the data flow editor from the pipeline editor.":::
92
92
@@ -128,7 +128,7 @@ In this step, you'll build a data flow that takes the moviesDB.csv in ADLS stora
1. Name your filter transformation **FilterYears**. Select the expression box next to **Filter on** and then **Open expression builder**. Here you'll specify your filtering condition.
131
+
1. Name your filter transformation **FilterYears**. Select the expression box next to **Filter on** and then **Open expression builder**. Here you specify your filtering condition.
132
132
133
133
:::image type="content" source="media/tutorial-data-flow/filter1.png" alt-text="Screenshot that shows the Filter on expression box.":::
134
134
@@ -142,7 +142,7 @@ In this step, you'll build a data flow that takes the moviesDB.csv in ADLS stora
If you've a debug cluster active, you can verify your logic by selecting **Refresh** to see expression output compared to the inputs used. There's more than one right answer on how you can accomplish this logic using the data flow expression language.
145
+
If you have a debug cluster active, you can verify your logic by selecting **Refresh** to see expression output compared to the inputs used. There's more than one right answer on how you can accomplish this logic using the data flow expression language.
@@ -200,7 +200,7 @@ Now you've finished building your data flow. You're ready to run it in your pipe
200
200
201
201
## Running and monitoring the Data Flow
202
202
203
-
You can debug a pipeline before you publish it. In this step, you're going to trigger a debug run of the data flow pipeline. While data preview doesn't write data, a debug run will write data to your sink destination.
203
+
You can debug a pipeline before you publish it. In this step, you're going to trigger a debug run of the data flow pipeline. While data preview doesn't write data, a debug run writes data to your sink destination.
204
204
205
205
1. Go to the pipeline canvas. Select **Debug** to trigger a debug run.
0 commit comments