Skip to content

Commit f35fd77

Browse files
authored
Merge pull request #94505 from djpmsft/docUpdates
Adding Wrangling Data Flow documentation
2 parents 39c4459 + ffd8544 commit f35fd77

13 files changed

+349
-36
lines changed

articles/data-factory/TOC.yml

Lines changed: 35 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -80,8 +80,12 @@
8080
items:
8181
- name: Copy Data tool
8282
href: tutorial-incremental-copy-partitioned-file-name-copy-data-tool.md
83-
- name: Transform data with mapping data flow
84-
href: tutorial-data-flow.md
83+
- name: Transform and prepare data with data flows
84+
items:
85+
- name: Transform data with a mapping data flow
86+
href: tutorial-data-flow.md
87+
- name: Prepare data with a wrangling data flow
88+
href: wrangling-data-flow-tutorial.md
8589
- name: External transformations
8690
items:
8791
- name: HDInsight Spark
@@ -143,28 +147,36 @@
143147
href: concepts-pipeline-execution-triggers.md
144148
- name: Integration runtime
145149
href: concepts-integration-runtime.md
146-
- name: Mapping data flows
150+
- name: Data flows
147151
items:
148-
- name: Mapping data flow overview
149-
href: concepts-data-flow-overview.md
150-
- name: Debug mode
151-
href: concepts-data-flow-debug-mode.md
152-
- name: JSON handling
153-
href: concepts-data-flow-json.md
154-
- name: Schema drift
155-
href: concepts-data-flow-schema-drift.md
156-
- name: Column patterns
157-
href: concepts-data-flow-column-pattern.md
158-
- name: Data flow monitoring
159-
href: concepts-data-flow-monitoring.md
160-
- name: Data flow performance
161-
href: concepts-data-flow-performance.md
162-
- name: Move nodes
163-
href: concepts-data-flow-move-nodes.md
164-
- name: Expression builder
165-
href: concepts-data-flow-expression-builder.md
166-
- name: Expression language
167-
href: data-flow-expression-functions.md
152+
- name: Transform data with mapping data flows
153+
items:
154+
- name: Mapping data flow overview
155+
href: concepts-data-flow-overview.md
156+
- name: Debug mode
157+
href: concepts-data-flow-debug-mode.md
158+
- name: JSON handling
159+
href: concepts-data-flow-json.md
160+
- name: Schema drift
161+
href: concepts-data-flow-schema-drift.md
162+
- name: Column patterns
163+
href: concepts-data-flow-column-pattern.md
164+
- name: Data flow monitoring
165+
href: concepts-data-flow-monitoring.md
166+
- name: Data flow performance
167+
href: concepts-data-flow-performance.md
168+
- name: Move nodes
169+
href: concepts-data-flow-move-nodes.md
170+
- name: Expression builder
171+
href: concepts-data-flow-expression-builder.md
172+
- name: Expression language
173+
href: data-flow-expression-functions.md
174+
- name: Prepare data with Wrangling data flows
175+
items:
176+
- name: Wrangling data flow overview
177+
href: wrangling-data-flow-overview.md
178+
- name: Supported functions
179+
href: wrangling-data-flow-functions.md
168180
- name: Roles and permissions
169181
href: concepts-roles-permissions.md
170182
- name: Understanding pricing

articles/data-factory/frequently-asked-questions.md

Lines changed: 76 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -187,6 +187,82 @@ Use the Copy activity to stage data from any of the other connectors, and then e
187187

188188
Self-hosted IR is an ADF pipeline construct that you can use with the Copy Activity to acquire or move data to and from on-prem or VM-based data sources and sinks. Stage the data first with a Copy, then Data Flow for transformation, and then a subsequent copy if you need to move that transformed data back to the on-prem store.
189189

190+
## Wrangling data flows
191+
192+
### What are the supported regions for wrangling data flow?
193+
194+
Wrangling data flow is currently supported in data factories created in following regions:
195+
196+
* Australia East
197+
* Canada Central
198+
* Central India
199+
* Central US
200+
* East US
201+
* East US 2
202+
* Japan East
203+
* North Europe
204+
* Southeast Asia
205+
* South Central US
206+
* UK South
207+
* West Central US
208+
* West Europe
209+
* West US
210+
* West US 2
211+
212+
### What are the limitations and constraints with wrangling data flow?
213+
214+
Dataset names can only contain alpha-numeric characters. The following data stores are supported:
215+
216+
* DelimitedText dataset in Azure Blob Storage using account key authentication
217+
* DelimitedText dataset in Azure Data Lake Storage gen2 using account key or service principal authentication
218+
* DelimitedText dataset in Azure Data Lake Storage gen1 using service principal authentication
219+
* Azure SQL Database and Data Warehouse using sql authentication. See supported SQL types below. There is no PolyBase or staging support for data warehouse.
220+
221+
At this time, linked service Key Vault integration is not supported in wrangling data flows.
222+
223+
### What is the difference between mapping and wrangling data flows?
224+
225+
Mapping data flows provide a way to transform data at scale without any coding required. You can design a data transformation job in the data flow canvas by constructing a series of transformations. Start with any number of source transformations followed by data transformation steps. Complete your data flow with a sink to land your results in a destination. Mapping data flow is great at mapping and transforming data with both known and unknown schemas in the sinks and sources.
226+
227+
Wrangling data flows allow you to do agile data preparation and exploration using the Power Query Online mashup editor at scale via spark execution. With the rise of data lakes sometimes you just need to explore a data set or create a dataset in the lake. You aren't mapping to a known target. Wrangling data flows are used for less formal and model-based analytics scenarios.
228+
229+
### What is the difference between Power Platform Dataflows and wrangling data flows?
230+
231+
Power Platform Dataflows allow users to import and transform data from a wide range of data sources into the Common Data Service and Azure Data Lake to build PowerApps applications, Power BI reports or Flow automations. Power Platform Dataflows use the established Power Query data preparation experiences, similar to Power BI and Excel. Power Platform Dataflows also enable easy reuse within an organization and automatically handle orchestration (e.g. automatically refreshing dataflows that depend on another dataflow when the former one is refreshed).
232+
233+
Azure Data Factory (ADF) is a managed data integration service that allows data engineers and citizen data integrator to create complex hybrid extract-transform-load (ETL) and extract-load-transform (ELT) workflows. Wrangling data flow in ADF empowers users with a code-free, serverless environment that simplifies data preparation in the cloud and scales to any data size with no infrastructure management required. It uses the Power Query data preparation technology (also used in Power Platform dataflows, Excel, Power BI) to prepare and shape the data. Built to handle all the complexities and scale challenges of big data integration, wrangling data flows allow users to quickly prepare data at scale via spark execution. Users can build resilient data pipelines in an accessible visual environment with our browser-based interface and let ADF handle the complexities of Spark execution. Build schedules for your pipelines and monitor your data flow executions from the ADF monitoring portal. Easily manage data availability SLAs with ADF’s rich availability monitoring and alerts and leverage built-in continuous integration and deployment capabilities to save and manage your flows in a managed environment. Establish alerts and view execution plans to validate that your logic is performing as planned as you tune your data flows.
234+
235+
### Supported SQL Types
236+
237+
Wrangling data flow supports the following data types in SQL. You will get a validation error for using a data type that isn't supported.
238+
239+
* short
240+
* double
241+
* real
242+
* float
243+
* char
244+
* nchar
245+
* varchar
246+
* nvarchar
247+
* integer
248+
* int
249+
* bit
250+
* boolean
251+
* smallint
252+
* tinyint
253+
* bigint
254+
* long
255+
* text
256+
* date
257+
* datetime
258+
* datetime2
259+
* smalldatetime
260+
* timestamp
261+
* uniqueidentifier
262+
* xml
263+
264+
Other data types will be supported in the future.
265+
190266
## Next steps
191267
For step-by-step instructions to create a data factory, see the following tutorials:
192268

271 KB
Loading
153 KB
Loading
276 KB
Loading
233 KB
Loading
276 KB
Loading
598 KB
Loading
337 KB
Loading

articles/data-factory/transform-data.md

Lines changed: 25 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -32,52 +32,64 @@ This article explains data transformation activities in Azure Data Factory that
3232

3333
Data Factory supports the following data transformation activities that can be added to [pipelines](concepts-pipelines-activities.md) either individually or chained with another activity.
3434

35-
## HDInsight Hive activity
35+
## Transform natively in Azure Data Factory with data flows
36+
37+
### Mapping data flows
38+
39+
Mapping data flows are visually designed data transformations in Azure Data Factory. Data flows allow data engineers to develop graphical data transformation logic without writing code. The resulting data flows are executed as activities within Azure Data Factory pipelines that use scaled-out Spark clusters. Data flow activities can be operationalized via existing Data Factory scheduling, control, flow, and monitoring capabilities. For more information, see [mapping data flows](concepts-data-flow-overview.md).
40+
41+
### Wrangling data flows
42+
43+
Wrangling data flows in Azure Data Factory allow you to do code-free data preparation at cloud scale iteratively. Wrangling data flows integrate with [Power Query Online](https://docs.microsoft.com/power-query/) and makes Power Query M functions available for data wrangling at cloud scale via spark execution. For more information, see [wrangling data flows](wrangling-data-flow-overview.md).
44+
45+
## External transformations
46+
47+
### HDInsight Hive activity
3648
The HDInsight Hive activity in a Data Factory pipeline executes Hive queries on your own or on-demand Windows/Linux-based HDInsight cluster. See [Hive activity](transform-data-using-hadoop-hive.md) article for details about this activity.
3749

38-
## HDInsight Pig activity
50+
### HDInsight Pig activity
3951
The HDInsight Pig activity in a Data Factory pipeline executes Pig queries on your own or on-demand Windows/Linux-based HDInsight cluster. See [Pig activity](transform-data-using-hadoop-pig.md) article for details about this activity.
4052

41-
## HDInsight MapReduce activity
53+
### HDInsight MapReduce activity
4254
The HDInsight MapReduce activity in a Data Factory pipeline executes MapReduce programs on your own or on-demand Windows/Linux-based HDInsight cluster. See [MapReduce activity](transform-data-using-hadoop-map-reduce.md) article for details about this activity.
4355

44-
## HDInsight Streaming activity
56+
### HDInsight Streaming activity
4557
The HDInsight Streaming activity in a Data Factory pipeline executes Hadoop Streaming programs on your own or on-demand Windows/Linux-based HDInsight cluster. See [HDInsight Streaming activity](transform-data-using-hadoop-streaming.md) for details about this activity.
4658

47-
## HDInsight Spark activity
59+
### HDInsight Spark activity
4860
The HDInsight Spark activity in a Data Factory pipeline executes Spark programs on your own HDInsight cluster. For details, see [Invoke Spark programs from Azure Data Factory](transform-data-using-spark.md).
4961

50-
## Machine Learning activities
62+
### Machine Learning activities
5163
Azure Data Factory enables you to easily create pipelines that use a published Azure Machine Learning web service for predictive analytics. Using the [Batch Execution activity](transform-data-using-machine-learning.md) in an Azure Data Factory pipeline, you can invoke a Machine Learning web service to make predictions on the data in batch.
5264

5365
Over time, the predictive models in the Machine Learning scoring experiments need to be retrained using new input datasets. After you are done with retraining, you want to update the scoring web service with the retrained Machine Learning model. You can use the [Update Resource activity](update-machine-learning-models.md) to update the web service with the newly trained model.
5466

5567
See [Use Machine Learning activities](transform-data-using-machine-learning.md) for details about these Machine Learning activities.
5668

57-
## Stored procedure activity
69+
### Stored procedure activity
5870
You can use the SQL Server Stored Procedure activity in a Data Factory pipeline to invoke a stored procedure in one of the following data stores: Azure SQL Database, Azure SQL Data Warehouse, SQL Server Database in your enterprise or an Azure VM. See [Stored Procedure activity](transform-data-using-stored-procedure.md) article for details.
5971

60-
## Data Lake Analytics U-SQL activity
72+
### Data Lake Analytics U-SQL activity
6173
Data Lake Analytics U-SQL activity runs a U-SQL script on an Azure Data Lake Analytics cluster. See [Data Analytics U-SQL activity](transform-data-using-data-lake-analytics.md) article for details.
6274

63-
## Databricks Notebook activity
75+
### Databricks Notebook activity
6476

6577
The Azure Databricks Notebook Activity in a Data Factory pipeline runs a Databricks notebook in your Azure Databricks workspace.Azure Databricks is a managed platform for running Apache Spark. See [Transform data by running a Databricks notebook](transform-data-databricks-notebook.md).
6678

67-
## Databricks Jar activity
79+
### Databricks Jar activity
6880

6981
The Azure Databricks Jar Activity in a Data Factory pipeline runs a Spark Jar in your Azure Databricks cluster. Azure Databricks is a managed platform for running Apache Spark. See [Transform data by running a Jar activity in Azure Databricks](transform-data-databricks-jar.md).
7082

71-
## Databricks Python activity
83+
### Databricks Python activity
7284

7385
The Azure Databricks Python Activity in a Data Factory pipeline runs a Python file in your Azure Databricks cluster. Azure Databricks is a managed platform for running Apache Spark. See [Transform data by running a Python activity in Azure Databricks](transform-data-databricks-python.md).
7486

75-
## Custom activity
87+
### Custom activity
7688
If you need to transform data in a way that is not supported by Data Factory, you can create a custom activity with your own data processing logic and use the activity in the pipeline. You can configure the custom .NET activity to run using either an Azure Batch service or an Azure HDInsight cluster. See [Use custom activities](transform-data-using-dotnet-custom-activity.md) article for details.
7789

7890
You can create a custom activity to run R scripts on your HDInsight cluster with R installed. See [Run R Script using Azure Data Factory](https://github.com/Azure/Azure-DataFactory/tree/master/Samples/RunRScriptUsingADFSample).
7991

80-
## Compute environments
92+
### Compute environments
8193
You create a linked service for the compute environment and then use the linked service when defining a transformation activity. There are two types of compute environments supported by Data Factory.
8294

8395
- **On-Demand**: In this case, the computing environment is fully managed by Data Factory. It is automatically created by the Data Factory service before a job is submitted to process data and removed when the job is completed. You can configure and control granular settings of the on-demand compute environment for job execution, cluster management, and bootstrapping actions.

0 commit comments

Comments
 (0)