Skip to content

Commit 25b5749

Browse files
authored
Create tutorial-data-flow-sql-server.md
1 parent a32c61b commit 25b5749

File tree

1 file changed

+70
-0
lines changed

1 file changed

+70
-0
lines changed
Lines changed: 70 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,70 @@
1+
---
2+
title: Ingest and transform data from SQL Server to Azure SQL Database | Microsoft Docs
3+
description: This tutorial provides step-by-step instructions for transforming data coming from on-prem SQL Server using ADF with Mapping Data Flows.
4+
services: data-factory
5+
author: kromerm
6+
ms.service: data-factory
7+
ms.workload: data-services
8+
ms.date: 05/23/2019
9+
ms.author: makromer
10+
---
11+
12+
# Transform data from on-prem SQL Server and send it to Azure SQL Database
13+
14+
Azure Data Factory is a cloud-based data integration and ETL service that allows you to create data-driven workflows in the cloud for orchestrating and automating data movement and data transformation. Using Azure Data Factory, you can create and schedule data-driven workflows (called pipelines) that can ingest data from disparate data stores, process/transform the data at scale with a code-free interface.
15+
16+
In this tutorial, you create a Data Factory pipeline that copies data from SQL Server on-prem to Azure Blob for transformation using Mapping Data Flows. You'll then land the transformed data in Azure SQL Database.
17+
18+
You perform the following steps in this tutorial:
19+
20+
> [!div class="checklist"]
21+
> * Create a data factory with the incremental data loading pattern.
22+
> * Create Azure Storage and Azure SQL Database linked services.
23+
> * Create Azure Blob and Azure SQL Database datasets.
24+
> * Add a data flow activity to the incremental pipeline.
25+
> * Start a pipeline run.
26+
27+
This tutorial assumes that you've already gone through the getting stared quickstarts and have performed the actions in the [Incremental data loading with SQL change tracking](https://docs.microsoft.com/en-us/azure/data-factory/tutorial-incremental-copy-change-tracking-feature-portal) tutorial.
28+
29+
If you don't have an Azure subscription, create a [free](https://azure.microsoft.com/free/) account before you begin.
30+
31+
## Prerequisites
32+
33+
* **Azure Storage account**. You use the blob storage as **sink** data store in the copy activity as part of the incremental data load to stage the SQL Server data from an on-prem source. If you don't have an Azure storage account, see the [Create a storage account](../storage/common/storage-create-storage-account.md#create-a-storage-account) article for steps to create one.
34+
* **Azure SQL Database**. You use the database as **sink** data store. If you don't have an Azure SQL Database, see the [Create an Azure SQL database](../sql-database/sql-database-get-started-portal.md) article for steps to create one.
35+
36+
## Build the incremental pipeline with SQL Server source
37+
38+
![SQL Delta Load](media/data-flow/sql1.png "Delta SQL")
39+
40+
* Follow the [Incremental data loading with SQL change tracking](tutorial-incremental-copy-change-tracking-feature-portal.md) tutorial.
41+
* Change the source linked service from Azure SQL Database to SQL Server.
42+
* Change the linked service for the stored procedure activity to call the stored procedure on your local SQL Server. This stored procedure updates the version control table in SQL.
43+
* Use the [self-hosted integration runtime](create-self-hosted-integration-runtime.md) to connect to the on-premises SQL Server.
44+
* Keep the rest of the pipeline the same. You'll stage the changed data from your SQL Server in Azure Blob Store and then use that staged data with Data Flows for visual data transformation.
45+
46+
## Build the data flow to transform the data
47+
48+
![Data flow tutorial](media/data-flow/dataflowtutorial1.png "Data flow tutorial")
49+
50+
* Click new (+) Data Flow
51+
* Start by clicking on "Add Source" in the design surface
52+
* In the source settings, choose the dataset that you used for the Blob Store in the incremental pipeline above.
53+
*
54+
55+
## Next steps
56+
The pipeline in this sample copies data from one location to another location in an Azure blob storage and transforms data in ADF Mapping Data Flows code-free. You learned how to:
57+
58+
> [!div class="checklist"]
59+
> * Create a data factory.
60+
> * Create Azure Storage and Azure SQL Database linked services.
61+
> * Create Azure Blob and Azure SQL Database datasets.
62+
> * Design a Mapping Data Flow.
63+
> * Create a pipeline contains a Data Flow activity.
64+
> * Start a pipeline run.
65+
66+
67+
Advance to the following tutorial to learn about copying data from on-premises to cloud:
68+
69+
> [!div class="nextstepaction"]
70+
>[Copy data from on-premises to cloud](tutorial-hybrid-copy-powershell.md)

0 commit comments

Comments
 (0)