Skip to content

Commit 67ec16d

Browse files
committed
PR review fixes
1 parent 98d5dcf commit 67ec16d

File tree

3 files changed

+17
-17
lines changed

3 files changed

+17
-17
lines changed

articles/data-factory/how-to-data-flow-error-rows.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ There are two primary methods to graceful handle errors when writing data to you
2020
* Alternatively, use the following steps to provide logging of columns that don't fit into a target string column, allowing your data flow to continue.
2121

2222
> [!NOTE]
23-
> When enabling automatic error row handling, as opposed to the following method of writing your own error handling logic, there will be a small performance penalty incurred by and additional step taken by ADF to perform a 2-phase operation to trap errors.
23+
> When enabling automatic error row handling, as opposed to the following method of writing your own error handling logic, there will be a small performance penalty incurred by and additional step taken by the data factory to perform a 2-phase operation to trap errors.
2424
2525
## Scenario
2626

articles/data-factory/how-to-sqldb-to-cosmosdb.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -14,9 +14,9 @@ This guide explains how to take an existing normalized database schema in Azure
1414

1515
SQL schemas are typically modeled using third normal form, resulting in normalized schemas that provide high levels of data integrity and fewer duplicate data values. Queries can join entities together across tables for reading. Azure Cosmos DB is optimized for super-quick transactions and querying within a collection or container via denormalized schemas with data self-contained inside a document.
1616

17-
Using Azure Data Factory, we build a pipeline that uses a single Mapping Data Flow to read from two Azure SQL Database normalized tables that contain primary and foreign keys as the entity relationship. ADF will join those tables into a single stream using the data flow Spark engine, collect joined rows into arrays and produce individual cleansed documents for insert into a new Azure Cosmos DB container.
17+
Using Azure Data Factory, we build a pipeline that uses a single Mapping Data Flow to read from two Azure SQL Database normalized tables that contain primary and foreign keys as the entity relationship. Data factory will join those tables into a single stream using the data flow Spark engine, collect joined rows into arrays and produce individual cleansed documents for insert into a new Azure Cosmos DB container.
1818

19-
This guide builds a new container on the fly called "orders" that will use the ```SalesOrderHeader``` and ```SalesOrderDetail``` tables from the standard SQL Server [Adventure Works sample database](/sql/samples/adventureworks-install-configure?tabs=ssms). Those tables represent sales transactions joined by ```SalesOrderID```. Each unique detail records have its own primary key of ```SalesOrderDetailID```. The relationship between header and detail is ```1:M```. We join on ```SalesOrderID``` in ADF and then roll each related detail record into an array called "detail".
19+
This guide builds a new container on the fly called "orders" that will use the ```SalesOrderHeader``` and ```SalesOrderDetail``` tables from the standard SQL Server [Adventure Works sample database](/sql/samples/adventureworks-install-configure?tabs=ssms). Those tables represent sales transactions joined by ```SalesOrderID```. Each unique detail record has its own primary key of ```SalesOrderDetailID```. The relationship between header and detail is ```1:M```. We join on ```SalesOrderID``` in ADF and then roll each related detail record into an array called "detail".
2020

2121
The representative SQL query for this guide is:
2222

@@ -33,7 +33,7 @@ The representative SQL query for this guide is:
3333
FROM SalesLT.SalesOrderHeader o;
3434
```
3535

36-
The resulting Azure Cosmos DB container embeds the inner query into a single document and look like this:
36+
The resulting Azure Cosmos DB container embeds the inner query into a single document and looks like this:
3737

3838
:::image type="content" source="media/data-flow/cosmosb3.png" alt-text="Collection":::
3939

@@ -45,7 +45,7 @@ The resulting Azure Cosmos DB container embeds the inner query into a single doc
4545

4646
3. In the data flow activity, select **New mapping data flow**.
4747

48-
4. We construct this data flow graph below
48+
4. We construct this data flow graph:
4949

5050
:::image type="content" source="media/data-flow/cosmosb1.png" alt-text="Data Flow Graph":::
5151

@@ -61,7 +61,7 @@ The resulting Azure Cosmos DB container embeds the inner query into a single doc
6161

6262
9. Now, let's go to the sales header source. Add a Join transformation. For the right-side select "MakeStruct". Leave it set to inner join and choose ```SalesOrderID``` for both sides of the join condition.
6363

64-
10. Select on the Data Preview tab in the new join that you added so that you can see your results up to this point. You should see all of the header rows joined with the detail rows. This is the result of the join being formed from the ```SalesOrderID```. Next, we combine the details from the common rows into the details struct and aggregate the common rows.
64+
10. Select the Data Preview tab in the new join that you added so that you can see your results up to this point. You should see all of the header rows joined with the detail rows. This is the result of the join being formed from the ```SalesOrderID```. Next, we combine the details from the common rows into the details struct and aggregate the common rows.
6565

6666
:::image type="content" source="media/data-flow/cosmosb4.png" alt-text="Join":::
6767

@@ -91,7 +91,7 @@ The resulting Azure Cosmos DB container embeds the inner query into a single doc
9191

9292
:::image type="content" source="media/data-flow/cosmosb7.png" alt-text="Screenshot shows the Mapping tab.":::
9393

94-
20. Select on data preview to make sure that you're seeing these 32 rows set to insert as new documents into your new container:
94+
20. Select data preview to make sure that you're seeing these 32 rows set to insert as new documents into your new container:
9595

9696
:::image type="content" source="media/data-flow/cosmosb8.png" alt-text="Screenshot shows the Data preview tab.":::
9797

articles/data-factory/pipeline-trigger-troubleshoot-guide.md

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -21,11 +21,11 @@ Pipeline runs are typically instantiated by passing arguments to parameters that
2121

2222
### An Azure Functions app pipeline throws an error with private endpoint connectivity
2323

24-
You have Data Factory and a function app running on a private endpoint in Azure. You're trying to run a pipeline that interacts with the function app. You've tried three different methods, but one returns error "Bad Request," and the other two methods return "103 Error Forbidden."
24+
You have a data factory and a function app running on a private endpoint in Azure. You're trying to run a pipeline that interacts with the function app. You've tried three different methods, but one returns error "Bad Request," and the other two methods return "103 Error Forbidden."
2525

2626
**Cause**
2727

28-
Data Factory currently doesn't support a private endpoint connector for function apps. Azure Functions rejects calls because it's configured to allow only connections from a private link.
28+
Azure Data Factory currently doesn't support a private endpoint connector for function apps. Azure Functions rejects calls because it's configured to allow only connections from a private link.
2929

3030
**Resolution**
3131

@@ -45,7 +45,7 @@ Refresh the browser and apply the correct monitoring filters.
4545

4646
**Cause**
4747

48-
If a folder you're copying contains files with different schemas, such as variable number of columns, different delimiters, quote char settings, or some data issue, the Data Factory pipeline might throw this error:
48+
If a folder you're copying contains files with different schemas, such as variable number of columns, different delimiters, quote char settings, or some data issue, the pipeline might throw this error:
4949

5050
`
5151
Operation on target Copy_sks failed: Failure happened on 'Sink' side.
@@ -57,7 +57,7 @@ Source=Microsoft.DataTransfer.Common,'
5757

5858
**Resolution**
5959

60-
Select the **Binary Copy** option while creating the Copy activity. This way, for bulk copies or migrating your data from one data lake to another, Data Factory won't open the files to read the schema. Instead, Data Factory treats each file as binary and copy it to the other location.
60+
Select the **Binary Copy** option while creating the Copy activity. This way, for bulk copies or migrating your data from one data lake to another, Data Factory won't open the files to read the schema. Instead, Azure Data Factory treats each file as binary and copies it to the other location.
6161

6262
### A pipeline run fails when you reach the capacity limit of the integration runtime for data flow
6363

@@ -115,7 +115,7 @@ Azure Data Factory evaluates the outcome of all leaf-level activities. Pipeline
115115

116116
**Cause**
117117

118-
You might need to monitor failed Data Factory pipelines in intervals, say 5 minutes. You can query and filter the pipeline runs from a data factory by using the endpoint.
118+
You might need to monitor failed Azure Data Factory pipelines in intervals, say 5 minutes. You can query and filter the pipeline runs from a data factory by using the endpoint.
119119

120120
**Resolution**
121121
* You can set up an Azure logic app to query all of the failed pipelines every 5 minutes, as described in [Query By Factory](/rest/api/datafactory/pipelineruns/querybyfactory). Then, you can report incidents to your ticketing system.
@@ -152,7 +152,7 @@ Known Facts about *ForEach*
152152
**Resolution**
153153

154154
* **Concurrency Limit:** If your pipeline has a concurrency policy, verify that there are no old pipeline runs in progress.
155-
* **Monitoring limits**: Go to the ADF authoring canvas, select your pipeline, and determine if it has a concurrency property assigned to it. If it does, go to the Monitoring view, and make sure there's nothing in the past 45 days that's in progress. If there's something in progress, you can cancel it and the new pipeline run should start.
155+
* **Monitoring limits**: Go to the authoring canvas, select your pipeline, and determine if it has a concurrency property assigned to it. If it does, go to the Monitoring view, and make sure there's nothing in the past 45 days that's in progress. If there's something in progress, you can cancel it and the new pipeline run should start.
156156

157157
* **Transient Issues:** It's possible that your run was impacted by a transient network issue, credential failures, services outages etc. If this happens, Azure Data Factory has an internal recovery process that monitors all the runs and starts them when it notices something went wrong. You can rerun pipelines and activities as described [here.](monitor-visually.md#rerun-pipelines-and-activities). You can rerun activities if you had canceled activity or had a failure as per [Rerun from activity failures.](monitor-visually.md#rerun-from-failed-activity) This process happens every one hour, so if your run is stuck for more than an hour, create a support case.
158158

@@ -199,7 +199,7 @@ It's a user error because JSON payload that hits management.azure.com is corrupt
199199

200200
**Resolution**
201201

202-
Perform network tracing of your API call from ADF portal using Microsoft Edge/Chrome browser **Developer tools**. You'll see offending JSON payload, which could be due to a special character(for example $), spaces, and other types of user input. Once you fix the string expression, you'll proceed with rest of ADF usage calls in the browser.
202+
Perform network tracing of your API call from ADF portal using Microsoft Edge/Chrome browser **Developer tools**. You'll see offending JSON payload, which could be due to a special character (for example, ```$```), spaces, and other types of user input. Once you fix the string expression, you'll proceed with rest of ADF usage calls in the browser.
203203

204204
### ForEach activities don't run in parallel mode
205205

@@ -272,8 +272,8 @@ Input **execute pipeline** activity for pipeline parameter as *@createArray(
272272

273273
For more troubleshooting help, try these resources:
274274

275-
* [Data Factory blog](https://techcommunity.microsoft.com/t5/azure-data-factory-blog/bg-p/AzureDataFactoryBlog)
276-
* [Data Factory feature requests](/answers/topics/azure-data-factory.html)
275+
* [Azure Data Factory blog](https://techcommunity.microsoft.com/t5/azure-data-factory-blog/bg-p/AzureDataFactoryBlog)
276+
* [Azure Data Factory feature requests](/answers/topics/azure-data-factory.html)
277277
* [Azure videos](https://azure.microsoft.com/resources/videos/index/?sort=newest&services=data-factory)
278278
* [Microsoft Q&A question page](/answers/topics/azure-data-factory.html)
279-
* [X information about Data Factory](https://x.com/hashtag/DataFactory)
279+
* [X information about Azure Data Factory](https://x.com/hashtag/DataFactory)

0 commit comments

Comments
 (0)