Skip to content

Commit 0ba570c

Browse files
committed
Fixes for PR review
1 parent 7b436e0 commit 0ba570c

10 files changed

+40
-41
lines changed

articles/data-factory/frequently-asked-questions.yml

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ sections:
4949
- Looping containers:
5050
* The foreach activity will iterate over a specified collection of activities in a loop.
5151
- Trigger-based flows:
52-
- Pipelines can be triggered on demand, by wall-clock time, or in response to driven by event grid topics
52+
- Pipelines can be triggered on demand, by wall-clock time, or in response to driven by Event Grid topics
5353
- Delta flows:
5454
- Parameters can be used to define your high-water mark for delta copy while moving dimension or reference tables from a relational store, either on-premises or in the cloud, to load the data into the lake.
5555
@@ -234,7 +234,7 @@ sections:
234234
### I need help troubleshooting my data flow logic. What info do I need to provide to get help?
235235
236236
When Microsoft provides help or troubleshooting with data flows, please provide the ADF pipeline support files.
237-
This Zip file contains the code-behind script from your data flow graph. From the ADF UI, click **...** next to pipeline, and then click **Download support files**.
237+
This Zip file contains the code-behind script from your data flow graph. From the ADF UI, select **...** next to pipeline, and then select **Download support files**.
238238
239239
### How do I access data by using the other 90 dataset types in Data Factory?
240240
@@ -244,23 +244,23 @@ sections:
244244
245245
### Is the self-hosted integration runtime available for data flows?
246246
247-
Self-hosted IR is an ADF pipeline construct that you can use with the Copy Activity to acquire or move data to and from on-prem or VM-based data sources and sinks. The virtual machines that you use for a self-hosted IR can also be placed inside of the same VNET as your protected data stores for access to those data stores from ADF. With data flows, you'll achieve these same end-results using the Azure IR with managed VNET instead.
247+
Self-hosted IR is an ADF pipeline construct that you can use with the Copy Activity to acquire or move data to and from on-premises or VM-based data sources and sinks. The virtual machines that you use for a self-hosted IR can also be placed inside of the same VNET as your protected data stores for access to those data stores from ADF. With data flows, you'll achieve these same end-results using the Azure IR with managed VNET instead.
248248
249249
### Does the data flow compute engine serve multiple tenants?
250250
251251
Clusters are never shared. We guarantee isolation for each job run in production runs. In case of debug scenario one person gets one cluster, and all debugs will go to that cluster which are initiated by that user.
252252
253-
### Is there a way to write attributes in cosmos db in the same order as specified in the sink in ADF data flow?
253+
### Is there a way to write attributes in Cosmos DB in the same order as specified in the sink in ADF data flow?
254254
255-
For cosmos DB, the underlying format of each document is a JSON object which is an unordered set of name/value pairs, so the order cannot be reserved.
255+
For Cosmos DB, the underlying format of each document is a JSON object which is an unordered set of name/value pairs, so the order cannot be reserved.
256256
257257
### Why a user is unable to use data preview in the data flows?
258258
259259
You should check permissions for custom role. There are multiple actions involved in the dataflow data preview. You start by checking network traffic while debugging on your browser. Please follow all of the actions, for details, please refer to [Resource provider.](../role-based-access-control/resource-provider-operations.md#microsoftdatafactory)
260260
261261
### In ADF, can I calculate value for a new column from existing column from mapping?
262262
263-
You can use derive transformation in mapping data flow to create a new column on the logic you want. When creating a derived column, you can either generate a new column or update an existing one. In the Column textbox, enter in the column you are creating. To override an existing column in your schema, you can use the column dropdown. To build the derived column's expression, click on the Enter expression textbox. You can either start typing your expression or open up the expression builder to construct your logic.
263+
You can use derive transformation in mapping data flow to create a new column on the logic you want. When creating a derived column, you can either generate a new column or update an existing one. In the Column textbox, enter in the column you are creating. To override an existing column in your schema, you can use the column dropdown. To build the derived column's expression, select on the Enter expression textbox. You can either start typing your expression or open up the expression builder to construct your logic.
264264
265265
### Why mapping data flow preview failing with Gateway timeout?
266266
@@ -283,7 +283,7 @@ sections:
283283
Data factory is available in following [regions.](https://azure.microsoft.com/global-infrastructure/services/?products=data-factory)
284284
The Power Query feature is available in all data flow regions. If the feature is not available in your region, please check with support.
285285
286-
### What is the difference between mapping data flow and Power query actvity (data wrangling)?
286+
### What is the difference between mapping data flow and Power query activity (data wrangling)?
287287
288288
Mapping data flows provide a way to transform data at scale without any coding required. You can design a data transformation job in the data flow canvas by constructing a series of transformations. Start with any number of source transformations followed by data transformation steps. Complete your data flow with a sink to land your results in a destination. Mapping data flow is great at mapping and transforming data with both known and unknown schemas in the sinks and sources.
289289

articles/data-factory/plan-manage-costs.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -145,7 +145,7 @@ Budgets can be created with filters for specific resources or services in Azure
145145

146146
## Export cost data
147147

148-
You can also [export your cost data](../cost-management-billing/costs/tutorial-export-acm-data.md?WT.mc_id=costmanagementcontent_docsacmhorizontal_-inproduct-learn) to a storage account. This is helpful when you need or others to do additional data analysis for costs. For example, a finance teams can analyze the data using Excel or Power BI. You can export your costs on a daily, weekly, or monthly schedule and set a custom date range. Exporting cost data is the recommended way to retrieve cost datasets.
148+
You can also [export your cost data](../cost-management-billing/costs/tutorial-export-acm-data.md?WT.mc_id=costmanagementcontent_docsacmhorizontal_-inproduct-learn) to a storage account. This is helpful when you need or others to do additional data analysis for costs. For example, finance teams can analyze the data using Excel or Power BI. You can export your costs on a daily, weekly, or monthly schedule and set a custom date range. Exporting cost data is the recommended way to retrieve cost datasets.
149149

150150
## Next steps
151151

articles/data-factory/pricing-examples-copy-transform-azure-databricks.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -24,11 +24,11 @@ Refer to the [Azure Pricing Calculator](https://azure.microsoft.com/pricing/calc
2424

2525
To accomplish the scenario, you need to create a pipeline with the following items:
2626

27-
1. One copy activity with an input dataset for the data to be copied from AWS S3, and an output dataset for the data on Azure storage.
28-
2. One Azure Databricks activity for the data transformation.
29-
3. One schedule trigger to execute the pipeline every hour. When you want to run a pipeline, you can either [trigger it immediately or schedule it](concepts-pipeline-execution-triggers.md). In addition to the pipeline itself, each trigger instance counts as a single Activity run.
27+
- One copy activity with an input dataset for the data to be copied from AWS S3, and an output dataset for the data on Azure storage.
28+
- One Azure Databricks activity for the data transformation.
29+
- One schedule trigger to execute the pipeline every hour. When you want to run a pipeline, you can either [trigger it immediately or schedule it](concepts-pipeline-execution-triggers.md). In addition to the pipeline itself, each trigger instance counts as a single Activity run.
3030

31-
:::image type="content" source="media/pricing-concepts/scenario2.png" alt-text="Diagram shows a pipeline with a schedule trigger. In the pipeline, copy activity flows to an input dataset, an output dataset, and a DataBricks activity, which runs on Azure Databricks. The input dataset flows to an A W S S3 linked service. The output dataset flows to an Azure Storage linked service.":::
31+
:::image type="content" source="media/pricing-concepts/scenario2.png" alt-text="Diagram shows a pipeline with a schedule trigger. In the pipeline, copy activity flows to an input dataset, an output dataset, and a DataBricks activity, which runs on Azure Databricks. The input dataset flows to an AWS S3 linked service. The output dataset flows to an Azure Storage linked service.":::
3232

3333
## Costs estimation
3434

@@ -42,9 +42,9 @@ To accomplish the scenario, you need to create a pipeline with the following ite
4242

4343
**Total scenario pricing for 30 days: $122.03**
4444

45-
:::image type="content" source="media/pricing-concepts/scenario-2-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for a copy data and transform with Azure Databricks scenario.":::
45+
:::image type="content" source="media/pricing-concepts/scenario-2-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for a copy data and transform with Azure Databricks scenario." lightbox="media/pricing-concepts/scenario-2-pricing-calculator.png":::
4646

47-
## Next Steps
47+
## Next steps
4848

4949
- [Pricing example: Copy data from AWS S3 to Azure Blob storage hourly for 30 days](pricing-examples-s3-to-blob.md)
5050
- [Pricing example: Copy data and transform with dynamic parameters hourly for 30 days](pricing-examples-copy-transform-dynamic-parameters.md)

articles/data-factory/pricing-examples-copy-transform-dynamic-parameters.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -24,12 +24,12 @@ Refer to the [Azure Pricing Calculator](https://azure.microsoft.com/pricing/calc
2424

2525
To accomplish the scenario, you need to create a pipeline with the following items:
2626

27-
1. One copy activity with an input dataset for the data to be copied from AWS S3, an output dataset for the data on Azure storage.
28-
2. One Lookup activity for passing parameters dynamically to the transformation script.
29-
3. One Azure Databricks activity for the data transformation.
30-
4. One schedule trigger to execute the pipeline every hour. When you want to run a pipeline, you can either [trigger it immediately or schedule it](concepts-pipeline-execution-triggers.md). In addition to the pipeline itself, each trigger instance counts as a single Activity run.
27+
- One copy activity with an input dataset for the data to be copied from AWS S3, an output dataset for the data on Azure storage.
28+
- One Lookup activity for passing parameters dynamically to the transformation script.
29+
- One Azure Databricks activity for the data transformation.
30+
- One schedule trigger to execute the pipeline every hour. When you want to run a pipeline, you can either [trigger it immediately or schedule it](concepts-pipeline-execution-triggers.md). In addition to the pipeline itself, each trigger instance counts as a single Activity run.
3131

32-
:::image type="content" source="media/pricing-concepts/scenario3.png" alt-text="Diagram shows a pipeline with a schedule trigger. In the pipeline, copy activity flows to an input dataset, an output dataset, and lookup activity that flows to a DataBricks activity, which runs on Azure Databricks. The input dataset flows to an A W S S3 linked service. The output dataset flows to an Azure Storage linked service.":::
32+
:::image type="content" source="media/pricing-concepts/scenario3.png" alt-text="Diagram shows a pipeline with a schedule trigger. In the pipeline, copy activity flows to an input dataset, an output dataset, and lookup activity that flows to a DataBricks activity, which runs on Azure Databricks. The input dataset flows to an AWS S3 linked service. The output dataset flows to an Azure Storage linked service.":::
3333

3434
## Costs estimation
3535

@@ -44,9 +44,9 @@ To accomplish the scenario, you need to create a pipeline with the following ite
4444

4545
**Total scenario pricing for 30 days: $122.09**
4646

47-
:::image type="content" source="media/pricing-concepts/scenario-3-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for a copy data and transform with dynamic parameters scenario.":::
47+
:::image type="content" source="media/pricing-concepts/scenario-3-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for a copy data and transform with dynamic parameters scenario." lightbox="media/pricing-concepts/scenario-3-pricing-calculator.png":::
4848

49-
## Next Steps
49+
## Next steps
5050

5151
- [Pricing example: Copy data from AWS S3 to Azure Blob storage hourly for 30 days](pricing-examples-s3-to-blob.md)
5252
- [Pricing example: Copy data and transform with Azure Databricks hourly for 30 days](pricing-examples-copy-transform-azure-databricks.md)

articles/data-factory/pricing-examples-data-integration-managed-vnet.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -41,9 +41,9 @@ To accomplish the scenario, you need to create two pipelines with the following
4141

4242
**Total scenario pricing for 30 days: $129.02**
4343

44-
:::image type="content" source="media/pricing-concepts/scenario-5-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for data integration with Managed VNET.":::
44+
:::image type="content" source="media/pricing-concepts/scenario-5-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for data integration with Managed VNET." lightbox="media/pricing-concepts/scenario-5-pricing-calculator.png":::
4545

46-
## Next Steps
46+
## Next steps
4747

4848
- [Pricing example: Copy data from AWS S3 to Azure Blob storage hourly for 30 days](pricing-examples-s3-to-blob.md)
4949
- [Pricing example: Copy data and transform with Azure Databricks hourly for 30 days](pricing-examples-copy-transform-azure-databricks.md)

articles/data-factory/pricing-examples-get-delta-data-from-sap-ecc.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ To accomplish the scenario, you need to create a pipeline with the following ite
3030

3131
## Costs estimation
3232

33-
In order to load data from SAP ECC via SAP CDC connector in Mapping Data Flow, you need to install your Self-Hosted Integration Runtime on an on-premise machine, or VM to directly connect to your SAP ECC system. Given that, you'll be charged on both Self-Hosted Integration Runtime with $0.10/hour and Mapping Data Flow with its vCore-hour price unit.
33+
In order to load data from SAP ECC via SAP CDC connector in Mapping Data Flow, you need to install your Self-Hosted Integration Runtime on an on-premises machine, or VM to directly connect to your SAP ECC system. Given that, you'll be charged on both Self-Hosted Integration Runtime with $0.10/hour and Mapping Data Flow with its vCore-hour price unit.
3434

3535
Assuming every time it requires 15 minutes to complete the job, the cost estimations are as below.
3636

@@ -44,9 +44,9 @@ Assuming every time it requires 15 minutes to complete the job, the cost estimat
4444

4545
**Total scenario pricing for 30 days: $17.21**
4646

47-
:::image type="content" source="media/pricing-concepts/scenario-6-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for getting delta data from SAP ECC via SAP CDC in mapping data flows.":::
47+
:::image type="content" source="media/pricing-concepts/scenario-6-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for getting delta data from SAP ECC via SAP CDC in mapping data flows." lightbox="media/pricing-concepts/scenario-6-pricing-calculator.png":::
4848

49-
## Next Steps
49+
## Next steps
5050

5151
- [Pricing example: Copy data from AWS S3 to Azure Blob storage hourly for 30 days](pricing-examples-s3-to-blob.md)
5252
- [Pricing example: Copy data and transform with Azure Databricks hourly for 30 days](pricing-examples-copy-transform-azure-databricks.md)

articles/data-factory/pricing-examples-mapping-data-flow-debug-workday.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ A data factory engineer is responsible for designing, building, and testing mapp
2626

2727
**8 (hours) x 8 (compute-optimized cores) x $0.193 = $12.35**
2828

29-
## Next Steps
29+
## Next steps
3030

3131
- [Pricing example: Copy data from AWS S3 to Azure Blob storage hourly for 30 days](pricing-examples-s3-to-blob.md)
3232
- [Pricing example: Copy data and transform with Azure Databricks hourly for 30 days](pricing-examples-copy-transform-azure-databricks.md)

articles/data-factory/pricing-examples-s3-to-blob.md

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -24,9 +24,8 @@ Refer to the [Azure Pricing Calculator](https://azure.microsoft.com/pricing/calc
2424

2525
To accomplish the scenario, you need to create a pipeline with the following items:
2626

27-
1. I'll copy data from AWS S3 to Azure Blob storage, and this will move 10 GB of data from S3 to blob storage. I estimate it will run for 2-3 hours, and I plan to set DIU as Auto.
28-
29-
3. A schedule trigger to execute the pipeline every hour for 8 hours every day. When you want to run a pipeline, you can either [trigger it immediately or schedule it](concepts-pipeline-execution-triggers.md). In addition to the pipeline itself, each trigger instance counts as a single Activity run.
27+
- I'll copy data from AWS S3 to Azure Blob storage, and this will move 10 GB of data from S3 to blob storage. I estimate it will run for 2-3 hours, and I plan to set DIU as Auto.
28+
- A schedule trigger to execute the pipeline every hour for 8 hours every day. When you want to run a pipeline, you can either [trigger it immediately or schedule it](concepts-pipeline-execution-triggers.md). In addition to the pipeline itself, each trigger instance counts as a single Activity run.
3029

3130
:::image type="content" source="media/pricing-concepts/scenario1.png" alt-text="Diagram shows a pipeline with a schedule trigger.":::
3231

@@ -42,9 +41,9 @@ To accomplish the scenario, you need to create a pipeline with the following ite
4241

4342
**Total scenario pricing for 30 days: $122.00**
4443

45-
:::image type="content" source="media/pricing-concepts/scenario-1-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for an hourly pipeline run.":::
44+
:::image type="content" source="media/pricing-concepts/scenario-1-pricing-calculator.png" alt-text="Screenshot of the pricing calculator configured for an hourly pipeline run." lightbox="media/pricing-concepts/scenario-1-pricing-calculator.png":::
4645

47-
## Next Steps
46+
## Next steps
4847

4948
- [Pricing example: Copy data and transform with Azure Databricks hourly for 30 days](pricing-examples-copy-transform-azure-databricks.md)
5049
- [Pricing example: Copy data and transform with dynamic parameters hourly for 30 days](pricing-examples-copy-transform-dynamic-parameters.md)

0 commit comments

Comments
 (0)