Skip to content

Commit fd7c7ce

Browse files
authored
Merge pull request #116158 from markingmyname/danielcheck
[ADF] minor grammar tweaks
2 parents 5e1ea5f + ab01ac0 commit fd7c7ce

7 files changed

+25
-28
lines changed

articles/data-factory/author-visually.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ To open the **authoring canvas**, click on the pencil icon.
2626

2727
![Authoring Canvas](media/author-visually/authoring-canvas.png)
2828

29-
Here, you will author the pipelines, activities, datasets, linked services, data flows, triggers, and integration runtimes that comprise your factory. To get started building a pipeline using the authoring canvas, see [Copy data using the copy Activity](tutorial-copy-data-portal.md).
29+
Here, you author the pipelines, activities, datasets, linked services, data flows, triggers, and integration runtimes that comprise your factory. To get started building a pipeline using the authoring canvas, see [Copy data using the copy Activity](tutorial-copy-data-portal.md).
3030

3131
The default visual authoring experience is directly working with the Data Factory service. Azure Repos Git or GitHub integration is also supported to allow source control and collaboration for work on your data factory pipelines. To learn more about the differences between these authoring experiences, see [Source control in Azure Data Factory](source-control.md).
3232

@@ -36,7 +36,7 @@ For top-level resources such as pipelines, datasets, and data flows, high-level
3636

3737
![Authoring Canvas](media/author-visually/properties-pane.png)
3838

39-
The properties pane will only be open by default on resource creation. To edit it, click on the properties pane icon located in the top-right corner of the canvas.
39+
The properties pane only opens by default on resource creation. To edit it, click on the properties pane icon located in the top-right corner of the canvas.
4040

4141
## Expressions and functions
4242

articles/data-factory/concepts-data-flow-monitoring.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -14,17 +14,17 @@ ms.date: 04/17/2020
1414

1515
[!INCLUDE[appliesto-adf-asa-md](includes/appliesto-adf-asa-md.md)]
1616

17-
After you have completed building and debugging your data flow, you will want to schedule your data flow to execute on a schedule within the context of a pipeline. You can schedule the pipeline from Azure Data Factory using Triggers. Or you can use the Trigger Now option from the Azure Data Factory Pipeline Builder to execute a single-run execution to test your data flow within the pipeline context.
17+
After you have completed building and debugging your data flow, you want to schedule your data flow to execute on a schedule within the context of a pipeline. You can schedule the pipeline from Azure Data Factory using Triggers. Or you can use the Trigger Now option from the Azure Data Factory Pipeline Builder to execute a single-run execution to test your data flow within the pipeline context.
1818

19-
When you execute your pipeline, you will be able to monitor the pipeline and all of the activities contained in the pipeline including the Data Flow activity. Click on the monitor icon in the left-hand Azure Data Factory UI panel. You will see a screen similar to the one below. The highlighted icons will allow you to drill into the activities in the pipeline, including the Data Flow activity.
19+
When you execute your pipeline, you can monitor the pipeline and all of the activities contained in the pipeline including the Data Flow activity. Click on the monitor icon in the left-hand Azure Data Factory UI panel. You can see a screen similar to the one below. The highlighted icons allow you to drill into the activities in the pipeline, including the Data Flow activity.
2020

2121
![Data Flow Monitoring](media/data-flow/mon001.png "Data Flow Monitoring")
2222

23-
You will see statistics at this level as well including the run times and status. The Run ID at the activity level is different that the Run ID at the pipeline level. The Run ID at the previous level is for the pipeline. Clicking the eyeglasses will give you deep details on your data flow execution.
23+
You see statistics at this level as well including the run times and status. The Run ID at the activity level is different than the Run ID at the pipeline level. The Run ID at the previous level is for the pipeline. Selecting the eyeglasses gives you deep details on your data flow execution.
2424

2525
![Data Flow Monitoring](media/data-flow/mon002.png "Data Flow Monitoring")
2626

27-
When you are in the graphical node monitoring view, you will see a simplified view-only version of your data flow graph.
27+
When you're in the graphical node monitoring view, you can see a simplified view-only version of your data flow graph.
2828

2929
![Data Flow Monitoring](media/data-flow/mon003.png "Data Flow Monitoring")
3030

@@ -34,18 +34,18 @@ Here is a video overview of monitoring performance of your data flows from the A
3434
3535
## View Data Flow Execution Plans
3636

37-
When your Data Flow is executed in Spark, Azure Data Factory determines optimal code paths based on the entirety of your data flow. Additionally, the execution paths may occur on different scale-out nodes and data partitions. Therefore, the monitoring graph represents the design of your flow, taking into account the execution path of your transformations. When you click on individual nodes, you will see "groupings" that represent code that was executed together on the cluster. The timings and counts that you see represent those groups as opposed to the individual steps in your design.
37+
When your Data Flow is executed in Spark, Azure Data Factory determines optimal code paths based on the entirety of your data flow. Additionally, the execution paths may occur on different scale-out nodes and data partitions. Therefore, the monitoring graph represents the design of your flow, taking into account the execution path of your transformations. When you select individual nodes, you can see "groupings" that represent code that was executed together on the cluster. The timings and counts that you see represent those groups as opposed to the individual steps in your design.
3838

3939
![Data Flow Monitoring](media/data-flow/mon004.png "Data Flow Monitoring")
4040

41-
* When you click on the open space in the monitoring window, the stats in the bottom pane will display timing and row counts for each Sink and the transformations that led to the sink data for transformation lineage.
41+
* When you select the open space in the monitoring window, the stats in the bottom pane display timing and row counts for each Sink and the transformations that led to the sink data for transformation lineage.
4242

43-
* When you select individual transformations, you will receive additional feedback on the right-hand panel that shows partition stats, column counts, skewness (how evenly is the data distributed across partitions), and kurtosis (how spiky is the data).
43+
* When you select individual transformations, you receive additional feedback on the right-hand panel that shows partition stats, column counts, skewness (how evenly is the data distributed across partitions), and kurtosis (how spiky is the data).
4444

45-
* When you click on the Sink in the node view, you will see column lineage. There are three different methods that columns are accumulated throughout your data flow to land in the Sink. They are:
45+
* When you select the Sink in the node view, you can see column lineage. There are three different methods that columns are accumulated throughout your data flow to land in the Sink. They are:
4646

47-
* Computed: You use the column for conditional processing or within an expression in your data flow, but do not land it in the Sink
48-
* Derived: The column is a new column that you generated in your flow, i.e. it was not present in the Source
47+
* Computed: You use the column for conditional processing or within an expression in your data flow, but don't land it in the Sink
48+
* Derived: The column is a new column that you generated in your flow, that is, it was not present in the Source
4949
* Mapped: The column originated from the source and your are mapping it to a sink field
5050
* Data flow status: The current status of your execution
5151
* Cluster startup time: Amount of time to acquire the JIT Spark compute environment for your data flow execution
@@ -59,4 +59,4 @@ This icon means that the transformation data was already cached on the cluster,
5959

6060
![Data Flow Monitoring](media/data-flow/mon004.png "Data Flow Monitoring")
6161

62-
You will also see green circle icons in the transformation. They represent a count of the number of sinks that data is flowing into.
62+
You also see green circle icons in the transformation. They represent a count of the number of sinks that data is flowing into.

articles/data-factory/concepts-datasets-linked-services.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -29,9 +29,9 @@ If you are new to Data Factory, see [Introduction to Azure Data Factory](introdu
2929
## Overview
3030
A data factory can have one or more pipelines. A **pipeline** is a logical grouping of **activities** that together perform a task. The activities in a pipeline define actions to perform on your data. Now, a **dataset** is a named view of data that simply points or references the data you want to use in your **activities** as inputs and outputs. Datasets identify data within different data stores, such as tables, files, folders, and documents. For example, an Azure Blob dataset specifies the blob container and folder in Blob storage from which the activity should read the data.
3131

32-
Before you create a dataset, you must create a [**linked service**](concepts-linked-services.md) to link your data store to the data factory. Linked services are much like connection strings, which define the connection information needed for Data Factory to connect to external resources. Think of it this way; the dataset represents the structure of the data within the linked data stores, and the linked service defines the connection to the data source. For example, an Azure Storage linked service links a storage account to the data factory. An Azure Blob dataset represents the blob container and the folder within that Azure storage account that contains the input blobs to be processed.
32+
Before you create a dataset, you must create a [**linked service**](concepts-linked-services.md) to link your data store to the data factory. Linked services are much like connection strings, which define the connection information needed for Data Factory to connect to external resources. Think of it this way; the dataset represents the structure of the data within the linked data stores, and the linked service defines the connection to the data source. For example, an Azure Storage linked service links a storage account to the data factory. An Azure Blob dataset represents the blob container and the folder within that Azure Storage account that contains the input blobs to be processed.
3333

34-
Here is a sample scenario. To copy data from Blob storage to a SQL database, you create two linked services: Azure Storage and Azure SQL Database. Then, create two datasets: Azure Blob dataset (which refers to the Azure Storage linked service) and Azure SQL Table dataset (which refers to the Azure SQL Database linked service). The Azure Storage and Azure SQL Database linked services contain connection strings that Data Factory uses at runtime to connect to your Azure Storage and Azure SQL Database, respectively. The Azure Blob dataset specifies the blob container and blob folder that contains the input blobs in your Blob storage. The Azure SQL Table dataset specifies the SQL table in your SQL database to which the data is to be copied.
34+
Here is a sample scenario. To copy data from Blob storage to a SQL Database, you create two linked services: Azure Storage and Azure SQL Database. Then, create two datasets: Azure Blob dataset (which refers to the Azure Storage linked service) and Azure SQL Table dataset (which refers to the Azure SQL Database linked service). The Azure Storage and Azure SQL Database linked services contain connection strings that Data Factory uses at runtime to connect to your Azure Storage and Azure SQL Database, respectively. The Azure Blob dataset specifies the blob container and blob folder that contains the input blobs in your Blob storage. The Azure SQL Table dataset specifies the SQL table in your SQL Database to which the data is to be copied.
3535

3636
The following diagram shows the relationships among pipeline, activity, dataset, and linked service in Data Factory:
3737

@@ -119,7 +119,7 @@ typeProperties | The type properties are different for each type (for example: A
119119

120120

121121
## Dataset example
122-
In the following example, the dataset represents a table named MyTable in a SQL database.
122+
In the following example, the dataset represents a table named MyTable in a SQL Database.
123123

124124
```json
125125
{
@@ -202,7 +202,7 @@ Define the Blob dataset structure as follows along with type definitions for the
202202
The following guidelines help you understand when to include structure information, and what to include in the **structure** section. Learn more on how data factory maps source data to sink and when to specify structure information from [Schema and type mapping](copy-activity-schema-and-type-mapping.md).
203203

204204
- **For strong schema data sources**, specify the structure section only if you want map source columns to sink columns, and their names are not the same. This kind of structured data source stores data schema and type information along with the data itself. Examples of structured data sources include SQL Server, Oracle, and Azure SQL Database.<br/><br/>As type information is already available for structured data sources, you should not include type information when you do include the structure section.
205-
- **For no/weak schema data sources e.g. text file in blob storage**, include structure when the dataset is an input for a copy activity, and data types of source dataset should be converted to native types for the sink. And include structure when you want to map source columns to sink columns..
205+
- **For no/weak schema data sources for example, text file in blob storage**, include structure when the dataset is an input for a copy activity, and data types of source dataset should be converted to native types for the sink. And include structure when you want to map source columns to sink columns
206206

207207
## Create datasets
208208
You can create datasets by using one of these tools or SDKs: [.NET API](quickstart-create-data-factory-dot-net.md), [PowerShell](quickstart-create-data-factory-powershell.md), [REST API](quickstart-create-data-factory-rest-api.md), Azure Resource Manager Template, and Azure portal

articles/data-factory/concepts-linked-services.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,15 +15,15 @@ ms.date: 04/25/2019
1515

1616
# Linked services in Azure Data Factory
1717

18-
> [!div class="op_single_selector" title1="Select the version of Data Factory service you are using:"]
18+
> [!div class="op_single_selector" title1="Select the version of Data Factory service you're using:"]
1919
> * [Version 1](v1/data-factory-create-datasets.md)
2020
> * [Current version](concepts-linked-services.md)
2121
2222
[!INCLUDE[appliesto-adf-asa-md](includes/appliesto-adf-asa-md.md)]
2323

2424
This article describes what linked services are, how they're defined in JSON format, and how they're used in Azure Data Factory pipelines.
2525

26-
If you are new to Data Factory, see [Introduction to Azure Data Factory](introduction.md) for an overview.
26+
If you're new to Data Factory, see [Introduction to Azure Data Factory](introduction.md) for an overview.
2727

2828
## Overview
2929

articles/data-factory/concepts-roles-permissions.md

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,6 @@ ms.service: data-factory
77
services: data-factory
88
documentationcenter: ''
99
ms.workload: data-services
10-
1110
author: djpmsft
1211
ms.author: daperlov
1312
manager: anandsub
@@ -65,7 +64,7 @@ Here are a few examples that demonstrate what you can achieve with custom roles:
6564

6665
- Let a user create, edit, or delete any data factory in a resource group from the Azure portal.
6766

68-
Assign the built-in **Data Factory contributor** role at the resource group level for the user. If you want to allow access to any data factory in a subscription, assign the role at the subscription level.
67+
Assign the built-in **Data Factory contributor** role at the resource group level for the user. If you want to allow access to any data factory in a subscription, assign the role at the subscription level.
6968

7069
- Let a user view (read) and monitor a data factory, but not edit or change it.
7170

@@ -80,7 +79,7 @@ Here are a few examples that demonstrate what you can achieve with custom roles:
8079

8180
- Let a user only be able to test connection in a linked service
8281

83-
Create a custom role role with permissions for the following actions: **Microsoft.DataFactory/factories/getFeatureValue/read** and **Microsoft.DataFactory/factories/getDataPlaneAccess/read**. Assign this custom role on the data factory resource for the user.
82+
Create a custom role with permissions for the following actions: **Microsoft.DataFactory/factories/getFeatureValue/read** and **Microsoft.DataFactory/factories/getDataPlaneAccess/read**. Assign this custom role on the data factory resource for the user.
8483

8584
- Let a user update a data factory from PowerShell or the SDK, but not in the Azure portal.
8685

articles/data-factory/control-flow-set-variable-activity.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,8 @@ Property | Description | Required
2424
name | Name of the activity in pipeline | yes
2525
description | Text describing what the activity does | no
2626
type | Must be set to **SetVariable** | yes
27-
value | String literal or expression object value that the variable will be assigned to | yes
28-
variableName | Name of the variable that will be set by this activity | yes
27+
value | String literal or expression object value that the variable is assigned to | yes
28+
variableName | Name of the variable that is set by this activity | yes
2929

3030
## Incrementing a variable
3131

articles/data-factory/v1/data-factory-build-your-first-pipeline-using-powershell.md

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -20,8 +20,6 @@ ms.date: 01/22/2018
2020
> * [PowerShell](data-factory-build-your-first-pipeline-using-powershell.md)
2121
> * [Resource Manager Template](data-factory-build-your-first-pipeline-using-arm.md)
2222
> * [REST API](data-factory-build-your-first-pipeline-using-rest-api.md)
23-
>
24-
>
2523
2624

2725
> [!NOTE]
@@ -112,7 +110,7 @@ In this step, you link your Azure Storage account to your data factory. You use
112110
}
113111
}
114112
```
115-
Replace **account name** with the name of your Azure storage account and **account key** with the access key of the Azure storage account. To learn how to get your storage access key, see [Manage storage account access keys](../../storage/common/storage-account-keys-manage.md).
113+
Replace **account name** with the name of your Azure Storage account and **account key** with the access key of the Azure Storage account. To learn how to get your storage access key, see [Manage storage account access keys](../../storage/common/storage-account-keys-manage.md).
116114
2. In Azure PowerShell, switch to the ADFGetStarted folder.
117115
3. You can use the **New-AzDataFactoryLinkedService** cmdlet that creates a linked service. This cmdlet and other Data Factory cmdlets you use in this tutorial requires you to pass values for the *ResourceGroupName* and *DataFactoryName* parameters. Alternatively, you can use **Get-AzDataFactory** to get a **DataFactory** object and pass the object without typing *ResourceGroupName* and *DataFactoryName* each time you run a cmdlet. Run the following command to assign the output of the **Get-AzDataFactory** cmdlet to a **$df** variable.
118116
@@ -310,7 +308,7 @@ In this step, you create your first pipeline with a **HDInsightHive** activity.
310308
```
311309
In the JSON snippet, you are creating a pipeline that consists of a single activity that uses Hive to process Data on an HDInsight cluster.
312310
313-
The Hive script file, **partitionweblogs.hql**, is stored in the Azure storage account (specified by the scriptLinkedService, called **StorageLinkedService**), and in **script** folder in the container **adfgetstarted**.
311+
The Hive script file, **partitionweblogs.hql**, is stored in the Azure Storage account (specified by the scriptLinkedService, called **StorageLinkedService**), and in **script** folder in the container **adfgetstarted**.
314312
315313
The **defines** section is used to specify the runtime settings that be passed to the hive script as Hive configuration values (e.g ${hiveconf:inputtable}, ${hiveconf:partitionedtable}).
316314

0 commit comments

Comments
 (0)