Skip to content

Commit 5044b04

Browse files
committed
Article tweaks
1 parent d8cf7d5 commit 5044b04

6 files changed

+14
-17
lines changed

articles/data-factory/author-visually.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@ To open the **authoring canvas**, click on the pencil icon.
2626

2727
![Authoring Canvas](media/author-visually/authoring-canvas.png)
2828

29-
Here, you will author the pipelines, activities, datasets, linked services, data flows, triggers, and integration runtimes that comprise your factory. To get started building a pipeline using the authoring canvas, see [Copy data using the copy Activity](tutorial-copy-data-portal.md).
29+
Here, you author the pipelines, activities, datasets, linked services, data flows, triggers, and integration runtimes that comprise your factory. To get started building a pipeline using the authoring canvas, see [Copy data using the copy Activity](tutorial-copy-data-portal.md).
3030

3131
The default visual authoring experience is directly working with the Data Factory service. Azure Repos Git or GitHub integration is also supported to allow source control and collaboration for work on your data factory pipelines. To learn more about the differences between these authoring experiences, see [Source control in Azure Data Factory](source-control.md).
3232

@@ -36,7 +36,7 @@ For top-level resources such as pipelines, datasets, and data flows, high-level
3636

3737
![Authoring Canvas](media/author-visually/properties-pane.png)
3838

39-
The properties pane will only be open by default on resource creation. To edit it, click on the properties pane icon located in the top-right corner of the canvas.
39+
The properties pane only opens by default on resource creation. To edit it, click on the properties pane icon located in the top-right corner of the canvas.
4040

4141
## Expressions and functions
4242

articles/data-factory/concepts-datasets-linked-services.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -29,9 +29,9 @@ If you are new to Data Factory, see [Introduction to Azure Data Factory](introdu
2929
## Overview
3030
A data factory can have one or more pipelines. A **pipeline** is a logical grouping of **activities** that together perform a task. The activities in a pipeline define actions to perform on your data. Now, a **dataset** is a named view of data that simply points or references the data you want to use in your **activities** as inputs and outputs. Datasets identify data within different data stores, such as tables, files, folders, and documents. For example, an Azure Blob dataset specifies the blob container and folder in Blob storage from which the activity should read the data.
3131

32-
Before you create a dataset, you must create a [**linked service**](concepts-linked-services.md) to link your data store to the data factory. Linked services are much like connection strings, which define the connection information needed for Data Factory to connect to external resources. Think of it this way; the dataset represents the structure of the data within the linked data stores, and the linked service defines the connection to the data source. For example, an Azure Storage linked service links a storage account to the data factory. An Azure Blob dataset represents the blob container and the folder within that Azure storage account that contains the input blobs to be processed.
32+
Before you create a dataset, you must create a [**linked service**](concepts-linked-services.md) to link your data store to the data factory. Linked services are much like connection strings, which define the connection information needed for Data Factory to connect to external resources. Think of it this way; the dataset represents the structure of the data within the linked data stores, and the linked service defines the connection to the data source. For example, an Azure Storage linked service links a storage account to the data factory. An Azure Blob dataset represents the blob container and the folder within that Azure Storage account that contains the input blobs to be processed.
3333

34-
Here is a sample scenario. To copy data from Blob storage to a SQL database, you create two linked services: Azure Storage and Azure SQL Database. Then, create two datasets: Azure Blob dataset (which refers to the Azure Storage linked service) and Azure SQL Table dataset (which refers to the Azure SQL Database linked service). The Azure Storage and Azure SQL Database linked services contain connection strings that Data Factory uses at runtime to connect to your Azure Storage and Azure SQL Database, respectively. The Azure Blob dataset specifies the blob container and blob folder that contains the input blobs in your Blob storage. The Azure SQL Table dataset specifies the SQL table in your SQL database to which the data is to be copied.
34+
Here is a sample scenario. To copy data from Blob storage to a SQL Database, you create two linked services: Azure Storage and Azure SQL Database. Then, create two datasets: Azure Blob dataset (which refers to the Azure Storage linked service) and Azure SQL Table dataset (which refers to the Azure SQL Database linked service). The Azure Storage and Azure SQL Database linked services contain connection strings that Data Factory uses at runtime to connect to your Azure Storage and Azure SQL Database, respectively. The Azure Blob dataset specifies the blob container and blob folder that contains the input blobs in your Blob storage. The Azure SQL Table dataset specifies the SQL table in your SQL Database to which the data is to be copied.
3535

3636
The following diagram shows the relationships among pipeline, activity, dataset, and linked service in Data Factory:
3737

@@ -119,7 +119,7 @@ typeProperties | The type properties are different for each type (for example: A
119119

120120

121121
## Dataset example
122-
In the following example, the dataset represents a table named MyTable in a SQL database.
122+
In the following example, the dataset represents a table named MyTable in a SQL Database.
123123

124124
```json
125125
{
@@ -202,7 +202,7 @@ Define the Blob dataset structure as follows along with type definitions for the
202202
The following guidelines help you understand when to include structure information, and what to include in the **structure** section. Learn more on how data factory maps source data to sink and when to specify structure information from [Schema and type mapping](copy-activity-schema-and-type-mapping.md).
203203

204204
- **For strong schema data sources**, specify the structure section only if you want map source columns to sink columns, and their names are not the same. This kind of structured data source stores data schema and type information along with the data itself. Examples of structured data sources include SQL Server, Oracle, and Azure SQL Database.<br/><br/>As type information is already available for structured data sources, you should not include type information when you do include the structure section.
205-
- **For no/weak schema data sources e.g. text file in blob storage**, include structure when the dataset is an input for a copy activity, and data types of source dataset should be converted to native types for the sink. And include structure when you want to map source columns to sink columns..
205+
- **For no/weak schema data sources for example, text file in blob storage**, include structure when the dataset is an input for a copy activity, and data types of source dataset should be converted to native types for the sink. And include structure when you want to map source columns to sink columns
206206

207207
## Create datasets
208208
You can create datasets by using one of these tools or SDKs: [.NET API](quickstart-create-data-factory-dot-net.md), [PowerShell](quickstart-create-data-factory-powershell.md), [REST API](quickstart-create-data-factory-rest-api.md), Azure Resource Manager Template, and Azure portal

articles/data-factory/concepts-linked-services.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,15 +15,15 @@ ms.date: 04/25/2019
1515

1616
# Linked services in Azure Data Factory
1717

18-
> [!div class="op_single_selector" title1="Select the version of Data Factory service you are using:"]
18+
> [!div class="op_single_selector" title1="Select the version of Data Factory service you're using:"]
1919
> * [Version 1](v1/data-factory-create-datasets.md)
2020
> * [Current version](concepts-linked-services.md)
2121
2222
[!INCLUDE[appliesto-adf-asa-md](includes/appliesto-adf-asa-md.md)]
2323

2424
This article describes what linked services are, how they're defined in JSON format, and how they're used in Azure Data Factory pipelines.
2525

26-
If you are new to Data Factory, see [Introduction to Azure Data Factory](introduction.md) for an overview.
26+
If you're new to Data Factory, see [Introduction to Azure Data Factory](introduction.md) for an overview.
2727

2828
## Overview
2929

articles/data-factory/concepts-roles-permissions.md

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,6 @@ ms.service: data-factory
77
services: data-factory
88
documentationcenter: ''
99
ms.workload: data-services
10-
1110
author: djpmsft
1211
ms.author: daperlov
1312
manager: anandsub
@@ -65,7 +64,7 @@ Here are a few examples that demonstrate what you can achieve with custom roles:
6564

6665
- Let a user create, edit, or delete any data factory in a resource group from the Azure portal.
6766

68-
Assign the built-in **Data Factory contributor** role at the resource group level for the user. If you want to allow access to any data factory in a subscription, assign the role at the subscription level.
67+
Assign the built-in **Data Factory contributor** role at the resource group level for the user. If you want to allow, access to any data factory in a subscription, assign the role at the subscription level.
6968

7069
- Let a user view (read) and monitor a data factory, but not edit or change it.
7170

@@ -80,7 +79,7 @@ Here are a few examples that demonstrate what you can achieve with custom roles:
8079

8180
- Let a user only be able to test connection in a linked service
8281

83-
Create a custom role role with permissions for the following actions: **Microsoft.DataFactory/factories/getFeatureValue/read** and **Microsoft.DataFactory/factories/getDataPlaneAccess/read**. Assign this custom role on the data factory resource for the user.
82+
Create a custom role with permissions for the following actions: **Microsoft.DataFactory/factories/getFeatureValue/read** and **Microsoft.DataFactory/factories/getDataPlaneAccess/read**. Assign this custom role on the data factory resource for the user.
8483

8584
- Let a user update a data factory from PowerShell or the SDK, but not in the Azure portal.
8685

articles/data-factory/control-flow-set-variable-activity.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,8 @@ Property | Description | Required
2424
name | Name of the activity in pipeline | yes
2525
description | Text describing what the activity does | no
2626
type | Must be set to **SetVariable** | yes
27-
value | String literal or expression object value that the variable will be assigned to | yes
28-
variableName | Name of the variable that will be set by this activity | yes
27+
value | String literal or expression object value that the variable is assigned to | yes
28+
variableName | Name of the variable that is set by this activity | yes
2929

3030
## Incrementing a variable
3131

articles/data-factory/v1/data-factory-build-your-first-pipeline-using-powershell.md

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -20,8 +20,6 @@ ms.date: 01/22/2018
2020
> * [PowerShell](data-factory-build-your-first-pipeline-using-powershell.md)
2121
> * [Resource Manager Template](data-factory-build-your-first-pipeline-using-arm.md)
2222
> * [REST API](data-factory-build-your-first-pipeline-using-rest-api.md)
23-
>
24-
>
2523
2624

2725
> [!NOTE]
@@ -112,7 +110,7 @@ In this step, you link your Azure Storage account to your data factory. You use
112110
}
113111
}
114112
```
115-
Replace **account name** with the name of your Azure storage account and **account key** with the access key of the Azure storage account. To learn how to get your storage access key, see [Manage storage account access keys](../../storage/common/storage-account-keys-manage.md).
113+
Replace **account name** with the name of your Azure Storage account and **account key** with the access key of the Azure Storage account. To learn how to get your storage access key, see [Manage storage account access keys](../../storage/common/storage-account-keys-manage.md).
116114
2. In Azure PowerShell, switch to the ADFGetStarted folder.
117115
3. You can use the **New-AzDataFactoryLinkedService** cmdlet that creates a linked service. This cmdlet and other Data Factory cmdlets you use in this tutorial requires you to pass values for the *ResourceGroupName* and *DataFactoryName* parameters. Alternatively, you can use **Get-AzDataFactory** to get a **DataFactory** object and pass the object without typing *ResourceGroupName* and *DataFactoryName* each time you run a cmdlet. Run the following command to assign the output of the **Get-AzDataFactory** cmdlet to a **$df** variable.
118116
@@ -310,7 +308,7 @@ In this step, you create your first pipeline with a **HDInsightHive** activity.
310308
```
311309
In the JSON snippet, you are creating a pipeline that consists of a single activity that uses Hive to process Data on an HDInsight cluster.
312310
313-
The Hive script file, **partitionweblogs.hql**, is stored in the Azure storage account (specified by the scriptLinkedService, called **StorageLinkedService**), and in **script** folder in the container **adfgetstarted**.
311+
The Hive script file, **partitionweblogs.hql**, is stored in the Azure Storage account (specified by the scriptLinkedService, called **StorageLinkedService**), and in **script** folder in the container **adfgetstarted**.
314312
315313
The **defines** section is used to specify the runtime settings that be passed to the hive script as Hive configuration values (e.g ${hiveconf:inputtable}, ${hiveconf:partitionedtable}).
316314

0 commit comments

Comments
 (0)