Skip to content

Commit 347b451

Browse files
authored
Merge pull request #85101 from linda33wj/master
Update ADF copy content
2 parents 034d06e + 5b764c3 commit 347b451

7 files changed

+33
-99
lines changed

articles/data-factory/connector-azure-blob-storage.md

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ ms.reviewer: craigg
88
ms.service: data-factory
99
ms.workload: data-services
1010
ms.topic: conceptual
11-
ms.date: 08/06/2019
11+
ms.date: 08/12/2019
1212
ms.author: jingwang
1313

1414
---
@@ -71,6 +71,9 @@ To use storage account key authentication, the following properties are supporte
7171
| connectionString | Specify the information needed to connect to Storage for the connectionString property. <br/>Mark this field as a SecureString to store it securely in Data Factory. You can also put account key in Azure Key Vault and pull the `accountKey` configuration out of the connection string. Refer to the following samples and [Store credentials in Azure Key Vault](store-credentials-in-key-vault.md) article with more details. |Yes |
7272
| connectVia | The [integration runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Azure Integration Runtime or Self-hosted Integration Runtime (if your data store is in a private network). If not specified, it uses the default Azure Integration Runtime. |No |
7373

74+
>[!NOTE]
75+
>Secondary Blob Service Endpoint is not supported when using account key authentication. You can use other authentication types.
76+
7477
>[!NOTE]
7578
>If you were using "AzureStorage" type linked service, it is still supported as-is, while you are suggested to use this new "AzureBlobStorage" linked service type going forward.
7679

articles/data-factory/connector-azure-data-lake-storage.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,13 +9,13 @@ ms.reviewer: douglasl
99
ms.service: data-factory
1010
ms.workload: data-services
1111
ms.topic: conceptual
12-
ms.date: 08/06/2019
12+
ms.date: 08/12/2019
1313
ms.author: jingwang
1414

1515
---
1616
# Copy data to or from Azure Data Lake Storage Gen2 using Azure Data Factory
1717

18-
Azure Data Lake Storage Gen2 is a set of capabilities dedicated to big data analytics built into [Azure Blob storage](../storage/blobs/storage-blobs-introduction.md). You can use it to interface with your data by using both file system and object storage paradigms.
18+
Azure Data Lake Storage Gen2 (ADLS Gen2) is a set of capabilities dedicated to big data analytics built into [Azure Blob storage](../storage/blobs/storage-blobs-introduction.md). You can use it to interface with your data by using both file system and object storage paradigms.
1919

2020
This article outlines how to copy data to and from Azure Data Lake Storage Gen2. To learn about Azure Data Factory, read the [introductory article](introduction.md).
2121

@@ -70,6 +70,9 @@ To use storage account key authentication, the following properties are supporte
7070
| accountKey | Account key for Data Lake Storage Gen2. Mark this field as a SecureString to store it securely in Data Factory, or [reference a secret stored in Azure Key Vault](store-credentials-in-key-vault.md). |Yes |
7171
| connectVia | The [integration runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use the Azure integration runtime or a self-hosted integration runtime if your data store is in a private network. If this property isn't specified, the default Azure integration runtime is used. |No |
7272

73+
>[!NOTE]
74+
>Secondary ADLS file system endpoint is not supported when using account key authentication. You can use other authentication types.
75+
7376
**Example:**
7477

7578
```json

articles/data-factory/connector-azure-sql-database.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 08/06/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -259,18 +259,18 @@ For a full list of sections and properties available for defining activities, se
259259

260260
### Azure SQL Database as the source
261261

262-
To copy data from Azure SQL Database, set the **type** property in the copy activity source to **SqlSource**. The following properties are supported in the copy activity **source** section:
262+
To copy data from Azure SQL Database, the following properties are supported in the copy activity **source** section:
263263

264264
| Property | Description | Required |
265265
|:--- |:--- |:--- |
266-
| type | The **type** property of the copy activity source must be set to **SqlSource**. | Yes |
266+
| type | The **type** property of the copy activity source must be set to **AzureSqlSource**. "SqlSource" type is still supported for backward compatibility. | Yes |
267267
| sqlReaderQuery | This property uses the custom SQL query to read data. An example is `select * from MyTable`. | No |
268268
| sqlReaderStoredProcedureName | The name of the stored procedure that reads data from the source table. The last SQL statement must be a SELECT statement in the stored procedure. | No |
269269
| storedProcedureParameters | Parameters for the stored procedure.<br/>Allowed values are name or value pairs. The names and casing of parameters must match the names and casing of the stored procedure parameters. | No |
270270

271271
**Points to note:**
272272

273-
- If **sqlReaderQuery** is specified for **SqlSource**, the copy activity runs this query against the Azure SQL Database source to get the data. You also can specify a stored procedure by specifying **sqlReaderStoredProcedureName** and **storedProcedureParameters** if the stored procedure takes parameters.
273+
- If **sqlReaderQuery** is specified for **AzureSqlSource**, the copy activity runs this query against the Azure SQL Database source to get the data. You also can specify a stored procedure by specifying **sqlReaderStoredProcedureName** and **storedProcedureParameters** if the stored procedure takes parameters.
274274
- If you don't specify either **sqlReaderQuery** or **sqlReaderStoredProcedureName**, the columns defined in the "structure" section of the dataset JSON are used to construct a query. The query `select column1, column2 from mytable` runs against Azure SQL Database. If the dataset definition doesn't have "structure," all columns are selected from the table.
275275

276276
#### SQL query example
@@ -294,7 +294,7 @@ To copy data from Azure SQL Database, set the **type** property in the copy acti
294294
],
295295
"typeProperties": {
296296
"source": {
297-
"type": "SqlSource",
297+
"type": "AzureSqlSource",
298298
"sqlReaderQuery": "SELECT * FROM MyTable"
299299
},
300300
"sink": {
@@ -326,7 +326,7 @@ To copy data from Azure SQL Database, set the **type** property in the copy acti
326326
],
327327
"typeProperties": {
328328
"source": {
329-
"type": "SqlSource",
329+
"type": "AzureSqlSource",
330330
"sqlReaderStoredProcedureName": "CopyTestSrcStoredProcedureWithParameters",
331331
"storedProcedureParameters": {
332332
"stringData": { "value": "str3" },
@@ -365,11 +365,11 @@ GO
365365
> [!TIP]
366366
> Learn more about the supported write behaviors, configurations, and best practices from [Best practice for loading data into Azure SQL Database](#best-practice-for-loading-data-into-azure-sql-database).
367367

368-
To copy data to Azure SQL Database, set the **type** property in the copy activity sink to **SqlSink**. The following properties are supported in the copy activity **sink** section:
368+
To copy data to Azure SQL Database, the following properties are supported in the copy activity **sink** section:
369369

370370
| Property | Description | Required |
371371
|:--- |:--- |:--- |
372-
| type | The **type** property of the copy activity sink must be set to **SqlSink**. | Yes |
372+
| type | The **type** property of the copy activity sink must be set to **AzureSqlSink**. "SqlSink" type is still supported for backward compatibility. | Yes |
373373
| writeBatchSize | Number of rows to insert into the SQL table *per batch*.<br/> The allowed value is **integer** (number of rows). By default, Azure Data Factory dynamically determines the appropriate batch size based on the row size. | No |
374374
| writeBatchTimeout | The wait time for the batch insert operation to finish before it times out.<br/> The allowed value is **timespan**. An example is “00:30:00” (30 minutes). | No |
375375
| preCopyScript | Specify a SQL query for the copy activity to run before writing data into Azure SQL Database. It's invoked only once per copy run. Use this property to clean up the preloaded data. | No |
@@ -402,7 +402,7 @@ To copy data to Azure SQL Database, set the **type** property in the copy activi
402402
"type": "<source type>"
403403
},
404404
"sink": {
405-
"type": "SqlSink",
405+
"type": "AzureSqlSink",
406406
"writeBatchSize": 100000
407407
}
408408
}
@@ -436,7 +436,7 @@ Learn more details from [Invoke a stored procedure from a SQL sink](#invoke-a-st
436436
"type": "<source type>"
437437
},
438438
"sink": {
439-
"type": "SqlSink",
439+
"type": "AzureSqlSink",
440440
"sqlWriterStoredProcedureName": "CopyTestStoredProcedureWithParameters",
441441
"storedProcedureTableTypeParameterName": "MyTable",
442442
"sqlWriterTableType": "MyTableType",
@@ -550,7 +550,7 @@ The following sample shows how to use a stored procedure to do an upsert into a
550550

551551
```json
552552
"sink": {
553-
"type": "SqlSink",
553+
"type": "AzureSqlSink",
554554
"SqlWriterStoredProcedureName": "spOverwriteMarketing",
555555
"storedProcedureTableTypeParameterName": "Marketing",
556556
"SqlWriterTableType": "MarketingType",

articles/data-factory/connector-odbc.md

Lines changed: 1 addition & 75 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 11/19/2018
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -229,80 +229,6 @@ To copy data to ODBC-compatible data store, set the sink type in the copy activi
229229
]
230230
```
231231

232-
## IBM Informix source
233-
234-
You can copy data from IBM Informix database using the generic ODBC connector.
235-
236-
Set up a Self-hosted Integration Runtime on a machine with access to your data store. The Integration Runtime uses the ODBC driver for Informix to connect to the data store. Therefore, install the driver if it is not already installed on the same machine. For example, you can use driver "IBM INFORMIX ODBC DRIVER (64-bit)". See [Prerequisites](#prerequisites) section for details.
237-
238-
Before you use the Informix source in a Data Factory solution, verify whether the Integration Runtime can connect to the data store using instructions in [Troubleshoot connectivity issues](#troubleshoot-connectivity-issues) section.
239-
240-
Create an ODBC linked service to link an IBM Informix data store to an Azure data factory as shown in the following example:
241-
242-
```json
243-
{
244-
"name": "InformixLinkedService",
245-
"properties": {
246-
"type": "Odbc",
247-
"typeProperties": {
248-
"connectionString": {
249-
"type": "SecureString",
250-
"value": "<Informix connection string or DSN>"
251-
},
252-
"authenticationType": "Basic",
253-
"userName": "<username>",
254-
"password": {
255-
"type": "SecureString",
256-
"value": "<password>"
257-
}
258-
},
259-
"connectVia": {
260-
"referenceName": "<name of Integration Runtime>",
261-
"type": "IntegrationRuntimeReference"
262-
}
263-
}
264-
}
265-
```
266-
267-
Read the article from the beginning for a detailed overview of using ODBC data stores as source/sink data stores in a copy operation.
268-
269-
## Microsoft Access source
270-
271-
You can copy data from Microsoft Access database using the generic ODBC connector.
272-
273-
Set up a Self-hosted Integration Runtime on a machine with access to your data store. The Integration Runtime uses the ODBC driver for Microsoft Access to connect to the data store. Therefore, install the driver if it is not already installed on the same machine. See [Prerequisites](#prerequisites) section for details.
274-
275-
Before you use the Microsoft Access source in a Data Factory solution, verify whether the Integration Runtime can connect to the data store using instructions in [Troubleshoot connectivity issues](#troubleshoot-connectivity-issues) section.
276-
277-
Create an ODBC linked service to link a Microsoft Access database to an Azure data factory as shown in the following example:
278-
279-
```json
280-
{
281-
"name": "MicrosoftAccessLinkedService",
282-
"properties": {
283-
"type": "Odbc",
284-
"typeProperties": {
285-
"connectionString": {
286-
"type": "SecureString",
287-
"value": "Driver={Microsoft Access Driver (*.mdb, *.accdb)};Dbq=<path to your DB file e.g. C:\\mydatabase.accdb>;"
288-
},
289-
"authenticationType": "Basic",
290-
"userName": "<username>",
291-
"password": {
292-
"type": "SecureString",
293-
"value": "<password>"
294-
}
295-
},
296-
"connectVia": {
297-
"referenceName": "<name of Integration Runtime>",
298-
"type": "IntegrationRuntimeReference"
299-
}
300-
}
301-
}
302-
```
303-
304-
Read the article from the beginning for a detailed overview of using ODBC data stores as source/sink data stores in a copy operation.
305-
306232
## SAP HANA sink
307233

308234
>[!NOTE]

articles/data-factory/connector-oracle.md

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 06/25/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -30,11 +30,13 @@ You can copy data from an Oracle database to any supported sink data store. You
3030
Specifically, this Oracle connector supports:
3131

3232
- The following versions of an Oracle database:
33-
- Oracle 12c R1 (12.1)
34-
- Oracle 11g R1, R2 (11.1, 11.2)
35-
- Oracle 10g R1, R2 (10.1, 10.2)
36-
- Oracle 9i R1, R2 (9.0.1, 9.2)
37-
- Oracle 8i R3 (8.1.7)
33+
- Oracle 18c R1 (18.1) and higher
34+
- Oracle 12c R1 (12.1) and higher
35+
- Oracle 11g R1 (11.1) and higher
36+
- Oracle 10g R1 (10.1) and higher
37+
- Oracle 9i R2 (9.2) and higher
38+
- Oracle 8i R3 (8.1.7) and higher
39+
- Oracle Database Cloud Exadata Service
3840
- Copying data by using Basic or OID authentications.
3941
- Parallel copying from an Oracle source. See the [Parallel copy from Oracle](#parallel-copy-from-oracle) section for details.
4042

articles/data-factory/control-flow-get-metadata-activity.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ ms.workload: data-services
1313
ms.tgt_pltfrm: na
1414

1515
ms.topic: conceptual
16-
ms.date: 08/06/2019
16+
ms.date: 08/12/2019
1717
ms.author: jingwang
1818

1919
---
@@ -135,8 +135,8 @@ Property | Description | Required
135135
-------- | ----------- | --------
136136
fieldList | Lists the types of metadata information required. See details in [Metadata options](#metadata-options) section on supported metadata. | Yes
137137
dataset | The reference dataset whose metadata activity is to be retrieved by the GetMetadata Activity. See [Supported capabilities](#supported-capabilities) section on supported connectors, and refer to connector topic on dataset syntax details. | Yes
138-
formatSettings | Apply when using format type dataset (Parquet, DelimitedText). | No
139-
storeSettings | Apply when using format type dataset (Parquet, DelimitedText). | No
138+
formatSettings | Apply when using format type dataset. | No
139+
storeSettings | Apply when using format type dataset. | No
140140

141141
## Sample output
142142

articles/data-factory/copy-activity-overview.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -190,7 +190,7 @@ Copy activity execution details and performance characteristics are also returne
190190
| usedDataIntegrationUnits | The effective Data Integration Units during copy. | Int32 value |
191191
| usedParallelCopies | The effective parallelCopies during copy. | Int32 value |
192192
| redirectRowPath | Path to the log of skipped incompatible rows in the blob storage you configure under "redirectIncompatibleRowSettings". See below example. | Text (string) |
193-
| executionDetails | More details on the stages copy activity goes through, and the corresponding steps, duration, used configurations, etc. It's not recommended to parse this section as it may change.<br/><br/>ADF also reports the detailed durations (in seconds) spent on respective steps under `detailedDurations`:<br/>- **Queuing duration** (`queuingDuration`): The time until the copy activity actually starts on integration runtime. If you use Self-hosted IR and this value is large, suggest to check the IR capacity and usage, and scale up/out according to your workload. <br/>- **Pre-copy script duration** (`preCopyScriptDuration`): The time spent on executing the pre-copy script in sink data store. Apply when you configure the pre-copy script. <br/>- **Time-to-first-byte** (`timeToFirstByte`): The time that integration runtime receives the first byte from the source data store. Apply to non-file-based source. If this value is large, suggest to check and optimize the query or server.<br/>- **Transfer duration** (`transferDuration`): The time for integration runtime to transfer all the data from source to sink after getting the first byte. | Array |
193+
| executionDetails | More details on the stages copy activity goes through, and the corresponding steps, duration, used configurations, etc. It's not recommended to parse this section as it may change.<br/><br/>ADF also reports the detailed durations (in seconds) spent on respective steps under `detailedDurations`. The durations of these steps are exclusive and only those that apply to the given copy activity run would show up:<br/>- **Queuing duration** (`queuingDuration`): The elapsed time until the copy activity actually starts on the integration runtime. If you use Self-hosted IR and this value is large, suggest to check the IR capacity and usage, and scale up/out according to your workload. <br/>- **Pre-copy script duration** (`preCopyScriptDuration`): The elapsed time between copy activity starting on IR and copy activity finishing executing the pre-copy script in sink data store. Apply when you configure the pre-copy script. <br/>- **Time-to-first-byte** (`timeToFirstByte`): The elapsed time between the end of the previous step and the IR receiving the first byte from the source data store. Apply to non-file-based source. If this value is large, suggest to check and optimize the query or server.<br/>- **Transfer duration** (`transferDuration`): The elapsed time between the end of the previous step and the IR transferring all the data from source to sink. | Array |
194194
| perfRecommendation | Copy performance tuning tips. See [Performance and tuning](#performance-and-tuning) section on details. | Array |
195195

196196
```json

0 commit comments

Comments
 (0)