You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/data-factory/connector-azure-blob-storage.md
+4-1Lines changed: 4 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -8,7 +8,7 @@ ms.reviewer: craigg
8
8
ms.service: data-factory
9
9
ms.workload: data-services
10
10
ms.topic: conceptual
11
-
ms.date: 08/06/2019
11
+
ms.date: 08/12/2019
12
12
ms.author: jingwang
13
13
14
14
---
@@ -71,6 +71,9 @@ To use storage account key authentication, the following properties are supporte
71
71
| connectionString | Specify the information needed to connect to Storage for the connectionString property. <br/>Mark this field as a SecureString to store it securely in Data Factory. You can also put account key in Azure Key Vault and pull the `accountKey` configuration out of the connection string. Refer to the following samples and [Store credentials in Azure Key Vault](store-credentials-in-key-vault.md) article with more details. |Yes |
72
72
| connectVia | The [integration runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Azure Integration Runtime or Self-hosted Integration Runtime (if your data store is in a private network). If not specified, it uses the default Azure Integration Runtime. |No |
73
73
74
+
>[!NOTE]
75
+
>Secondary Blob Service Endpoint is not supported when using account key authentication. You can use other authentication types.
76
+
74
77
>[!NOTE]
75
78
>If you were using "AzureStorage" type linked service, it is still supported as-is, while you are suggested to use this new "AzureBlobStorage" linked service type going forward.
Copy file name to clipboardExpand all lines: articles/data-factory/connector-azure-data-lake-storage.md
+5-2Lines changed: 5 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -9,13 +9,13 @@ ms.reviewer: douglasl
9
9
ms.service: data-factory
10
10
ms.workload: data-services
11
11
ms.topic: conceptual
12
-
ms.date: 08/06/2019
12
+
ms.date: 08/12/2019
13
13
ms.author: jingwang
14
14
15
15
---
16
16
# Copy data to or from Azure Data Lake Storage Gen2 using Azure Data Factory
17
17
18
-
Azure Data Lake Storage Gen2 is a set of capabilities dedicated to big data analytics built into [Azure Blob storage](../storage/blobs/storage-blobs-introduction.md). You can use it to interface with your data by using both file system and object storage paradigms.
18
+
Azure Data Lake Storage Gen2 (ADLS Gen2) is a set of capabilities dedicated to big data analytics built into [Azure Blob storage](../storage/blobs/storage-blobs-introduction.md). You can use it to interface with your data by using both file system and object storage paradigms.
19
19
20
20
This article outlines how to copy data to and from Azure Data Lake Storage Gen2. To learn about Azure Data Factory, read the [introductory article](introduction.md).
21
21
@@ -70,6 +70,9 @@ To use storage account key authentication, the following properties are supporte
70
70
| accountKey | Account key for Data Lake Storage Gen2. Mark this field as a SecureString to store it securely in Data Factory, or [reference a secret stored in Azure Key Vault](store-credentials-in-key-vault.md). |Yes |
71
71
| connectVia | The [integration runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use the Azure integration runtime or a self-hosted integration runtime if your data store is in a private network. If this property isn't specified, the default Azure integration runtime is used. |No |
72
72
73
+
>[!NOTE]
74
+
>Secondary ADLS file system endpoint is not supported when using account key authentication. You can use other authentication types.
Copy file name to clipboardExpand all lines: articles/data-factory/connector-azure-sql-database.md
+11-11Lines changed: 11 additions & 11 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
12
12
ms.tgt_pltfrm: na
13
13
14
14
ms.topic: conceptual
15
-
ms.date: 08/06/2019
15
+
ms.date: 08/12/2019
16
16
ms.author: jingwang
17
17
18
18
---
@@ -259,18 +259,18 @@ For a full list of sections and properties available for defining activities, se
259
259
260
260
### Azure SQL Database as the source
261
261
262
-
To copy data from Azure SQL Database, setthe**type** property in the copy activity source to **SqlSource**. The following properties are supported in the copy activity **source** section:
262
+
To copy data from Azure SQL Database, the following properties are supported in the copy activity **source** section:
263
263
264
264
| Property | Description | Required |
265
265
|:--- |:--- |:--- |
266
-
| type | The **type** property of the copy activity source must be set to **SqlSource**. | Yes |
266
+
| type | The **type** property of the copy activity source must be set to **AzureSqlSource**. "SqlSource" type is still supported for backward compatibility. | Yes |
267
267
| sqlReaderQuery | This property uses the custom SQL query to read data. An example is `select * from MyTable`. | No |
268
268
| sqlReaderStoredProcedureName | The name of the stored procedure that reads data from the source table. The last SQL statement must be a SELECT statement in the stored procedure. | No |
269
269
| storedProcedureParameters | Parameters for the stored procedure.<br/>Allowed values are name or value pairs. The names and casing of parameters must match the names and casing of the stored procedure parameters. | No |
270
270
271
271
**Points to note:**
272
272
273
-
- If **sqlReaderQuery** is specified for **SqlSource**, the copy activity runs this query against the Azure SQL Database source to get the data. You also can specify a stored procedure by specifying **sqlReaderStoredProcedureName**and**storedProcedureParameters** if the stored procedure takes parameters.
273
+
- If **sqlReaderQuery** is specified for **AzureSqlSource**, the copy activity runs this query against the Azure SQL Database source to get the data. You also can specify a stored procedure by specifying **sqlReaderStoredProcedureName**and**storedProcedureParameters** if the stored procedure takes parameters.
274
274
- If you don't specify either **sqlReaderQuery** or **sqlReaderStoredProcedureName**, the columns defined in the "structure" section of the dataset JSON are used to construct a query. The query `select column1, column2 from mytable` runs against Azure SQL Database. If the dataset definition doesn't have "structure," all columns are selected from the table.
275
275
276
276
#### SQL query example
@@ -294,7 +294,7 @@ To copy data from Azure SQL Database, set the **type** property in the copy acti
294
294
],
295
295
"typeProperties": {
296
296
"source": {
297
-
"type": "SqlSource",
297
+
"type": "AzureSqlSource",
298
298
"sqlReaderQuery": "SELECT * FROM MyTable"
299
299
},
300
300
"sink": {
@@ -326,7 +326,7 @@ To copy data from Azure SQL Database, set the **type** property in the copy acti
> Learn more about the supported write behaviors, configurations, and best practices from [Best practice for loading data into Azure SQL Database](#best-practice-for-loading-data-into-azure-sql-database).
367
367
368
-
To copy data to Azure SQL Database, setthe**type** property in the copy activity sink to **SqlSink**. The following properties are supported in the copy activity **sink** section:
368
+
To copy data to Azure SQL Database, the following properties are supported in the copy activity **sink** section:
369
369
370
370
| Property | Description | Required |
371
371
|:--- |:--- |:--- |
372
-
| type | The **type** property of the copy activity sink must be set to **SqlSink**. | Yes |
372
+
| type | The **type** property of the copy activity sink must be set to **AzureSqlSink**. "SqlSink" type is still supported for backward compatibility. | Yes |
373
373
| writeBatchSize | Number of rows to insert into the SQL table *per batch*.<br/> The allowed value is **integer** (number of rows). By default, Azure Data Factory dynamically determines the appropriate batch size based on the row size. | No |
374
374
| writeBatchTimeout | The wait time for the batch insert operation to finish before it times out.<br/> The allowed value is **timespan**. An example is “00:30:00” (30 minutes). | No |
375
375
| preCopyScript | Specify a SQL query for the copy activity to run before writing data into Azure SQL Database. It's invoked only once per copy run. Use this property to clean up the preloaded data. | No |
@@ -402,7 +402,7 @@ To copy data to Azure SQL Database, set the **type** property in the copy activi
402
402
"type": "<source type>"
403
403
},
404
404
"sink": {
405
-
"type": "SqlSink",
405
+
"type": "AzureSqlSink",
406
406
"writeBatchSize": 100000
407
407
}
408
408
}
@@ -436,7 +436,7 @@ Learn more details from [Invoke a stored procedure from a SQL sink](#invoke-a-st
Copy file name to clipboardExpand all lines: articles/data-factory/connector-odbc.md
+1-75Lines changed: 1 addition & 75 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
12
12
ms.tgt_pltfrm: na
13
13
14
14
ms.topic: conceptual
15
-
ms.date: 11/19/2018
15
+
ms.date: 08/12/2019
16
16
ms.author: jingwang
17
17
18
18
---
@@ -229,80 +229,6 @@ To copy data to ODBC-compatible data store, set the sink type in the copy activi
229
229
]
230
230
```
231
231
232
-
## IBM Informix source
233
-
234
-
You can copy data from IBM Informix database using the generic ODBC connector.
235
-
236
-
Set up a Self-hosted Integration Runtime on a machine with access to your data store. The Integration Runtime uses the ODBC driver for Informix to connect to the data store. Therefore, install the driver if it is not already installed on the same machine. For example, you can use driver "IBM INFORMIX ODBC DRIVER (64-bit)". See [Prerequisites](#prerequisites) section for details.
237
-
238
-
Before you use the Informix source in a Data Factory solution, verify whether the Integration Runtime can connect to the data store using instructions in [Troubleshoot connectivity issues](#troubleshoot-connectivity-issues) section.
239
-
240
-
Create an ODBC linked service to link an IBM Informix data store to an Azure data factory as shown in the following example:
241
-
242
-
```json
243
-
{
244
-
"name": "InformixLinkedService",
245
-
"properties": {
246
-
"type": "Odbc",
247
-
"typeProperties": {
248
-
"connectionString": {
249
-
"type": "SecureString",
250
-
"value": "<Informix connection string or DSN>"
251
-
},
252
-
"authenticationType": "Basic",
253
-
"userName": "<username>",
254
-
"password": {
255
-
"type": "SecureString",
256
-
"value": "<password>"
257
-
}
258
-
},
259
-
"connectVia": {
260
-
"referenceName": "<name of Integration Runtime>",
261
-
"type": "IntegrationRuntimeReference"
262
-
}
263
-
}
264
-
}
265
-
```
266
-
267
-
Read the article from the beginning for a detailed overview of using ODBC data stores as source/sink data stores in a copy operation.
268
-
269
-
## Microsoft Access source
270
-
271
-
You can copy data from Microsoft Access database using the generic ODBC connector.
272
-
273
-
Set up a Self-hosted Integration Runtime on a machine with access to your data store. The Integration Runtime uses the ODBC driver for Microsoft Access to connect to the data store. Therefore, install the driver if it is not already installed on the same machine. See [Prerequisites](#prerequisites) section for details.
274
-
275
-
Before you use the Microsoft Access source in a Data Factory solution, verify whether the Integration Runtime can connect to the data store using instructions in [Troubleshoot connectivity issues](#troubleshoot-connectivity-issues) section.
276
-
277
-
Create an ODBC linked service to link a Microsoft Access database to an Azure data factory as shown in the following example:
278
-
279
-
```json
280
-
{
281
-
"name": "MicrosoftAccessLinkedService",
282
-
"properties": {
283
-
"type": "Odbc",
284
-
"typeProperties": {
285
-
"connectionString": {
286
-
"type": "SecureString",
287
-
"value": "Driver={Microsoft Access Driver (*.mdb, *.accdb)};Dbq=<path to your DB file e.g. C:\\mydatabase.accdb>;"
288
-
},
289
-
"authenticationType": "Basic",
290
-
"userName": "<username>",
291
-
"password": {
292
-
"type": "SecureString",
293
-
"value": "<password>"
294
-
}
295
-
},
296
-
"connectVia": {
297
-
"referenceName": "<name of Integration Runtime>",
298
-
"type": "IntegrationRuntimeReference"
299
-
}
300
-
}
301
-
}
302
-
```
303
-
304
-
Read the article from the beginning for a detailed overview of using ODBC data stores as source/sink data stores in a copy operation.
fieldList | Lists the types of metadata information required. See details in [Metadata options](#metadata-options) section on supported metadata. | Yes
137
137
dataset | The reference dataset whose metadata activity is to be retrieved by the GetMetadata Activity. See [Supported capabilities](#supported-capabilities) section on supported connectors, and refer to connector topic on dataset syntax details. | Yes
138
-
formatSettings | Apply when using format type dataset (Parquet, DelimitedText). | No
139
-
storeSettings | Apply when using format type dataset (Parquet, DelimitedText). | No
138
+
formatSettings | Apply when using format type dataset. | No
139
+
storeSettings | Apply when using format type dataset. | No
Copy file name to clipboardExpand all lines: articles/data-factory/copy-activity-overview.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -190,7 +190,7 @@ Copy activity execution details and performance characteristics are also returne
190
190
| usedDataIntegrationUnits | The effective Data Integration Units during copy. | Int32 value |
191
191
| usedParallelCopies | The effective parallelCopies during copy. | Int32 value |
192
192
| redirectRowPath | Path to the log of skipped incompatible rows in the blob storage you configure under "redirectIncompatibleRowSettings". See below example. | Text (string) |
193
-
| executionDetails | More details on the stages copy activity goes through, and the corresponding steps, duration, used configurations, etc. It's not recommended to parse this section as it may change.<br/><br/>ADF also reports the detailed durations (in seconds) spent on respective steps under `detailedDurations`:<br/>- **Queuing duration** (`queuingDuration`): The time until the copy activity actually starts on integration runtime. If you use Self-hosted IR and this value is large, suggest to check the IR capacity and usage, and scale up/out according to your workload. <br/>- **Pre-copy script duration** (`preCopyScriptDuration`): The time spent on executing the pre-copy script in sink data store. Apply when you configure the pre-copy script. <br/>- **Time-to-first-byte** (`timeToFirstByte`): The time that integration runtime receives the first byte from the source data store. Apply to non-file-based source. If this value is large, suggest to check and optimize the query or server.<br/>- **Transfer duration** (`transferDuration`): The time for integration runtime to transfer all the data from source to sink after getting the first byte. | Array |
193
+
| executionDetails | More details on the stages copy activity goes through, and the corresponding steps, duration, used configurations, etc. It's not recommended to parse this section as it may change.<br/><br/>ADF also reports the detailed durations (in seconds) spent on respective steps under `detailedDurations`. The durations of these steps are exclusive and only those that apply to the given copy activity run would show up:<br/>- **Queuing duration** (`queuingDuration`): The elapsed time until the copy activity actually starts on the integration runtime. If you use Self-hosted IR and this value is large, suggest to check the IR capacity and usage, and scale up/out according to your workload. <br/>- **Pre-copy script duration** (`preCopyScriptDuration`): The elapsed time between copy activity starting on IR and copy activity finishing executing the pre-copy script in sink data store. Apply when you configure the pre-copy script. <br/>- **Time-to-first-byte** (`timeToFirstByte`): The elapsed time between the end of the previous step and the IR receiving the first byte from the source data store. Apply to non-file-based source. If this value is large, suggest to check and optimize the query or server.<br/>- **Transfer duration** (`transferDuration`): The elapsed time between the end of the previous step and the IR transferring all the data from source to sink. | Array |
194
194
| perfRecommendation | Copy performance tuning tips. See [Performance and tuning](#performance-and-tuning) section on details. | Array |
0 commit comments