Skip to content

Commit a1553f3

Browse files
authored
Merge pull request #85117 from linda33wj/master
Update ADF connector articles
2 parents 29ef5d3 + 0c5b253 commit a1553f3

34 files changed

+183
-87
lines changed

articles/data-factory/connector-azure-database-for-mariadb.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ ms.service: data-factory
1010
ms.workload: data-services
1111
ms.tgt_pltfrm: na
1212
ms.topic: conceptual
13-
ms.date: 02/01/2019
13+
ms.date: 08/12/2019
1414
ms.author: jingwang
1515

1616
---
@@ -38,7 +38,7 @@ The following properties are supported for Azure Database for MariaDB linked ser
3838
|:--- |:--- |:--- |
3939
| type | The type property must be set to: **MariaDB** | Yes |
4040
| connectionString | A connection string to connect to Azure Database for MariaDB. You can find it from the Azure portal -> your Azure Database for MariaDB -> Connection strings -> ADO.NET one. <br/>Mark this field as a SecureString to store it securely in Data Factory. You can also put password in Azure Key Vault and pull the `pwd` configuration out of the connection string. Refer to the following samples and [Store credentials in Azure Key Vault](store-credentials-in-key-vault.md) article with more details. | Yes |
41-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Self-hosted Integration Runtime or Azure Integration Runtime (if your data store is publicly accessible). If not specified, it uses the default Azure Integration Runtime. |No |
41+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. If not specified, it uses the default Azure Integration Runtime. |No |
4242

4343
**Example:**
4444

articles/data-factory/connector-cassandra.md

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 08/01/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -37,7 +37,9 @@ Specifically, this Cassandra connector supports:
3737
3838
## Prerequisites
3939

40-
To copy data from a Cassandra database that is not publicly accessible, you need to set up a Self-hosted Integration Runtime. See [Self-hosted Integration Runtime](create-self-hosted-integration-runtime.md) article to learn details. The Integration Runtime provides a built-in Cassandra driver, therefore you don't need to manually install any driver when copying data from/to Cassandra.
40+
[!INCLUDE [data-factory-v2-integration-runtime-requirements](../../includes/data-factory-v2-integration-runtime-requirements.md)]
41+
42+
The Integration Runtime provides a built-in Cassandra driver, therefore you don't need to manually install any driver when copying data from/to Cassandra.
4143

4244
## Getting started
4345

@@ -57,7 +59,7 @@ The following properties are supported for Cassandra linked service:
5759
| authenticationType | Type of authentication used to connect to the Cassandra database.<br/>Allowed values are: **Basic**, and **Anonymous**. |Yes |
5860
| username |Specify user name for the user account. |Yes, if authenticationType is set to Basic. |
5961
| password |Specify password for the user account. Mark this field as a SecureString to store it securely in Data Factory, or [reference a secret stored in Azure Key Vault](store-credentials-in-key-vault.md). |Yes, if authenticationType is set to Basic. |
60-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Self-hosted Integration Runtime or Azure Integration Runtime (if your data store is publicly accessible). If not specified, it uses the default Azure Integration Runtime. |No |
62+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. Learn more from [Prerequisites](#prerequisites) section. If not specified, it uses the default Azure Integration Runtime. |No |
6163

6264
>[!NOTE]
6365
>Currently connection to Cassandra using SSL is not supported.

articles/data-factory/connector-couchbase.md

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 08/01/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -29,6 +29,10 @@ You can copy data from Couchbase to any supported sink data store. For a list of
2929

3030
Azure Data Factory provides a built-in driver to enable connectivity, therefore you don't need to manually install any driver using this connector.
3131

32+
## Prerequisites
33+
34+
[!INCLUDE [data-factory-v2-integration-runtime-requirements](../../includes/data-factory-v2-integration-runtime-requirements.md)]
35+
3236
## Getting started
3337

3438
[!INCLUDE [data-factory-v2-connector-get-started](../../includes/data-factory-v2-connector-get-started.md)]
@@ -43,7 +47,7 @@ The following properties are supported for Couchbase linked service:
4347
|:--- |:--- |:--- |
4448
| type | The type property must be set to: **Couchbase** | Yes |
4549
| connectionString | An ODBC connection string to connect to Couchbase. <br/>Mark this field as a SecureString to store it securely in Data Factory. You can also put credential string in Azure Key Vault and pull the `credString` configuration out of the connection string. Refer to the following samples and [Store credentials in Azure Key Vault](store-credentials-in-key-vault.md) article with more details. | Yes |
46-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Self-hosted Integration Runtime or Azure Integration Runtime (if your data store is publicly accessible). If not specified, it uses the default Azure Integration Runtime. |No |
50+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. Learn more from [Prerequisites](#prerequisites) section. If not specified, it uses the default Azure Integration Runtime. |No |
4751

4852
**Example:**
4953

articles/data-factory/connector-db2.md

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 08/17/2018
15+
ms.date: 08/12/2019
1616

1717
ms.author: jingwang
1818

@@ -46,7 +46,9 @@ Specifically, this DB2 connector supports the following IBM DB2 platforms and ve
4646
4747
## Prerequisites
4848

49-
To use copy data from a DB2 database that is not publicly accessible, you need to set up a Self-hosted Integration Runtime. To learn about Self-hosted integration runtimes, see [Self-hosted Integration Runtime](create-self-hosted-integration-runtime.md) article. The Integration Runtime provides a built-in DB2 driver, therefore you don't need to manually install any driver when copying data from DB2.
49+
[!INCLUDE [data-factory-v2-integration-runtime-requirements](../../includes/data-factory-v2-integration-runtime-requirements.md)]
50+
51+
The Integration Runtime provides a built-in DB2 driver, therefore you don't need to manually install any driver when copying data from DB2.
5052

5153
## Getting started
5254

@@ -66,7 +68,7 @@ The following properties are supported for DB2 linked service:
6668
| authenticationType |Type of authentication used to connect to the DB2 database.<br/>Allowed value is: **Basic**. |Yes |
6769
| username |Specify user name to connect to the DB2 database. |Yes |
6870
| password |Specify password for the user account you specified for the username. Mark this field as a SecureString to store it securely in Data Factory, or [reference a secret stored in Azure Key Vault](store-credentials-in-key-vault.md). |Yes |
69-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Self-hosted Integration Runtime or Azure Integration Runtime (if your data store is publicly accessible). If not specified, it uses the default Azure Integration Runtime. |No |
71+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. Learn more from [Prerequisites](#prerequisites) section. If not specified, it uses the default Azure Integration Runtime. |No |
7072

7173
**Example:**
7274

articles/data-factory/connector-drill.md

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 02/01/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -29,6 +29,10 @@ You can copy data from Drill to any supported sink data store. For a list of dat
2929

3030
Azure Data Factory provides a built-in driver to enable connectivity, therefore you don't need to manually install any driver using this connector.
3131

32+
## Prerequisites
33+
34+
[!INCLUDE [data-factory-v2-integration-runtime-requirements](../../includes/data-factory-v2-integration-runtime-requirements.md)]
35+
3236
## Getting started
3337

3438
[!INCLUDE [data-factory-v2-connector-get-started](../../includes/data-factory-v2-connector-get-started.md)]
@@ -43,7 +47,7 @@ The following properties are supported for Drill linked service:
4347
|:--- |:--- |:--- |
4448
| type | The type property must be set to: **Drill** | Yes |
4549
| connectionString | An ODBC connection string to connect to Drill. <br/>Mark this field as a SecureString to store it securely in Data Factory. You can also put password in Azure Key Vault and pull the `pwd` configuration out of the connection string. Refer to the following samples and [Store credentials in Azure Key Vault](store-credentials-in-key-vault.md) article with more details. | Yes |
46-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Self-hosted Integration Runtime or Azure Integration Runtime (if your data store is publicly accessible). If not specified, it uses the default Azure Integration Runtime. |No |
50+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. Learn more from [Prerequisites](#prerequisites) section. If not specified, it uses the default Azure Integration Runtime. |No |
4751

4852
**Example:**
4953

articles/data-factory/connector-file-system.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 08/06/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -40,7 +40,7 @@ Specifically, this file system connector supports:
4040

4141
## Prerequisites
4242

43-
To copy data from/to a file system that is not publicly accessible, you need to set up a Self-hosted Integration Runtime. See [Self-hosted Integration Runtime](create-self-hosted-integration-runtime.md) article for details.
43+
[!INCLUDE [data-factory-v2-integration-runtime-requirements](../../includes/data-factory-v2-integration-runtime-requirements.md)]
4444

4545
## Getting started
4646

@@ -58,7 +58,7 @@ The following properties are supported for file system linked service:
5858
| host | Specifies the root path of the folder that you want to copy. Use the escape character "\" for special characters in the string. See [Sample linked service and dataset definitions](#sample-linked-service-and-dataset-definitions) for examples. | Yes |
5959
| userid | Specify the ID of the user who has access to the server. | Yes |
6060
| password | Specify the password for the user (userid). Mark this field as a SecureString to store it securely in Data Factory, or [reference a secret stored in Azure Key Vault](store-credentials-in-key-vault.md). | Yes |
61-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Self-hosted Integration Runtime or Azure Integration Runtime (if your data store is publicly accessible). If not specified, it uses the default Azure Integration Runtime. |No |
61+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. Learn more from [Prerequisites](#prerequisites) section. If not specified, it uses the default Azure Integration Runtime. |No |
6262

6363
### Sample linked service and dataset definitions
6464

articles/data-factory/connector-ftp.md

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 08/06/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -37,6 +37,10 @@ Specifically, this FTP connector supports:
3737
- Copying files using **Basic** or **Anonymous** authentication.
3838
- Copying files as-is or parsing files with the [supported file formats and compression codecs](supported-file-formats-and-compression-codecs.md).
3939

40+
## Prerequisites
41+
42+
[!INCLUDE [data-factory-v2-integration-runtime-requirements](../../includes/data-factory-v2-integration-runtime-requirements.md)]
43+
4044
## Get started
4145

4246
[!INCLUDE [data-factory-v2-connector-get-started](../../includes/data-factory-v2-connector-get-started.md)]
@@ -57,7 +61,7 @@ The following properties are supported for FTP linked service:
5761
| authenticationType | Specify the authentication type.<br/>Allowed values are: **Basic**, **Anonymous** | Yes |
5862
| userName | Specify the user who has access to the FTP server. | No |
5963
| password | Specify the password for the user (userName). Mark this field as a SecureString to store it securely in Data Factory, or [reference a secret stored in Azure Key Vault](store-credentials-in-key-vault.md). | No |
60-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Azure Integration Runtime or Self-hosted Integration Runtime (if your data store is located in private network). If not specified, it uses the default Azure Integration Runtime. |No |
64+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. Learn more from [Prerequisites](#prerequisites) section. If not specified, it uses the default Azure Integration Runtime. |No |
6165

6266
>[!NOTE]
6367
>The FTP connector supports accessing FTP server with either no encryption or explicit SSL/TLS encryption; it doesn’t support implicit SSL/TLS encryption.

articles/data-factory/connector-greenplum.md

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 02/01/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -26,6 +26,10 @@ You can copy data from Greenplum to any supported sink data store. For a list of
2626

2727
Azure Data Factory provides a built-in driver to enable connectivity, therefore you don't need to manually install any driver using this connector.
2828

29+
## Prerequisites
30+
31+
[!INCLUDE [data-factory-v2-integration-runtime-requirements](../../includes/data-factory-v2-integration-runtime-requirements.md)]
32+
2933
## Getting started
3034

3135
[!INCLUDE [data-factory-v2-connector-get-started](../../includes/data-factory-v2-connector-get-started.md)]
@@ -40,7 +44,7 @@ The following properties are supported for Greenplum linked service:
4044
|:--- |:--- |:--- |
4145
| type | The type property must be set to: **Greenplum** | Yes |
4246
| connectionString | An ODBC connection string to connect to Greenplum. <br/>Mark this field as a SecureString to store it securely in Data Factory. You can also put password in Azure Key Vault and pull the `pwd` configuration out of the connection string. Refer to the following samples and [Store credentials in Azure Key Vault](store-credentials-in-key-vault.md) article with more details. | Yes |
43-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Self-hosted Integration Runtime or Azure Integration Runtime (if your data store is publicly accessible). If not specified, it uses the default Azure Integration Runtime. |No |
47+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. Learn more from [Prerequisites](#prerequisites) section. If not specified, it uses the default Azure Integration Runtime. |No |
4448

4549
**Example:**
4650

articles/data-factory/connector-hbase.md

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 08/01/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -26,6 +26,10 @@ You can copy data from HBase to any supported sink data store. For a list of dat
2626

2727
Azure Data Factory provides a built-in driver to enable connectivity, therefore you don't need to manually install any driver using this connector.
2828

29+
## Prerequisites
30+
31+
[!INCLUDE [data-factory-v2-integration-runtime-requirements](../../includes/data-factory-v2-integration-runtime-requirements.md)]
32+
2933
## Getting started
3034

3135
[!INCLUDE [data-factory-v2-connector-get-started](../../includes/data-factory-v2-connector-get-started.md)]
@@ -49,7 +53,7 @@ The following properties are supported for HBase linked service:
4953
| trustedCertPath | The full path of the .pem file containing trusted CA certificates for verifying the server when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The default value is the cacerts.pem file installed with the IR. | No |
5054
| allowHostNameCNMismatch | Specifies whether to require a CA-issued SSL certificate name to match the host name of the server when connecting over SSL. The default value is false. | No |
5155
| allowSelfSignedServerCert | Specifies whether to allow self-signed certificates from the server. The default value is false. | No |
52-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Self-hosted Integration Runtime or Azure Integration Runtime (if your data store is publicly accessible). If not specified, it uses the default Azure Integration Runtime. |No |
56+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. Learn more from [Prerequisites](#prerequisites) section. If not specified, it uses the default Azure Integration Runtime. |No |
5357

5458
>[!NOTE]
5559
>If your cluster doesn't support sticky session e.g. HDInsight, explicitly add node index at the end of the http path setting, e.g. specify `/hbaserest0` instead of `/hbaserest`.

articles/data-factory/connector-hdfs.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ms.workload: data-services
1212
ms.tgt_pltfrm: na
1313

1414
ms.topic: conceptual
15-
ms.date: 08/06/2019
15+
ms.date: 08/12/2019
1616
ms.author: jingwang
1717

1818
---
@@ -38,7 +38,7 @@ Specifically, this HDFS connector supports:
3838

3939
## Prerequisites
4040

41-
To copy data from an HDFS that is not publicly accessible, you need to set up a Self-hosted Integration Runtime. See [Self-hosted Integration Runtime](concepts-integration-runtime.md) article to learn details.
41+
[!INCLUDE [data-factory-v2-integration-runtime-requirements](../../includes/data-factory-v2-integration-runtime-requirements.md)]
4242

4343
> [!NOTE]
4444
> Make sure the Integration Runtime can access to **ALL** the [name node server]:[name node port] and [data node servers]:[data node port] of the Hadoop cluster. Default [name node port] is 50070, and default [data node port] is 50075.
@@ -60,7 +60,7 @@ The following properties are supported for HDFS linked service:
6060
| authenticationType | Allowed values are: **Anonymous**, or **Windows**. <br><br> To use **Kerberos authentication** for HDFS connector, refer to [this section](#use-kerberos-authentication-for-hdfs-connector) to set up your on-premises environment accordingly. |Yes |
6161
| userName |Username for Windows authentication. For Kerberos authentication, specify `<username>@<domain>.com`. |Yes (for Windows Authentication) |
6262
| password |Password for Windows authentication. Mark this field as a SecureString to store it securely in Data Factory, or [reference a secret stored in Azure Key Vault](store-credentials-in-key-vault.md). |Yes (for Windows Authentication) |
63-
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. You can use Self-hosted Integration Runtime or Azure Integration Runtime (if your data store is publicly accessible). If not specified, it uses the default Azure Integration Runtime. |No |
63+
| connectVia | The [Integration Runtime](concepts-integration-runtime.md) to be used to connect to the data store. Learn more from [Prerequisites](#prerequisites) section. If not specified, it uses the default Azure Integration Runtime. |No |
6464

6565
**Example: using Anonymous authentication**
6666

0 commit comments

Comments
 (0)