Skip to content

Commit e60dd45

Browse files
Merge pull request #303885 from Clare-Zheng82/0806-Update_auto_upgrade_table
[Doc update] Update auto upgrade table
2 parents ed0c56b + 26e5981 commit e60dd45

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

articles/data-factory/connector-upgrade-guidance.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.topic: concept-article
99
ms.custom:
1010
- references_regions
1111
- build-2025
12-
ms.date: 07/31/2025
12+
ms.date: 08/06/2025
1313
---
1414

1515
# Connector upgrade guidance
@@ -107,10 +107,12 @@ You can find more details from the table below on the connector list that is pla
107107

108108
| Connector | Scenario |
109109
|------------------|----------|
110+
| [Amazon RDS for Oracle](connector-amazon-rds-for-oracle.md) | Scenario that doesn't rely on capability below in Oracle (version 1.0):<br><br>• Use procedureRetResults, batchFailureReturnsError, truststore, and truststorepassword as connection properties.<br>• Use query ends with the semicolon.<br><br>If your pipeline runs on self-hosted integration runtime, it requires SHIR version 5.56 or above. |
110111
| [Amazon Redshift](connector-amazon-redshift.md) | Scenario that doesn't rely on below capability in Amazon Redshift (version 1.0):<br><br>• Linked service that uses Azure integration runtime.<br>• Use [UNLOAD](connector-amazon-redshift.md#use-unload-to-copy-data-from-amazon-redshift).<br><br>Automatic upgrade is only applicable when the driver is installed in your machine that installs the self-hosted integration runtime.<br><br> For more information, go to [Install Amazon Redshift ODBC driver for the version 2.0](connector-amazon-redshift.md#install-amazon-redshift-odbc-driver-for-the-version-20).|
111112
| [Google BigQuery](connector-google-bigquery.md) | Scenario that doesn't rely on below capability in Google BigQuery V1:<br><br> • Use `trustedCertsPath`, `additionalProjects`, `requestgoogledrivescope` connection properties.<br> • Set `useSystemTrustStore` connection property as `false`.<br> • Use **STRUCT** and **ARRAY** data types. <br><br>If your pipeline runs on self-hosted integration runtime, it requires SHIR version 5.55 or above. |
112113
| [Hive](connector-hive.md) | Scenario that doesn't rely on below capability in Hive (version 1.0):<br><br>• Authentication types:<br>&nbsp;&nbsp;• Username<br>• Thrift transport protocol:<br>&nbsp;&nbsp;• HiveServer1<br>• Service discovery mode: True<br>• Use native query: True <br><br>If your pipeline runs on self-hosted integration runtime, it requires SHIR version 5.55 or above.|
113114
| [Impala](connector-impala.md) | Scenario that doesn't rely on below capability in Impala (version 1.0):<br><br>• Authentication types:<br>&nbsp;&nbsp;• SASL Username<br><br>If your pipeline runs on self-hosted integration runtime, it requires SHIR version 5.55 or above. |
115+
| [Oracle](connector-oracle.md) | Scenario that doesn't rely on capability below in Oracle (version 1.0):<br><br>• Use procedureRetResults, batchFailureReturnsError, truststore, and truststorepassword as connection properties.<br>• Use Oracle connector as sink.<br>• Use query ends with the semicolon.<br>• Use PL/SQL command in Script activity<br>• Use script parameters in Script activity<br><br>If your pipeline runs on self-hosted integration runtime, it requires SHIR version 5.56 or above. |
114116
| [Salesforce](connector-salesforce.md) | Scenario that doesn't rely on capability below in Salesforce V1:<br><br>• SOQL queries that use:<br>&nbsp;&nbsp;• TYPEOF clauses<br>&nbsp;&nbsp;• Compound address/geolocations fields<br>• All SQL-92 query<br>• Report query {call "\<report name>"}<br>• Use Self-hosted integration runtime (To be supported) |
115117
| [Salesforce Service Cloud](connector-salesforce-service-cloud.md) | Scenario that doesn't rely on capability below in Salesforce Service Cloud V1:<br><br>• SOQL queries that use:<br>&nbsp;&nbsp;• TYPEOF clauses<br>&nbsp;&nbsp;• Compound address/geolocations fields<br>• All SQL-92 query<br>• Report query {call "\<report name>"}<br>• Use Self-hosted integration runtime (To be supported) |
116118
| [Spark](connector-spark.md) | Scenario that doesn't rely on below capability in Spark (version 1.0):<br><br>• Authentication types:<br>&nbsp;&nbsp;• Username<br>• Thrift transport protocol:<br>&nbsp;&nbsp;• SASL<br>&nbsp;&nbsp;• Binary<br>• Thrift transport protocol:<br>&nbsp;&nbsp;• SharkServer<br>&nbsp;&nbsp;• SharkServer2<br><br>If your pipeline runs on self-hosted integration runtime, it requires SHIR version 5.55 or above.|

0 commit comments

Comments
 (0)