Skip to content

Commit 6a22c19

Browse files
authored
acrolinx
1 parent 599b466 commit 6a22c19

File tree

1 file changed

+14
-14
lines changed

1 file changed

+14
-14
lines changed

articles/data-factory/connector-snowflake-legacy.md

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
---
22
title: Copy and transform data in Snowflake using legacy
3-
titleSuffix: Azure Data Factory & Azure Synapse
3+
titleSuffix: Azure Data Factory and Azure Synapse
44
description: Learn how to copy and transform data in Snowflake using legacy Data Factory or Azure Synapse Analytics.
55
ms.author: jianleishen
66
author: jianleishen
@@ -198,7 +198,7 @@ To copy data from Snowflake, the following properties are supported in the Copy
198198
| Property | Description | Required |
199199
| :--------------------------- | :----------------------------------------------------------- | :------- |
200200
| type | The type property of the Copy activity source must be set to **SnowflakeSource**. | Yes |
201-
| query | Specifies the SQL query to read data from Snowflake. If the names of the schema, table and columns contain lower case, quote the object identifier in query e.g. `select * from "schema"."myTable"`.<br>Executing stored procedure is not supported. | No |
201+
| query | Specifies the SQL query to read data from Snowflake. If the names of the schema, table and columns contain lower case, quote the object identifier in query e.g. `select * from "schema"."myTable"`.<br>Executing stored procedure isn't supported. | No |
202202
| exportSettings | Advanced settings used to retrieve data from Snowflake. You can configure the ones supported by the COPY into command that the service will pass through when you invoke the statement. | Yes |
203203
| ***Under `exportSettings`:*** | | |
204204
| type | The type of export command, set to **SnowflakeExportCopyCommand**. | Yes |
@@ -212,7 +212,7 @@ To copy data from Snowflake, the following properties are supported in the Copy
212212
213213
#### Direct copy from Snowflake
214214

215-
If your sink data store and format meet the criteria described in this section, you can use the Copy activity to directly copy from Snowflake to sink. The service checks the settings and fails the Copy activity run if the following criteria is not met:
215+
If your sink data store and format meet the criteria described in this section, you can use the Copy activity to directly copy from Snowflake to sink. The service checks the settings and fails the Copy activity run if the following criteria isn't met:
216216

217217
- The **sink linked service** is [**Azure Blob storage**](connector-azure-blob-storage.md) with **shared access signature** authentication. If you want to directly copy data to Azure Data Lake Storage Gen2 in the following supported format, you can create an Azure Blob linked service with SAS authentication against your ADLS Gen2 account, to avoid using [staged copy from Snowflake](#staged-copy-from-snowflake).
218218

@@ -228,8 +228,8 @@ If your sink data store and format meet the criteria described in this section,
228228
- `compression` can be **no compression**, **gzip**, **bzip2**, or **deflate**.
229229
- `encodingName` is left as default or set to **utf-8**.
230230
- `filePattern` in copy activity sink is left as default or set to **setOfObjects**.
231-
- In copy activity source, `additionalColumns` is not specified.
232-
- Column mapping is not specified.
231+
- In copy activity source, `additionalColumns` isn't specified.
232+
- Column mapping isn't specified.
233233

234234
**Example:**
235235

@@ -275,7 +275,7 @@ If your sink data store and format meet the criteria described in this section,
275275

276276
#### Staged copy from Snowflake
277277

278-
When your sink data store or format is not natively compatible with the Snowflake COPY command, as mentioned in the last section, enable the built-in staged copy using an interim Azure Blob storage instance. The staged copy feature also provides you with better throughput. The service exports data from Snowflake into staging storage, then copies the data to sink, and finally cleans up your temporary data from the staging storage. See [Staged copy](copy-activity-performance-features.md#staged-copy) for details about copying data by using staging.
278+
When your sink data store or format isn't natively compatible with the Snowflake COPY command, as mentioned in the last section, enable the built-in staged copy using an interim Azure Blob storage instance. The staged copy feature also provides you with better throughput. The service exports data from Snowflake into staging storage, then copies the data to sink, and finally cleans up your temporary data from the staging storage. See [Staged copy](copy-activity-performance-features.md#staged-copy) for details about copying data by using staging.
279279

280280
To use this feature, create an [Azure Blob storage linked service](connector-azure-blob-storage.md#linked-service-properties) that refers to the Azure storage account as the interim staging. Then specify the `enableStaging` and `stagingSettings` properties in the Copy activity.
281281

@@ -354,7 +354,7 @@ To copy data to Snowflake, the following properties are supported in the Copy ac
354354
355355
#### Direct copy to Snowflake
356356

357-
If your source data store and format meet the criteria described in this section, you can use the Copy activity to directly copy from source to Snowflake. The service checks the settings and fails the Copy activity run if the following criteria is not met:
357+
If your source data store and format meet the criteria described in this section, you can use the Copy activity to directly copy from source to Snowflake. The service checks the settings and fails the Copy activity run if the following criteria isn't met:
358358

359359
- The **source linked service** is [**Azure Blob storage**](connector-azure-blob-storage.md) with **shared access signature** authentication. If you want to directly copy data from Azure Data Lake Storage Gen2 in the following supported format, you can create an Azure Blob linked service with SAS authentication against your ADLS Gen2 account, to avoid using [staged copy to Snowflake](#staged-copy-to-snowflake).
360360

@@ -363,20 +363,20 @@ If your source data store and format meet the criteria described in this section
363363
- For **Parquet** format, the compression codec is **None**, or **Snappy**.
364364

365365
- For **delimited text** format:
366-
- `rowDelimiter` is **\r\n**, or any single character. If row delimiter is not “\r\n”, `firstRowAsHeader` need to be **false**, and `skipLineCount` is not specified.
366+
- `rowDelimiter` is **\r\n**, or any single character. If row delimiter isn't “\r\n”, `firstRowAsHeader` need to be **false**, and `skipLineCount` isn't specified.
367367
- `compression` can be **no compression**, **gzip**, **bzip2**, or **deflate**.
368368
- `encodingName` is left as default or set to "UTF-8", "UTF-16", "UTF-16BE", "UTF-32", "UTF-32BE", "BIG5", "EUC-JP", "EUC-KR", "GB18030", "ISO-2022-JP", "ISO-2022-KR", "ISO-8859-1", "ISO-8859-2", "ISO-8859-5", "ISO-8859-6", "ISO-8859-7", "ISO-8859-8", "ISO-8859-9", "WINDOWS-1250", "WINDOWS-1251", "WINDOWS-1252", "WINDOWS-1253", "WINDOWS-1254", "WINDOWS-1255".
369369
- `quoteChar` is **double quote**, **single quote**, or **empty string** (no quote char).
370370
- For **JSON** format, direct copy only supports the case that sink Snowflake table only has single column and the data type of this column is **VARIANT**, **OBJECT**, or **ARRAY**.
371371
- `compression` can be **no compression**, **gzip**, **bzip2**, or **deflate**.
372372
- `encodingName` is left as default or set to **utf-8**.
373-
- Column mapping is not specified.
373+
- Column mapping isn't specified.
374374

375375
- In the Copy activity source:
376376

377-
- `additionalColumns` is not specified.
377+
- `additionalColumns` isn't specified.
378378
- If your source is a folder, `recursive` is set to true.
379-
- `prefix`, `modifiedDateTimeStart`, `modifiedDateTimeEnd`, and `enablePartitionDiscovery` are not specified.
379+
- `prefix`, `modifiedDateTimeStart`, `modifiedDateTimeEnd`, and `enablePartitionDiscovery` aren't specified.
380380

381381
**Example:**
382382

@@ -421,7 +421,7 @@ If your source data store and format meet the criteria described in this section
421421

422422
#### Staged copy to Snowflake
423423

424-
When your source data store or format is not natively compatible with the Snowflake COPY command, as mentioned in the last section, enable the built-in staged copy using an interim Azure Blob storage instance. The staged copy feature also provides you with better throughput. The service automatically converts the data to meet the data format requirements of Snowflake. It then invokes the COPY command to load data into Snowflake. Finally, it cleans up your temporary data from the blob storage. See [Staged copy](copy-activity-performance-features.md#staged-copy) for details about copying data using staging.
424+
When your source data store or format isn't natively compatible with the Snowflake COPY command, as mentioned in the last section, enable the built-in staged copy using an interim Azure Blob storage instance. The staged copy feature also provides you with better throughput. The service automatically converts the data to meet the data format requirements of Snowflake. It then invokes the COPY command to load data into Snowflake. Finally, it cleans up your temporary data from the blob storage. See [Staged copy](copy-activity-performance-features.md#staged-copy) for details about copying data using staging.
425425

426426
To use this feature, create an [Azure Blob storage linked service](connector-azure-blob-storage.md#linked-service-properties) that refers to the Azure storage account as the interim staging. Then specify the `enableStaging` and `stagingSettings` properties in the Copy activity.
427427

@@ -485,7 +485,7 @@ The below table lists the properties supported by Snowflake source. You can edit
485485
| Enable incremental extract (Preview) | Use this option to tell ADF to only process rows that have changed since the last time that the pipeline executed. | No | Boolean | enableCdc |
486486
| Incremental Column | When using the incremental extract feature, you must choose the date/time/numeric column that you wish to use as the watermark in your source table. | No | String | waterMarkColumn |
487487
| Enable Snowflake Change Tracking (Preview) | This option enables ADF to leverage Snowflake change data capture technology to process only the delta data since the previous pipeline execution. This option automatically loads the delta data with row insert, update and deletion operations without requiring any incremental column. | No | Boolean | enableNativeCdc |
488-
| Net Changes | When using snowflake change tracking, you can use this option to get deduped changed rows or exhaustive changes. Deduped changed rows will show only the latest versions of the rows that have changed since a given point in time, while exhaustive changes will show you all the versions of each row that has changed, including the ones that were deleted or updated. For example, if you update a row, you will see a delete version and an insert version in exhaustive changes, but only the insert version in deduped changed rows. Depending on your use case, you can choose the option that suits your needs. The default option is false, which means exhaustive changes. | No | Boolean | netChanges |
488+
| Net Changes | When using snowflake change tracking, you can use this option to get deduped changed rows or exhaustive changes. Deduped changed rows will show only the latest versions of the rows that have changed since a given point in time, while exhaustive changes will show you all the versions of each row that has changed, including the ones that were deleted or updated. For example, if you update a row, you'll see a delete version and an insert version in exhaustive changes, but only the insert version in deduped changed rows. Depending on your use case, you can choose the option that suits your needs. The default option is false, which means exhaustive changes. | No | Boolean | netChanges |
489489
| Include system Columns | When using snowflake change tracking, you can use the systemColumns option to control whether the metadata stream columns provided by Snowflake are included or excluded in the change tracking output. By default, systemColumns is set to true, which means the metadata stream columns are included. You can set systemColumns to false if you want to exclude them. | No | Boolean | systemColumns |
490490
| Start reading from beginning | Setting this option with incremental extract and change tracking will instruct ADF to read all rows on first execution of a pipeline with incremental extract turned on. | No | Boolean | skipInitialLoad |
491491

@@ -516,7 +516,7 @@ Azure Data Factory now supports a native feature in Snowflake known as change tr
516516

517517
### Sink transformation
518518

519-
The below table lists the properties supported by Snowflake sink. You can edit these properties in the **Settings** tab. When using inline dataset, you will see additional settings, which are the same as the properties described in [dataset properties](#dataset-properties) section. The connector utilizes Snowflake [internal data transfer](https://docs.snowflake.com/en/user-guide/spark-connector-overview.html#internal-data-transfer).
519+
The below table lists the properties supported by Snowflake sink. You can edit these properties in the **Settings** tab. When using inline dataset, you'll see additional settings, which are the same as the properties described in [dataset properties](#dataset-properties) section. The connector utilizes Snowflake [internal data transfer](https://docs.snowflake.com/en/user-guide/spark-connector-overview.html#internal-data-transfer).
520520

521521
| Name | Description | Required | Allowed values | Data flow script property |
522522
| ---- | ----------- | -------- | -------------- | ---------------- |

0 commit comments

Comments
 (0)