Skip to content

Commit dcbacf2

Browse files
Merge pull request #208723 from kromerm/deltaupdate
Deltaupdate
2 parents b139386 + e13ffb5 commit dcbacf2

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

articles/data-factory/format-delta.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ author: kromerm
55
ms.service: data-factory
66
ms.subservice: data-flows
77
ms.topic: conceptual
8-
ms.date: 08/05/2022
8+
ms.date: 08/22/2022
99
ms.author: makromer
1010
---
1111

@@ -70,6 +70,7 @@ The below table lists the properties supported by a delta sink. You can edit the
7070
| Compression type | The compression type of the delta table | no | `bzip2`<br>`gzip`<br>`deflate`<br>`ZipDeflate`<br>`snappy`<br>`lz4` | compressionType |
7171
| Compression level | Choose whether the compression completes as quickly as possible or if the resulting file should be optimally compressed. | required if `compressedType` is specified. | `Optimal` or `Fastest` | compressionLevel |
7272
| Vacuum | Specify retention threshold in hours for older versions of table. A value of 0 or less defaults to 30 days | yes | Integer | vacuum |
73+
| Table action | Tells ADF what to do with the target Delta table in your sink. You can leave it as-is and append new rows, overwrite the existing table definition and data with new metadata and data, or keep the existing table structure but first truncate all rows, then insert the new rows. | no | None, Truncate, Overwrite | truncate, overwrite |
7374
| Update method | Specify which update operations are allowed on the delta lake. For methods that aren't insert, a preceding alter row transformation is required to mark rows. | yes | `true` or `false` | deletable <br> insertable <br> updateable <br> merge |
7475
| Optimized Write | Achieve higher throughput for write operation via optimizing internal shuffle in Spark executors. As a result, you may notice fewer partitions and files that are of a larger size | no | `true` or `false` | optimizedWrite: true |
7576
| Auto Compact | After any write operation has completed, Spark will automatically execute the ```OPTIMIZE``` command to re-organize the data, resulting in more partitions if necessary, for better reading performance in the future | no | `true` or `false` | autoCompact: true |

0 commit comments

Comments
 (0)