Skip to content

Commit 036fe73

Browse files
committed
Freshness and format updates
1 parent 1c73774 commit 036fe73

File tree

1 file changed

+5
-7
lines changed

1 file changed

+5
-7
lines changed
Lines changed: 5 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
11
---
22
title: What is Delta Lake?
3-
description: Overview of Delta Lake and how it works as part of Azure Synapse Analytics
3+
description: Overview of Delta Lake's key features and how it brings atomicity, consistency, isolation, and durability to Azure Synapse Analytics.
44
author: jovanpop-msft
55
ms.author: jovanpop
66
ms.service: azure-synapse-analytics
7-
ms.topic: conceptual
7+
ms.topic: concept-article
88
ms.subservice: spark
9-
ms.date: 12/06/2022
9+
ms.date: 11/15/2024
1010
ms.reviewer: euang
1111
---
1212

@@ -16,15 +16,13 @@ Delta Lake is an open-source storage layer that brings ACID (atomicity, consiste
1616

1717
The current version of Delta Lake included with Azure Synapse has language support for Scala, PySpark, and .NET and is compatible with Linux Foundation Delta Lake. There are links at the bottom of the page to more detailed examples and documentation. You can learn more from the [Introduction to Delta Tables video](https://www.youtube.com/watch?v=B_wyRXlLKok).
1818

19-
## Key features
20-
2119
| Feature | Description |
2220
| --- | --- |
2321
| **ACID Transactions** | Data lakes are typically populated through multiple processes and pipelines, some of which are writing data concurrently with reads. Prior to Delta Lake and the addition of transactions, data engineers had to go through a manual error prone process to ensure data integrity. Delta Lake brings familiar ACID transactions to data lakes. It provides serializability, the strongest level of isolation level. Learn more at [Diving into Delta Lake: Unpacking the Transaction Log](https://databricks.com/blog/2019/08/21/diving-into-delta-lake-unpacking-the-transaction-log.html).|
2422
| **Scalable Metadata Handling** | In big data, even the metadata itself can be "big data." Delta Lake treats metadata just like data, leveraging Spark's distributed processing power to handle all its metadata. As a result, Delta Lake can handle petabyte-scale tables with billions of partitions and files at ease. |
2523
| **Time Travel (data versioning)** | The ability to "undo" a change or go back to a previous version is one of the key features of transactions. Delta Lake provides snapshots of data enabling you to revert to earlier versions of data for audits, rollbacks or to reproduce experiments. Learn more in [Introducing Delta Lake Time Travel for Large Scale Data Lakes](https://databricks.com/blog/2019/02/04/introducing-delta-time-travel-for-large-scale-data-lakes.html). |
2624
| **Open Format** | Apache Parquet is the baseline format for Delta Lake, enabling you to leverage the efficient compression and encoding schemes that are native to the format. |
27-
| **Unified Batch and Streaming Source and Sink** | A table in Delta Lake is both a batch table, as well as a streaming source and sink. Streaming data ingest, batch historic backfill, and interactive queries all just work out of the box. |
25+
| **Unified Batch and Streaming Source and Sink** | A table in Delta Lake is both a batch table, and a streaming source and sink. Streaming data ingest, batch historic backfill, and interactive queries all just work out of the box. |
2826
| **Schema Enforcement** | Schema enforcement helps ensure that the data types are correct and required columns are present, preventing bad data from causing data inconsistency. For more information, see [Diving Into Delta Lake: Schema Enforcement & Evolution](https://databricks.com/blog/2019/09/24/diving-into-delta-lake-schema-enforcement-evolution.html) |
2927
| **Schema Evolution** | Delta Lake enables you to make changes to a table schema that can be applied automatically, without having to write migration DDL. For more information, see [Diving Into Delta Lake: Schema Enforcement & Evolution](https://databricks.com/blog/2019/09/24/diving-into-delta-lake-schema-enforcement-evolution.html) |
3028
| **Audit History** | Delta Lake transaction log records details about every change made to data providing a full audit trail of the changes. |
@@ -35,7 +33,7 @@ For full documentation, see the [Delta Lake Documentation Page](https://docs.del
3533

3634
For more information, see [Delta Lake Project](https://github.com/delta-io/delta).
3735

38-
## Next steps
36+
## Related content
3937

4038
- [.NET for Apache Spark documentation](/previous-versions/dotnet/spark/what-is-apache-spark-dotnet)
4139
- [Azure Synapse Analytics](../index.yml)

0 commit comments

Comments
 (0)