Skip to content

Commit aa83f2b

Browse files
Merge pull request #221 from databrickslabs/feature/v0.0.10
- Replacing DLT with Lakeflow Declarative Pipelines - Changed direct publishing mode to default publishing mode in docs
2 parents 7fd18ae + 4141ee7 commit aa83f2b

File tree

15 files changed

+32
-32
lines changed

15 files changed

+32
-32
lines changed

README.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -32,10 +32,10 @@ In practice, a single generic pipeline reads the Dataflowspec and uses it to orc
3232
#### Generic Lakeflow Declarative Pipeline
3333

3434
- Apply appropriate readers based on input metadata
35-
- Apply data quality rules with DLT expectations
35+
- Apply data quality rules with Lakeflow Declarative Pipeline expectations
3636
- Apply CDC apply changes if specified in metadata
3737
- Builds Lakeflow Declarative Pipeline graph based on input/output metadata
38-
- Launch Lakeflow Declarative Pipeline pipeline
38+
- Launch Lakeflow Declarative pipeline
3939

4040
## High-Level Process Flow:
4141

@@ -45,7 +45,7 @@ In practice, a single generic pipeline reads the Dataflowspec and uses it to orc
4545

4646
![DLT-META Stages](./docs/static/images/dlt-meta_stages.png)
4747

48-
## DLT-META DLT Features support
48+
## DLT-META Lakeflow Declarative Pipeline Features support
4949
| Features | DLT-META Support |
5050
| ------------- | ------------- |
5151
| Input data sources | Autoloader, Delta, Eventhub, Kafka, snapshot |
@@ -58,7 +58,7 @@ In practice, a single generic pipeline reads the Dataflowspec and uses it to orc
5858
| [append_flow](https://docs.databricks.com/en/delta-live-tables/flows.html#use-append-flow-to-write-to-a-streaming-table-from-multiple-source-streams) API support | Bronze layer|
5959
| Liquid cluster support | Bronze, Bronze Quarantine, Silver tables|
6060
| [DLT-META CLI](https://databrickslabs.github.io/dlt-meta/getting_started/dltmeta_cli/) | ```databricks labs dlt-meta onboard```, ```databricks labs dlt-meta deploy``` |
61-
| Bronze and Silver pipeline chaining | Deploy dlt-meta pipeline with ```layer=bronze_silver``` option using Direct publishing mode |
61+
| Bronze and Silver pipeline chaining | Deploy dlt-meta pipeline with ```layer=bronze_silver``` option using default publishing mode |
6262
| [create_sink](https://docs.databricks.com/aws/en/dlt-ref/dlt-python-ref-sink) API support |Supported formats:```external delta table , kafka``` Bronze, Silver layers|
6363
| [Databricks Asset Bundles](https://docs.databricks.com/aws/en/dev-tools/bundles/) | Supported
6464
| [DLT-META UI](https://github.com/databrickslabs/dlt-meta/tree/main/lakehouse_app#dlt-meta-lakehouse-app-setup) | Uses Databricks Lakehouse DLT-META App
@@ -166,7 +166,7 @@ The command will prompt you to provide pipeline configuration details.
166166
![deployingDLTMeta_bronze_silver.gif](docs/static/images/deployingDLTMeta_bronze_silver.gif)
167167

168168
Above deploy cli command will:
169-
1. Deploy Lakeflow Declarative pipeline with dlt-meta configuration like ```layer```, ```group```, ```dataflowSpec table details``` etc to your databricks workspace
169+
1. Deploy Lakeflow Declarative Pipeline with dlt-meta configuration like ```layer```, ```group```, ```dataflowSpec table details``` etc to your databricks workspace
170170
2. Display message: ```dlt-meta pipeline={pipeline_id} created and launched with update_id={pipeline_update_id}, url=https://{databricks workspace url}/#joblist/pipelines/{pipeline_id}```
171171
3. Pipline URL will automatically open in your defaul browser.
172172

demo/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
## [DAIS 2023 Session Recording](https://www.youtube.com/watch?v=WYv5haxLlfA)
1414
This Demo launches Bronze and Silver pipelines with following activities:
1515
- Customer and Transactions feeds for initial load
16-
- Adds new feeds Product and Stores to existing Bronze and Silver DLT pipelines with metadata changes.
16+
- Adds new feeds Product and Stores to existing Bronze and Silver Lakeflow Declarative pipeline with metadata changes.
1717
- Runs Bronze and Silver pipeline for incremental load for CDC events
1818

1919
### Steps:

demo/notebooks/afam_cloudfiles_runners/validate.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@
2323
}
2424

2525

26-
log_list.append("Validating DLT Bronze and Silver Table Counts...")
26+
log_list.append("Validating Lakeflow Declarative Pipeline Bronze and Silver Table Counts...")
2727
for table, counts in UC_TABLES.items():
2828
query = spark.sql(f"SELECT count(*) as cnt FROM {table}")
2929
cnt = query.collect()[0].cnt

demo/notebooks/afam_eventhub_runners/validate.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
f"{uc_catalog_name}.{bronze_schema}.bronze_{run_id}_iot_quarantine": 2
2222
}
2323

24-
log_list.append("Validating DLT EVenthub Bronze Table Counts...")
24+
log_list.append("Validating Lakeflow Declarative Pipeline EVenthub Bronze Table Counts...")
2525
tables = UC_TABLES if uc_enabled else NON_UC_TABLES
2626
for table, counts in tables.items():
2727
query = spark.sql(f"SELECT count(*) as cnt FROM {table}")

demo/notebooks/dlt_sink_runners/validate.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
f"{uc_catalog_name}.{bronze_schema}.bronze_{run_id}_iot_quarantine": 2,
1616
}
1717

18-
log_list.append("Validating DLT EVenthub Bronze Table Counts...")
18+
log_list.append("Validating Lakeflow Declarative Pipeline EVenthub Bronze Table Counts...")
1919
for table, counts in TABLES.items():
2020
query = spark.sql(f"SELECT count(*) as cnt FROM {table}")
2121
cnt = query.collect()[0].cnt

docs/content/_index.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ In practice, a single generic pipeline reads the Dataflowspec and uses it to orc
3636
- Option#2: [Manual Job](https://databrickslabs.github.io/dlt-meta/getting_started/dltmeta_manual/#onboardjob)
3737
- option#3: [Databricks Notebook](https://databrickslabs.github.io/dlt-meta/getting_started/dltmeta_manual/#option2-databricks-notebook)
3838

39-
- Dataflow DLT Pipeline
39+
- Dataflow Lakeflow Declarative Pipeline
4040
- Option#1: [DLT-META CLI](https://databrickslabs.github.io/dlt-meta/getting_started/dltmeta_cli/#dataflow-dlt-pipeline)
4141
- Option#2: [DLT-META MANUAL](https://databrickslabs.github.io/dlt-meta/getting_started/dltmeta_manual/#dataflow-dlt-pipeline)
4242

@@ -53,7 +53,7 @@ In practice, a single generic pipeline reads the Dataflowspec and uses it to orc
5353
| [append_flow](https://docs.databricks.com/aws/en/dlt-ref/dlt-python-ref-append-flow) API support | Bronze layer|
5454
| Liquid cluster support | Bronze, Bronze Quarantine, Silver, Silver Quarantine tables|
5555
| [DLT-META CLI](https://databrickslabs.github.io/dlt-meta/getting_started/dltmeta_cli/) | ```databricks labs dlt-meta onboard```, ```databricks labs dlt-meta deploy``` |
56-
| Bronze and Silver pipeline chaining | Deploy dlt-meta pipeline with ```layer=bronze_silver``` option using Direct publishing mode |
56+
| Bronze and Silver pipeline chaining | Deploy dlt-meta pipeline with ```layer=bronze_silver``` option using default publishing mode |
5757
| [DLT Sinks](https://docs.databricks.com/aws/en/dlt/dlt-sinks) | Supported formats:external ```delta table```, ```kafka```.Bronze, Silver layers|
5858
## How much does it cost ?
5959
DLT-META does not have any **direct cost** associated with it other than the cost to run the Databricks Lakeflow Declarative Pipelines

docs/content/additionals/integration_tests.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@ draft: false
7171
,0
7272
0,Completed Bronze Lakeflow Declarative Pipeline.
7373
1,Completed Silver Lakeflow Declarative Pipeline.
74-
2,Validating DLT Bronze and Silver Table Counts...
74+
2,Validating Lakeflow Declarative Pipeline Bronze and Silver Table Counts...
7575
3,Validating Counts for Table bronze_7d1d3ccc9e144a85b07c23110ea50133.transactions.
7676
4,Expected: 10002 Actual: 10002. Passed!
7777
5,Validating Counts for Table bronze_7d1d3ccc9e144a85b07c23110ea50133.transactions_quarantine.

docs/content/demo/Apply_Changes_From_Snapshot.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,11 +9,11 @@ draft: false
99
- This demo will perform following steps
1010
- Showcase onboarding process for apply changes from snapshot pattern
1111
- Run onboarding for the bronze stores and products tables, which contains data snapshot data in csv files.
12-
- Run Bronze DLT to load initial snapshot (LOAD_1.csv)
12+
- Run Bronze Lakeflow Declarative Pipeline to load initial snapshot (LOAD_1.csv)
1313
- Upload incremental snapshot LOAD_2.csv version=2 for stores and product
14-
- Run Bronze DLT to load incremental snapshot (LOAD_2.csv). Stores is scd_type=2 so updated records will expired and added new records with version_number. Products is scd_type=1 so in case records missing for scd_type=1 will be deleted.
14+
- Run Bronze Lakeflow Declarative Pipeline to load incremental snapshot (LOAD_2.csv). Stores is scd_type=2 so updated records will expired and added new records with version_number. Products is scd_type=1 so in case records missing for scd_type=1 will be deleted.
1515
- Upload incremental snapshot LOAD_3.csv version=3 for stores and product
16-
- Run Bronze DLT to load incremental snapshot (LOAD_3.csv). Stores is scd_type=2 so updated records will expired and added new records with version_number. Products is scd_type=1 so in case records missing for scd_type=1 will be deleted.
16+
- Run Bronze Lakeflow Declarative Pipeline to load incremental snapshot (LOAD_3.csv). Stores is scd_type=2 so updated records will expired and added new records with version_number. Products is scd_type=1 so in case records missing for scd_type=1 will be deleted.
1717

1818

1919
### Steps:

docs/content/demo/DAIS.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -8,10 +8,10 @@ draft: false
88
### DAIS 2023 DEMO:
99
#### [DAIS 2023 Session Recording](https://www.youtube.com/watch?v=WYv5haxLlfA)
1010

11-
This demo showcases DLT-META's capabilities of creating Bronze and Silver DLT pipelines with initial and incremental mode automatically.
11+
This demo showcases DLT-META's capabilities of creating Bronze and Silver Lakeflow Declarative pipeline with initial and incremental mode automatically.
1212
- Customer and Transactions feeds for initial load
13-
- Adds new feeds Product and Stores to existing Bronze and Silver DLT pipelines with metadata changes.
14-
- Runs Bronze and Silver DLT for incremental load for CDC events
13+
- Adds new feeds Product and Stores to existing Bronze and Silver Lakeflow Declarative pipeline with metadata changes.
14+
- Runs Bronze and Silver Lakeflow Declarative Pipeline for incremental load for CDC events
1515

1616
#### Steps to launch DAIS demo in your Databricks workspace:
1717
1. Launch Command Prompt

docs/content/demo/Techsummit.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ draft: false
66
---
77

88
### Databricks Tech Summit FY2024 DEMO:
9-
This demo will launch auto generated tables(100s) inside single bronze and silver DLT pipeline using dlt-meta.
9+
This demo will launch auto generated tables(100s) inside single bronze and silver Lakeflow Declarative Pipeline using dlt-meta.
1010

1111
1. Launch Command Prompt
1212

0 commit comments

Comments
 (0)