|
3 | 3 | 2. [Databricks Techsummit Demo](#databricks-tech-summit-fy2024-demo): 100s of data sources ingestion in bronze and silver DLT pipelines automatically. |
4 | 4 | 3. [Append FLOW Autoloader Demo](#append-flow-autoloader-file-metadata-demo): Write to same target from multiple sources using [dlt.append_flow](https://docs.databricks.com/en/delta-live-tables/flows.html#append-flows) and adding [File metadata column](https://docs.databricks.com/en/ingestion/file-metadata-column.html) |
5 | 5 | 4. [Append FLOW Eventhub Demo](#append-flow-eventhub-demo): Write to same target from multiple sources using [dlt.append_flow](https://docs.databricks.com/en/delta-live-tables/flows.html#append-flows) and adding [File metadata column](https://docs.databricks.com/en/ingestion/file-metadata-column.html) |
| 6 | + 5. [Silver Fanout Demo](#silver-fanout-demo): This demo showcases the implementation of fanout architecture in the silver layer. |
6 | 7 |
|
7 | 8 |
|
8 | 9 |
|
@@ -35,7 +36,7 @@ This Demo launches Bronze and Silver DLT pipelines with following activities: |
35 | 36 | export PYTHONPATH=$dlt_meta_home |
36 | 37 | ``` |
37 | 38 |
|
38 | | -6. Run the command ```python demo/launch_dais_demo.py --source=cloudfiles --uc_catalog_name=<<uc catalog name>> --cloud_provider_name=aws --dbr_version=15.3.x-scala2.12 --dbfs_path=dbfs:/dais-dlt-meta-demo-automated_new``` |
| 39 | +6. Run the command ```python demo/launch_dais_demo.py --source=cloudfiles --uc_catalog_name=<<uc catalog name>> --cloud_provider_name=aws --dbr_version=15.3.x-scala2.12 --dbfs_path=dbfs:/dais-dlt-meta-demo-automated``` |
39 | 40 | - cloud_provider_name : aws or azure or gcp |
40 | 41 | - db_version : Databricks Runtime Version |
41 | 42 | - dbfs_path : Path on your Databricks workspace where demo will be copied for launching DLT-META Pipelines |
@@ -202,3 +203,56 @@ This demo will perform following tasks: |
202 | 203 | ``` |
203 | 204 |
|
204 | 205 |  |
| 206 | +
|
| 207 | +
|
| 208 | +# Silver Fanout Demo |
| 209 | +- This demo will showcase the onboarding process for the silver fanout pattern. |
| 210 | + - Run the onboarding process for the bronze cars table, which contains data from various countries. |
| 211 | + - Run the onboarding process for the silver tables, which have a `where_clause` based on the country condition specified in [silver_transformations_cars.json](https://github.com/databrickslabs/dlt-meta/blob/main/demo/conf/silver_transformations_cars.json). |
| 212 | + - Run the Bronze DLT pipeline which will produce cars table. |
| 213 | + - Run Silver DLT pipeline, fanning out from the bronze cars table to country-specific tables such as cars_usa, cars_uk, cars_germany, and cars_japan. |
| 214 | +
|
| 215 | +### Steps: |
| 216 | +1. Launch Terminal/Command prompt |
| 217 | +
|
| 218 | +2. Install [Databricks CLI](https://docs.databricks.com/dev-tools/cli/index.html) |
| 219 | +
|
| 220 | +3. ```commandline |
| 221 | + git clone https://github.com/databrickslabs/dlt-meta.git |
| 222 | + ``` |
| 223 | +
|
| 224 | +4. ```commandline |
| 225 | + cd dlt-meta |
| 226 | + ``` |
| 227 | +5. Set python environment variable into terminal |
| 228 | + ```commandline |
| 229 | + dlt_meta_home=$(pwd) |
| 230 | + ``` |
| 231 | + ```commandline |
| 232 | + export PYTHONPATH=$dlt_meta_home |
| 233 | +
|
| 234 | +6. Run the command ```python demo/launch_silver_fanout_demo.py --source=cloudfiles --uc_catalog_name=<<uc catalog name>> --cloud_provider_name=aws --dbr_version=15.3.x-scala2.12 --dbfs_path=dbfs:/dais-dlt-meta-silver-fanout``` |
| 235 | + - cloud_provider_name : aws or azure |
| 236 | + - db_version : Databricks Runtime Version |
| 237 | + - dbfs_path : Path on your Databricks workspace where demo will be copied for launching DLT-META Pipelines |
| 238 | + - you can provide `--profile=databricks_profile name` in case you already have databricks cli otherwise command prompt will ask host and token. |
| 239 | +
|
| 240 | + - - 6a. Databricks Workspace URL: |
| 241 | + - - Enter your workspace URL, with the format https://<instance-name>.cloud.databricks.com. To get your workspace URL, see Workspace instance names, URLs, and IDs. |
| 242 | +
|
| 243 | + - - 6b. Token: |
| 244 | + - In your Databricks workspace, click your Databricks username in the top bar, and then select User Settings from the drop down. |
| 245 | +
|
| 246 | + - On the Access tokens tab, click Generate new token. |
| 247 | +
|
| 248 | + - (Optional) Enter a comment that helps you to identify this token in the future, and change the token’s default lifetime of 90 days. To create a token with no lifetime (not recommended), leave the Lifetime (days) box empty (blank). |
| 249 | +
|
| 250 | + - Click Generate. |
| 251 | +
|
| 252 | + - Copy the displayed token |
| 253 | +
|
| 254 | + - Paste to command prompt |
| 255 | +
|
| 256 | +  |
| 257 | + |
| 258 | +  |
0 commit comments