Skip to content

Commit bdaacbd

Browse files
committed
metadata
1 parent 9d8af8d commit bdaacbd

File tree

5 files changed

+3
-362
lines changed

5 files changed

+3
-362
lines changed

METADATA_PLAN.md

Lines changed: 0 additions & 138 deletions
This file was deleted.

uc-neo4j-test-suite/DBX_CLUSTER_JOBS.md

Lines changed: 0 additions & 55 deletions
This file was deleted.

uc-neo4j-test-suite/enable_system_tables.sh

Lines changed: 0 additions & 96 deletions
This file was deleted.

uc-neo4j-test-suite/metadata_sync_delta.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
{
44
"cell_type": "markdown",
55
"metadata": {},
6-
"source": "# Metadata Sync: Materialized Delta Tables (Approach 3)\n\nThis notebook materializes Neo4j node labels and relationship types as **managed Delta tables**\nin Unity Catalog. When data is written as a Delta table, UC automatically registers the full\nschema metadata — column names, types, nullability, row counts, and statistics — making it\nbrowsable in **Catalog Explorer** and queryable via `INFORMATION_SCHEMA`.\n\n**What this proves:** Neo4j graph schema can be synchronized into Unity Catalog with zero\ncustom API calls. The Spark Connector infers the schema, and `saveAsTable()` does the rest.\n\n### Steps\n\n1. Load configuration from Databricks Secrets\n2. Verify Neo4j connectivity\n3. Discover all node labels and their properties\n4. Create target UC schemas (catalog must already exist — see README)\n5. Materialize a single label as a Delta table (test)\n6. Verify metadata in `INFORMATION_SCHEMA`\n7. Materialize all discovered labels\n8. Materialize relationship types\n9. Final verification and summary\n\n### Prerequisites\n\n- **Target catalog `neo4j_metadata` must already exist** — see `METADATA_SYNC_README.md`\n- `neo4j-uc-creds` secret scope configured via `setup.sh`\n- Neo4j Spark Connector installed on cluster (`org.neo4j:neo4j-connector-apache-spark_2.12:5.4.0_for_spark_3`)\n- Neo4j Python driver installed (`neo4j`)\n- **Single user** access mode cluster (required by Spark Connector)",
6+
"source": "# Metadata Sync: Materialized Delta Tables\n\nThis notebook materializes Neo4j node labels and relationship types as **managed Delta tables**\nin Unity Catalog. When data is written as a Delta table, UC automatically registers the full\nschema metadata — column names, types, nullability, row counts, and statistics — making it\nbrowsable in **Catalog Explorer** and queryable via `INFORMATION_SCHEMA`.\n\n**What this proves:** Neo4j graph schema can be synchronized into Unity Catalog with zero\ncustom API calls. The Spark Connector infers the schema, and `saveAsTable()` does the rest.\n\n### Steps\n\n1. Load configuration from Databricks Secrets\n2. Verify Neo4j connectivity\n3. Discover all node labels and their properties\n4. Create target UC schemas (catalog must already exist — see README)\n5. Materialize a single label as a Delta table (test)\n6. Verify metadata in `INFORMATION_SCHEMA`\n7. Materialize all discovered labels\n8. Materialize relationship types\n9. Final verification and summary\n\n### Prerequisites\n\n- **Target catalog `neo4j_metadata` must already exist** — see `METADATA_SYNC_README.md`\n- `neo4j-uc-creds` secret scope configured via `setup.sh`\n- Neo4j Spark Connector installed on cluster (`org.neo4j:neo4j-connector-apache-spark_2.12:5.4.0_for_spark_3`)\n- Neo4j Python driver installed (`neo4j`)\n- **Single user** access mode cluster (required by Spark Connector)",
77
"id": "intro"
88
},
99
{

0 commit comments

Comments
 (0)