Skip to content

Commit e62bedd

Browse files
Improve README notebook documentaiton (#260)
Tune up the documentation - 5 documentation updated for new functions / methods
1 parent 0bc27e5 commit e62bedd

File tree

1 file changed

+14
-15
lines changed

1 file changed

+14
-15
lines changed

src/databricks/labs/ucx/runtime.py

Lines changed: 14 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -70,45 +70,44 @@ def inventorize_mounts(cfg: MigrationConfig):
7070

7171
@task("assessment", depends_on=[inventorize_mounts, crawl_tables])
7272
def inventorize_external_locations(cfg: MigrationConfig):
73-
"""In this part of the assessment, we're going to look up the location of all the tables.
74-
Locations that use
75-
Mount Points will be looked up. We will then find out the External Locations required for a proper migration and
76-
save these to the `$inventory.external_locations` Table.
77-
The logic this assessment follow is extract all the locations for tables that are not using DBFS
78-
(we capture the locations for tables using mount points).
79-
We then scan all the locations and find the common folders that accommodate them.
80-
These external_locations will be created in a later stage before the table can be migrated.
81-
"""
73+
"""In this section of the assessment, our objective is to determine the whereabouts of all the tables.
74+
Specifically, we will focus on identifying locations that utilize Mount Points. Our goal is to identify the
75+
External Locations necessary for a successful migration and store this information in the
76+
$inventory.external_locations Table.
77+
78+
The approach taken in this assessment involves the following steps:
79+
- Extracting all the locations associated with tables that do not use DBFS (with a focus on those
80+
using mount points).
81+
- Scanning all these locations to identify common folders that can accommodate them.
82+
- These identified external locations will be created subsequently prior to the actual table migration"""
8283
ws = WorkspaceClient(config=cfg.to_databricks_config())
8384
crawler = ExternalLocationCrawler(ws, RuntimeBackend(), cfg.inventory_database)
8485
crawler.snapshot()
8586

8687

8788
@task("assessment", depends_on=[setup_schema])
8889
def inventorize_jobs(cfg: MigrationConfig):
89-
"""This part scan through all the jobs and locate ones that are not compatible with UC.
90+
"""This module scans through all the jobs and identifies those that are not compatible with UC.
9091
It looks for:
9192
- Clusters with DBR version earlier than 11.3
9293
- Clusters using Passthru Authentication
9394
- Clusters with incompatible spark config tags
9495
- Clusters referencing DBFS locations in one or more config options
95-
A report with a list of all the Jobs is saved to the `$inventory.jobs` table.
96-
"""
96+
Subsequently, the list of all the jobs is stored in the `$inventory.jobs` table."""
9797
ws = WorkspaceClient(config=cfg.to_databricks_config())
9898
crawler = JobsCrawler(ws, RuntimeBackend(), cfg.inventory_database)
9999
crawler.snapshot()
100100

101101

102102
@task("assessment", depends_on=[setup_schema])
103103
def inventorize_clusters(cfg: MigrationConfig):
104-
"""This part scan through all the clusters and locate ones that are not compatible with UC.
104+
"""This module scan through all the clusters and identifies those that are not compatible with UC.
105105
It looks for:
106106
- Clusters with DBR version earlier than 11.3
107107
- Clusters using Passthru Authentication
108108
- Clusters with incompatible spark config tags
109109
- Clusters referencing DBFS locations in one or more config options
110-
A report with a list of all the Jobs is saved to the `$inventory.clusters` table.
111-
"""
110+
Subsequently, the list of all the clusters is stored in the`$inventory.clusters` table."""
112111
ws = WorkspaceClient(config=cfg.to_databricks_config())
113112
crawler = ClustersCrawler(ws, RuntimeBackend(), cfg.inventory_database)
114113
crawler.snapshot()

0 commit comments

Comments
 (0)