Skip to content

Commit 4fe6ed8

Browse files
authored
Added filtering out for job clusters in clusters crawler (#298)
Closed #279
1 parent b1cd014 commit 4fe6ed8

File tree

2 files changed

+15
-1
lines changed

2 files changed

+15
-1
lines changed

src/databricks/labs/ucx/assessment/crawlers.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,7 @@
22
from dataclasses import dataclass
33

44
from databricks.sdk import WorkspaceClient
5+
from databricks.sdk.service.compute import ClusterSource
56
from databricks.sdk.service.jobs import BaseJob
67

78
from databricks.labs.ucx.framework.crawlers import CrawlerBase, SqlBackend
@@ -61,6 +62,8 @@ def _crawl(self) -> list[ClusterInfo]:
6162

6263
def _assess_clusters(self, all_clusters):
6364
for cluster in all_clusters:
65+
if cluster.cluster_source == ClusterSource.JOB:
66+
continue
6467
cluster_info = ClusterInfo(cluster.cluster_id, cluster.cluster_name, cluster.creator_user_name, 1, "")
6568
support_status = spark_version_compatibility(cluster.spark_version)
6669
failures = []

tests/unit/assessment/test_assessment.py

Lines changed: 12 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
from unittest.mock import Mock
22

3-
from databricks.sdk.service.compute import AutoScale, ClusterDetails
3+
from databricks.sdk.service.compute import AutoScale, ClusterDetails, ClusterSource
44
from databricks.sdk.service.jobs import BaseJob, JobSettings, NotebookTask, Task
55

66
from databricks.labs.ucx.assessment.crawlers import ClustersCrawler, JobsCrawler
@@ -82,6 +82,7 @@ def test_job_assessment():
8282
spark_env_vars=None,
8383
spark_version="13.3.x-cpu-ml-scala2.12",
8484
cluster_id="0807-225846-motto493",
85+
cluster_source=ClusterSource.UI,
8586
),
8687
ClusterDetails(
8788
autoscale=AutoScale(min_workers=1, max_workers=6),
@@ -90,6 +91,16 @@ def test_job_assessment():
9091
spark_env_vars=None,
9192
spark_version="9.3.x-cpu-ml-scala2.12",
9293
cluster_id="0810-225833-atlanta69",
94+
cluster_source=ClusterSource.UI,
95+
),
96+
ClusterDetails(
97+
autoscale=AutoScale(min_workers=1, max_workers=6),
98+
spark_conf={"spark.databricks.delta.preview.enabled": "true"},
99+
spark_context_id=5134472582179566666,
100+
spark_env_vars=None,
101+
spark_version="13.3.x-cpu-ml-scala2.12",
102+
cluster_id="0810-229933-chicago12",
103+
cluster_source=ClusterSource.JOB,
93104
),
94105
]
95106
result_set = JobsCrawler(Mock(), MockBackend(), "ucx")._assess_jobs(

0 commit comments

Comments
 (0)