You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: content/en/data_jobs/databricks.md
+44-1Lines changed: 44 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -44,11 +44,26 @@ Follow these steps to enable Data Jobs Monitoring for Databricks.
44
44
1. On the **Configure** tab, click **Add Databricks Workspace**.
45
45
1. Enter a workspace name, your Databricks workspace URL, account ID, and the client ID and secret you generated.
46
46
{{< img src="data_jobs/databricks/configure-workspace-form-m2m.png" alt="In the Datadog-Databricks integration tile, a Databricks workspace is displayed. This workspace has a name, URL, account ID, client ID, and client secret." style="width:100%;" >}}
47
+
1. To gain visibility into your Databricks costs in Data Jobs Monitoring or [Cloud Cost Management][18], provide the ID of a [Databricks SQL Warehouse][19] that Datadog can use to query your [system tables][20].
48
+
- The service principal must have access to the SQL Warehouse. In the Warehouse configuration page, go to **Permissions** (top right) and grant it `CAN USE` permission.
49
+
- Grant the service principal read access to the Unity Catalog [system tables][20] by running the following commands:
50
+
```sql
51
+
GRANT USE CATALOG ON CATALOG system TO <service_principal>;
52
+
GRANTSELECTON CATALOG system TO <service_principal>;
53
+
GRANT USE SCHEMA ON CATALOG system TO <service_principal>;
54
+
```
55
+
The user granting these must have `MANAGE` privilege on `CATALOG system`.
56
+
57
+
- The SQL Warehouse must be Pro or Serverless. Classic Warehouses are **NOT** supported. A 2XS warehouse is recommended, with Auto Stop set to 5-10 minutes to reduce cost.
47
58
1. In the **Select products to set up integration** section, ensure that Data Jobs Monitoring is **Enabled**.
48
59
1. In the **Datadog Agent Setup** section, choose either
49
60
-[Managed by Datadog (recommended)](?tab=datadogmanagedglobalinitscriptrecommended#install-the-datadog-agent): Datadog installs and manages the Agent with a global init script in the workspace.
50
61
-[Manually](?tab=manuallyinstallaglobalinitscript#install-the-datadog-agent): Follow the [instructions below](?tab=manuallyinstallaglobalinitscript#install-the-datadog-agent) to install and manage the init script for installing the Agent globally or on specific Databricks clusters.
{{% tab "Use a Personal Access Token (Legacy)" %}}
@@ -67,6 +82,17 @@ Follow these steps to enable Data Jobs Monitoring for Databricks.
67
82
1. On the **Configure** tab, click **Add Databricks Workspace**.
68
83
1. Enter a workspace name, your Databricks workspace URL, and the Databricks token you generated.
69
84
{{< img src="data_jobs/databricks/configure-workspace-form.png" alt="In the Datadog-Databricks integration tile, a Databricks workspace is displayed. This workspace has a name, URL, and API token." style="width:100%;" >}}
85
+
1. To gain visibility into your Databricks costs in Data Jobs Monitoring or [Cloud Cost Management][18], provide the ID of a [Databricks SQL Warehouse][19] that Datadog can use to query your [system tables][20].
86
+
87
+
- The token's principal must have access to the SQL Warehouse. Give it `CAN USE` permission from **Permissions** at the top right of the Warehouse configuration page.
88
+
- Grant the service principal read access to the Unity Catalog [system tables][20] by running the following commands::
89
+
```sql
90
+
GRANT USE CATALOG ON CATALOG system TO <token_principal>;
91
+
GRANTSELECTON CATALOG system TO <token_principal>;
92
+
GRANT USE SCHEMA ON CATALOG system TO <token_principal>;
93
+
```
94
+
The user granting these must have `MANAGE` privilege on `CATALOG system`.
95
+
- The SQL Warehouse must be Pro or Serverless. Classic Warehouses are **NOT** supported. A 2XS size warehouse is recommended, with Auto Stop configured for 5-10 minutes to minimize cost.
70
96
1. In the **Select products to set up integration** section, make sure the Data Jobs Monitoring product is **Enabled**.
71
97
1. In the **Datadog Agent Setup** section, choose either
72
98
-[Managed by Datadog (recommended)](?tab=datadogmanagedglobalinitscriptrecommended#install-the-datadog-agent): Datadog installs and manages the Agent with a global init script in the workspace.
@@ -76,14 +102,18 @@ Follow these steps to enable Data Jobs Monitoring for Databricks.
The Datadog Agent must be installed on Databricks clusters to monitor Databricks jobs that run on all-purpose or job clusters.
116
+
The Datadog Agent must be installed on Databricks clusters to monitor Databricks jobs that run on all-purpose or job clusters. This step is not required to monitor jobs on [serverless compute][4].
87
117
88
118
{{< tabs >}}
89
119
{{% tab "Datadog managed global init script (Recommended)" %}}
@@ -129,6 +159,7 @@ Optionally, you can add tags to your Databricks cluster and Spark performance me
@@ -296,6 +327,18 @@ If you need more granular control, grant these minimal permissions to the follow
296
327
| Query | [CAN VIEW][23]
297
328
| SQL warehouse | [CAN MONITOR][24]
298
329
330
+
Additionally, for Datadog to access your Databricks cost data in Data Jobs Monitoring or [Cloud Cost Management][18], the user or service principal used to query [system tables][20] must have the following permissions:
331
+
-`CAN USE` permission on the SQL Warehouse.
332
+
- Read access to the [system tables][20] within Unity Catalog. This can be granted with:
333
+
```sql
334
+
GRANT USE CATALOG ON CATALOG system TO <service_principal>;
335
+
GRANTSELECTON CATALOG system TO <service_principal>;
336
+
GRANT USE SCHEMA ON CATALOG system TO <service_principal>;
337
+
```
338
+
The user granting these must have `MANAGE` privilege on `CATALOG system`.
0 commit comments