Skip to content

Commit c321830

Browse files
committed
Extend err message
1 parent 3e05103 commit c321830

File tree

1 file changed

+17
-1
lines changed

1 file changed

+17
-1
lines changed

kedro-datasets/kedro_datasets/_utils/spark_utils.py

Lines changed: 17 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,23 @@ def get_spark() -> Union[SparkSession, "DatabricksSession"]:
1919
from databricks.connect import DatabricksSession
2020

2121
# We can't test this as there's no Databricks test env available
22-
spark = DatabricksSession.builder.getOrCreate() # pragma: no cover
22+
try:
23+
spark = DatabricksSession.builder.getOrCreate() # pragma: no cover
24+
# this can't be narrowed down since databricks-connect throws error of Exception type
25+
except Exception as e:
26+
error_message = str(e)
27+
if (
28+
error_message
29+
== "Cluster id or serverless are required but were not specified."
30+
):
31+
raise type(e)(
32+
"DatabricksSession is expected to behave as singleton but it didn't. "
33+
"Either set up DATABRICKS_CONFIG_PROFILE or DATABRICKS_PROFILE and DATABRICKS_SERVERLESS_COMPUTE_ID "
34+
"env variables in your hooks prior to using the spark session. "
35+
"Read more about these variables here: "
36+
"https://docs.databricks.com/aws/en/dev-tools/databricks-connect/cluster-config#config-profile-env-var"
37+
) from e
38+
pass
2339

2440
except ImportError:
2541
# For "normal" spark sessions that don't use databricks-connect

0 commit comments

Comments
 (0)