You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
> If you enabled [Data exfiltration protection](../security/workspace-data-exfiltration-protection.md), you should upload all your dependencies as workspace libraries.
51
-
52
49
## Pool packages
53
50
54
51
In some cases, you might want to standardize the packages that are used on an Apache Spark pool. This standardization can be useful if the same packages are commonly installed by multiple people on your team.
@@ -63,7 +60,39 @@ To learn more about these capabilities, see [Manage Spark pool packages](./apach
63
60
>
64
61
> - If the package you are installing is large or takes a long time to install, this fact affects the Spark instance start up time.
65
62
> - Altering the PySpark, Python, Scala/Java, .NET, or Spark version is not supported.
66
-
> - Installing packages from PyPI is not supported within DEP-enabled workspaces.
63
+
64
+
### Manage dependencies for DEP-enabled Synapse Spark pools
65
+
66
+
> [!NOTE]
67
+
>
68
+
> - Installing packages from public repo is not supported within [DEP-enabled workspaces](../security/workspace-data-exfiltration-protection.md), you should upload all your dependencies as workspace libraries and install to your Spark pool.
69
+
>
70
+
Please follow the steps below if you have trouble to identify the required dependencies:
71
+
72
+
-**Step1: Run the following script to set up a local Python environment same with Synapse Spark environment**
73
+
The setup script requires [Synapse-Python38-CPU.yml](https://github.com/Azure-Samples/Synapse/blob/main/Spark/Python/Synapse-Python38-CPU.yml) which is the list of libraries shipped in the default python env in Synapse spark.
-**Step2: Run the following script to identify the required dependencies**
86
+
The below snippet can be used to pass your requirement.txt which has all the packages and version you intend to install in the spark 3.1/spark3.2 spark pool. It will print the names of the *new* wheel files/dependencies needed for your input library requirements. Note this will list out only the dependencies that are not already present in the spark pool by default.
87
+
88
+
```python
89
+
# command to list out wheels needed for your input libraries
90
+
# this command will list out only *new* dependencies that are
91
+
# not already part of the built-in synapse environment
0 commit comments