Skip to content

Commit 28c9dd4

Browse files
authored
Merge pull request #206288 from kevinjaku/main
adding troubleshooting Synapse Spark library management failures through Spark History Server
2 parents 9ab2665 + 22d8566 commit 28c9dd4

File tree

6 files changed

+18
-2
lines changed

6 files changed

+18
-2
lines changed

articles/synapse-analytics/spark/apache-spark-troubleshoot-library-errors.md

Lines changed: 18 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,6 +36,22 @@ To view these logs:
3636
4. Within the results, you'll see the logs related to the installation of your packages.
3737
![Screenshot that highlights system reserved library job results.](./media/apache-spark-azure-portal-add-libraries/system-reserved-library-job-results.png "View system library job progress")
3838

39+
## Track installation failures
40+
In certain cases, users can also inspect the full installation logs available in the Spark History Server to identify complicated dependency conflicts. The logs available through the Spark UI could be truncated and accessing the full installation logs through the Spark History Server would be useful in complex library installation scenarios.
41+
42+
To view the full installation logs:
43+
1. Navigate to the Spark applications list in the **Monitor** tab.
44+
2. Select the system Spark application job that corresponds to the failed pool update. These system jobs run under the *SystemReservedJob-LibraryManagement* title.
45+
![Screenshot that highlights the failed system reserved library job.](./media/apache-spark-azure-portal-add-libraries/system-reserved-library-job-failure.png "View failed system library job")
46+
3. Select the highlighted **Spark history server** option which would open the Spark history server details page in a new tab.
47+
![Screenshot that highlights the details of the failed system reserved library job.](./media/apache-spark-azure-portal-add-libraries/system-reserved-library-job-failure-details.png "View details of failed system library job")
48+
4. In this page, you will see 2 attempts, select **Attempt 1** as shown below.
49+
![Screenshot that highlights the executor details in the spark history server page for the failed system reserved library job.](./media/apache-spark-azure-portal-add-libraries/spark-history-server-executors.png "View executor detaols in spark history server page")
50+
5. On the top navigation bar in the Spark history server page, switch to the **Executors** tab.
51+
![Screenshot that highlights the job details in the spark history server page for the failed system reserved library job.](./media/apache-spark-azure-portal-add-libraries/spark-history-server-page.png "View the job details in the spark history server page")
52+
6. Download the **stdout** and **stderr** log files to access the full library management output and error logs.
53+
![Screenshot that highlights the spark history server page for the failed system reserved library job.](./media/apache-spark-azure-portal-add-libraries/spark-history-server-executors-details.png "View stdout and stderr logs in the spark history server page")
54+
3955
## Validate your permissions
4056
To install and update libraries, you must have the **Storage Blob Data Contributor** or **Storage Blob Data Owner** permissions on the primary Azure Data Lake Storage Gen2 Storage account that is linked to the Azure Synapse Analytics workspace.
4157

@@ -93,10 +109,10 @@ To recreate the environment and validate your updates:
93109
conda activate myenv
94110
```
95111
96-
3. Use ``pip install -r <provide your req.txt file>`` to update the virtual environment with your specified packages. If the installation results in an error, then there may be a a conflict between what is pre-installed in the Synapse base runtime and what is specified in the provided requirements file. These dependency conflicts must be resolved in order to get the updated libraries on your serverless Apache Spark pool.
112+
3. Use ``pip install -r <provide your req.txt file>`` to update the virtual environment with your specified packages. If the installation results in an error, then there may be a conflict between what is pre-installed in the Synapse base runtime and what is specified in the provided requirements file. These dependency conflicts must be resolved in order to get the updated libraries on your serverless Apache Spark pool.
97113
98114
>[!IMPORTANT]
99115
>Issues may arrise when using pip and conda together. When combining pip and conda, it's best to follow these [recommended best practices](https://conda.io/projects/conda/en/latest/user-guide/tasks/manage-environments.html#activating-an-environment).
100116
101117
## Next steps
102-
- View the default libraries: [Apache Spark version support](apache-spark-version-support.md)
118+
- View the default libraries: [Apache Spark version support](apache-spark-version-support.md)
Loading
Loading
81.7 KB
Loading
Loading
Loading

0 commit comments

Comments
 (0)