You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/hdinsight/spark/apache-azure-spark-history-server.md
+59-60Lines changed: 59 additions & 60 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,10 +1,10 @@
1
1
---
2
2
title: Extended Spark History Server to debug Spark applications - Azure HDInsight
3
3
description: Use extended Spark History Server to debug and diagnose Spark applications - Azure HDInsight.
4
-
ms.service: hdinsight
5
4
author: hrasheed-msft
6
5
ms.author: hrasheed
7
6
ms.reviewer: jasonh
7
+
ms.service: hdinsight
8
8
ms.custom: hdinsightactive,hdiseo17may2017
9
9
ms.topic: conceptual
10
10
ms.date: 09/04/2019
@@ -16,16 +16,17 @@ This article provides guidance on how to use extended Apache Spark History Serve
16
16
17
17
## Get access to Apache Spark History Server
18
18
19
-
Apache Spark History Server is the web UI for completed and running Spark applications.
19
+
Apache Spark History Server is the web UI for completed and running Spark applications.
20
20
21
21
### Open the Apache Spark History Server Web UI from Azure portal
22
22
23
23
1. From the [Azure portal](https://portal.azure.com/), open the Spark cluster. For more information, see [List and show clusters](../hdinsight-administer-use-portal-linux.md#showClusters).
24
-
2. From **Quick Links**, click **Cluster Dashboard**, and then click **Spark History Server**. When prompted, enter the admin credentials for the Spark cluster.
24
+
2. From **Quick Links**, click **Cluster Dashboard**, and then click **Spark History Server**. When prompted, enter the admin credentials for the Spark cluster.
25
25
26
-

26
+

27
27
28
28
### Open the Spark History Server Web UI by URL
29
+
29
30
Open the Spark History Server by browsing to the following URL, replace `<ClusterName>` with Spark cluster name of customer.
30
31
31
32
```
@@ -36,67 +37,67 @@ The Spark History Server web UI looks like:
36
37
37
38

38
39
39
-
40
40
## Data tab in Spark History Server
41
+
41
42
Select job ID then click **Data** on the tool menu to get the data view.
42
43
43
44
+ Check the **Inputs**, **Outputs**, and **Table Operations** by selecting the tabs separately.

62
63
63
64
+ Download single file by clicking button **Partial Download** that place at the right, then the selected file will be downloaded to local, if the file does not exist any more, it will open a new tab to show the error messages.

66
67
67
68
+ Copy full path or relative path by selecting the **Copy Full Path**, **Copy Relative Path** that expands from download menu. For azure data lake storage files, **Open in Azure Storage Explorer** will launch Azure Storage Explorer, and locate to the folder when sign-in.

100
101
101
102
+ Play back the job by clicking the **Playback** button and stop anytime by clicking the stop button. The task display in color to show different status when playback:
102
103
@@ -106,30 +107,29 @@ Select job ID then click **Graph** on the tool menu to get the job graph view.
106
107
+ White for waiting or skipped: The task is waiting to run, or the stage has skipped.
107
108
+ Red for failed: The task has failed.
108
109
109
-

110
-
110
+

111
+
111
112
The skipped stage display in white.
112
-

113
+

114
+
115
+

113
116
114
-

115
-
116
117
> [!NOTE]
117
118
> Playback for each job is allowed. For incomplete job, playback is not supported.
118
119
119
-
120
120
+ Mouse scrolls to zoom in/out the job graph, or click **Zoom to fit** to make it fit to screen.
121
-
122
-

121
+
122
+

123
123
124
124
+ Hover on graph node to see the tooltip when there are failed tasks, and click on stage to open stage page.

154
153
155
154
## Diagnosis tab in Apache Spark History Server
155
+
156
156
Select job ID then click **Diagnosis** on the tool menu to get the job Diagnosis view. The diagnosis tab includes **Data Skew**, **Time Skew**, and **Executor Usage Analysis**.
157
-
157
+
158
158
+ Check the **Data Skew**, **Time Skew**, and **Executor Usage Analysis** by selecting the tabs respectively.

161
161
162
162
### Data Skew
163
-
Click **Data Skew** tab, the corresponding skewed tasks are displayed based on the specified parameters.
163
+
164
+
Click **Data Skew** tab, the corresponding skewed tasks are displayed based on the specified parameters.
164
165
165
166
+**Specify Parameters** - The first section displays the parameters which are used to detect Data Skew. The built-in rule is: Task Data Read is greater than 3 times of the average task data read, and the task data read is more than 10MB. If you want to define your own rule for skewed tasks, you can choose your parameters, the **Skewed Stage**, and **Skew Char** section will be refreshed accordingly.
166
167
167
168
+**Skewed Stage** - The second section displays stages which have skewed tasks meeting the criteria specified above. If there are more than one skewed task in a stage, the skewed stage table only displays the most skewed task (e.g. the largest data for data skew).

170
171
171
172
+**Skew Chart** – When a row in the skew stage table is selected, the skew chart displays more task distributions details based on data read and execution time. The skewed tasks are marked in red and the normal tasks are marked in blue. For performance consideration, the chart only displays up to 100 sample tasks. The task details are displayed in right bottom panel.

174
175
175
176
### Time Skew
176
-
The **Time Skew** tab displays skewed tasks based on task execution time.
177
+
178
+
The **Time Skew** tab displays skewed tasks based on task execution time.
177
179
178
180
+**Specify Parameters** - The first section displays the parameters which are used to detect Time Skew. The default criteria to detect time skew is: task execution time is greater than 3 times of average execution time and task execution time is greater than 30 seconds. You can change the parameters based on your needs. The **Skewed Stage** and **Skew Chart** display the corresponding stages and tasks information just like the **Data Skew** tab above.
179
181
180
182
+ Click **Time Skew**, then filtered result is displayed in **Skewed Stage** section according to the parameters set in section **Specify Parameters**. Click one item in **Skewed Stage** section, then the corresponding chart is drafted in section3, and the task details are displayed in right bottom panel.

183
185
184
186
### Executor Usage Analysis
187
+
185
188
The Executor Usage Graph visualizes the Spark job actual executor allocation and running status.
186
189
187
190
+ Click **Executor Usage Analysis**, then four types curves about executor usage are drafted, including **Allocated Executors**, **Running Executors**,**idle Executors**, and **Max Executor Instances**. Regarding allocated executors, each "Executor added" or "Executor removed" event will increase or decrease the allocated executors, you can check "Event Timeline" in the “Jobs" tab for more comparison.
If you have any feedback, or if you encounter any other problems when using this tool, send an email at ([[email protected]](mailto:[email protected])).
0 commit comments