You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
## What changes were proposed in this pull request?
Updated documentation for Spark on Kubernetes for the upcoming 2.4.0.
Please review http://spark.apache.org/contributing.html before opening a pull request.
mccheah erikerlandson
Closesapache#22224 from liyinan926/master.
Authored-by: Yinan Li <[email protected]>
Signed-off-by: Sean Owen <[email protected]>
Starting with Spark 2.4.0, users can mount the following types of Kubernetes [volumes](https://kubernetes.io/docs/concepts/storage/volumes/) into the driver and executor pods:
191
+
*[hostPath](https://kubernetes.io/docs/concepts/storage/volumes/#hostpath): mounts a file or directory from the host node’s filesystem into a pod.
192
+
*[emptyDir](https://kubernetes.io/docs/concepts/storage/volumes/#emptydir): an initially empty volume created when a pod is assigned to a node.
193
+
*[persistentVolumeClaim](https://kubernetes.io/docs/concepts/storage/volumes/#persistentvolumeclaim): used to mount a `PersistentVolume` into a pod.
194
+
195
+
To mount a volume of any of the types above into the driver pod, use the following configuration property:
Specifically, `VolumeType` can be one of the following values: `hostPath`, `emptyDir`, and `persistentVolumeClaim`. `VolumeName` is the name you want to use for the volume under the `volumes` field in the pod specification.
203
+
204
+
Each supported type of volumes may have some specific configuration options, which can be specified using configuration properties of the following form:
The configuration properties for mounting volumes into the executor pods use prefix `spark.kubernetes.executor.` instead of `spark.kubernetes.driver.`. For a complete list of available options for each supported type of volumes, please refer to the [Spark Properties](#spark-properties) section below.
217
+
188
218
## Introspection and Debugging
189
219
190
220
These are the different ways in which you can investigate a running/completed Spark application, monitor progress, and
@@ -299,21 +329,15 @@ RBAC authorization and how to configure Kubernetes service accounts for pods, pl
299
329
300
330
## Future Work
301
331
302
-
There are several Spark on Kubernetes features that are currently being incubated in a fork -
303
-
[apache-spark-on-k8s/spark](https://github.com/apache-spark-on-k8s/spark), which are expected to eventually make it into
304
-
future versions of the spark-kubernetes integration.
332
+
There are several Spark on Kubernetes features that are currently being worked on or planned to be worked on. Those features are expected to eventually make it into future versions of the spark-kubernetes integration.
305
333
306
334
Some of these include:
307
335
308
-
* R
309
-
* Dynamic Executor Scaling
336
+
* Dynamic Resource Allocation and External Shuffle Service
310
337
* Local File Dependency Management
311
338
* Spark Application Management
312
339
* Job Queues and Resource Management
313
340
314
-
You can refer to the [documentation](https://apache-spark-on-k8s.github.io/userdocs/) if you want to try these features
315
-
and provide feedback to the development team.
316
-
317
341
# Configuration
318
342
319
343
See the [configuration page](configuration.html) for information on Spark configurations. The following configurations are
0 commit comments