From a03ac08e094eb02288c3c9bfe4dfda8d4aca49dd Mon Sep 17 00:00:00 2001 From: Dongjoon Hyun Date: Wed, 16 Jul 2025 13:36:14 -0700 Subject: [PATCH] [SPARK-52827] Add `pi-with-template.yaml` example --- examples/pi-with-template.yaml | 41 ++++++++++++++++++++++++++++++++++ 1 file changed, 41 insertions(+) create mode 100644 examples/pi-with-template.yaml diff --git a/examples/pi-with-template.yaml b/examples/pi-with-template.yaml new file mode 100644 index 00000000..5570fad4 --- /dev/null +++ b/examples/pi-with-template.yaml @@ -0,0 +1,41 @@ +# Licensed to the Apache Software Foundation (ASF) under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +apiVersion: spark.apache.org/v1 +kind: SparkApplication +metadata: + name: pi-with-template +spec: + mainClass: "org.apache.spark.examples.SparkPi" + jars: "local:///opt/spark/examples/jars/spark-examples.jar" + sparkConf: + spark.dynamicAllocation.enabled: "true" + spark.dynamicAllocation.shuffleTracking.enabled: "true" + spark.dynamicAllocation.maxExecutors: "3" + spark.kubernetes.authenticate.driver.serviceAccountName: "spark" + spark.kubernetes.container.image: "apache/spark:4.0.0-scala" + applicationTolerations: + resourceRetainPolicy: OnFailure + driverSpec: + podTemplateSpec: + spec: + priorityClassName: system-cluster-critical + terminationGracePeriodSeconds: 0 + executorSpec: + podTemplateSpec: + spec: + priorityClassName: system-cluster-critical + terminationGracePeriodSeconds: 0 + runtimeVersions: + sparkVersion: "4.0.0"