spark MountVolumesFeatureStep 源码
spark MountVolumesFeatureStep 代码
文件路径:/resource-managers/kubernetes/core/src/main/scala/org/apache/spark/deploy/k8s/features/MountVolumesFeatureStep.scala
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.spark.deploy.k8s.features
import scala.collection.JavaConverters._
import scala.collection.mutable.ArrayBuffer
import io.fabric8.kubernetes.api.model._
import org.apache.spark.deploy.k8s._
import org.apache.spark.deploy.k8s.Constants.{ENV_EXECUTOR_ID, SPARK_APP_ID_LABEL}
import org.apache.spark.internal.config.EXECUTOR_INSTANCES
private[spark] class MountVolumesFeatureStep(conf: KubernetesConf)
extends KubernetesFeatureConfigStep {
import MountVolumesFeatureStep._
val additionalResources = ArrayBuffer.empty[HasMetadata]
override def configurePod(pod: SparkPod): SparkPod = {
val (volumeMounts, volumes) = constructVolumes(conf.volumes).unzip
val podWithVolumes = new PodBuilder(pod.pod)
.editSpec()
.addToVolumes(volumes.toSeq: _*)
.endSpec()
.build()
val containerWithVolumeMounts = new ContainerBuilder(pod.container)
.addToVolumeMounts(volumeMounts.toSeq: _*)
.build()
SparkPod(podWithVolumes, containerWithVolumeMounts)
}
private def constructVolumes(
volumeSpecs: Iterable[KubernetesVolumeSpec]
): Iterable[(VolumeMount, Volume)] = {
val duplicateMountPaths = volumeSpecs.map(_.mountPath).toSeq.groupBy(identity).collect {
case (x, ys) if ys.length > 1 => s"'$x'"
}
require(duplicateMountPaths.isEmpty,
s"Found duplicated mountPath: ${duplicateMountPaths.mkString(", ")}")
volumeSpecs.zipWithIndex.map { case (spec, i) =>
val volumeMount = new VolumeMountBuilder()
.withMountPath(spec.mountPath)
.withReadOnly(spec.mountReadOnly)
.withSubPath(spec.mountSubPath)
.withName(spec.volumeName)
.build()
val volumeBuilder = spec.volumeConf match {
case KubernetesHostPathVolumeConf(hostPath) =>
/* "" means that no checks will be performed before mounting the hostPath volume */
new VolumeBuilder()
.withHostPath(new HostPathVolumeSource(hostPath, ""))
case KubernetesPVCVolumeConf(claimNameTemplate, storageClass, size) =>
val claimName = conf match {
case c: KubernetesExecutorConf =>
checkPVCClaimName(claimNameTemplate)
claimNameTemplate
.replaceAll(PVC_ON_DEMAND,
s"${conf.resourceNamePrefix}-exec-${c.executorId}$PVC_POSTFIX-$i")
.replaceAll(ENV_EXECUTOR_ID, c.executorId)
case _ =>
claimNameTemplate
.replaceAll(PVC_ON_DEMAND, s"${conf.resourceNamePrefix}-driver$PVC_POSTFIX-$i")
}
if (storageClass.isDefined && size.isDefined) {
additionalResources.append(new PersistentVolumeClaimBuilder()
.withKind(PVC)
.withApiVersion("v1")
.withNewMetadata()
.withName(claimName)
.addToLabels(SPARK_APP_ID_LABEL, conf.appId)
.endMetadata()
.withNewSpec()
.withStorageClassName(storageClass.get)
.withAccessModes(PVC_ACCESS_MODE)
.withResources(new ResourceRequirementsBuilder()
.withRequests(Map("storage" -> new Quantity(size.get)).asJava).build())
.endSpec()
.build())
}
new VolumeBuilder()
.withPersistentVolumeClaim(
new PersistentVolumeClaimVolumeSource(claimName, spec.mountReadOnly))
case KubernetesEmptyDirVolumeConf(medium, sizeLimit) =>
new VolumeBuilder()
.withEmptyDir(
new EmptyDirVolumeSource(medium.getOrElse(""),
sizeLimit.map(new Quantity(_)).orNull))
case KubernetesNFSVolumeConf(path, server) =>
new VolumeBuilder()
.withNfs(new NFSVolumeSource(path, null, server))
}
val volume = volumeBuilder.withName(spec.volumeName).build()
(volumeMount, volume)
}
}
override def getAdditionalKubernetesResources(): Seq[HasMetadata] = {
additionalResources.toSeq
}
private def checkPVCClaimName(claimName: String): Unit = {
val executorInstances = conf.get(EXECUTOR_INSTANCES)
if (executorInstances.isDefined && executorInstances.get > 1) {
// PVC ClaimName should contain OnDemand or SPARK_EXECUTOR_ID
// when requiring multiple executors.
// Else, spark continues to try to create the executor pod.
if (!claimName.contains(PVC_ON_DEMAND) && !claimName.contains(ENV_EXECUTOR_ID)) {
throw new IllegalArgumentException(s"PVC ClaimName: $claimName " +
s"should contain $PVC_ON_DEMAND or $ENV_EXECUTOR_ID " +
"when requiring multiple executors")
}
}
}
}
private[spark] object MountVolumesFeatureStep {
val PVC_ON_DEMAND = "OnDemand"
val PVC = "PersistentVolumeClaim"
val PVC_POSTFIX = "-pvc"
val PVC_ACCESS_MODE = "ReadWriteOnce"
}
相关信息
相关文章
spark BasicDriverFeatureStep 源码
spark BasicExecutorFeatureStep 源码
spark DriverCommandFeatureStep 源码
spark DriverKubernetesCredentialsFeatureStep 源码
spark DriverServiceFeatureStep 源码
spark EnvSecretsFeatureStep 源码
spark ExecutorKubernetesCredentialsFeatureStep 源码
spark HadoopConfDriverFeatureStep 源码
0
赞
- 所属分类: 前端技术
- 本文标签:
热门推荐
-
2、 - 优质文章
-
3、 gate.io
-
7、 golang
-
9、 openharmony
-
10、 Vue中input框自动聚焦