mirror of
https://github.com/bitnami/charts.git
synced 2026-03-10 15:07:49 +08:00
329 lines
8.4 KiB
YAML
329 lines
8.4 KiB
YAML
## Global Docker image parameters
|
|
## Please, note that this will override the image parameters, including dependencies, configured to use the global value
|
|
## Current available global Docker image parameters: imageRegistry and imagePullSecrets
|
|
##
|
|
# global:
|
|
# imageRegistry: myRegistryName
|
|
# imagePullSecrets:
|
|
# - myRegistryKeySecretName
|
|
|
|
## Bitnami Spark image version
|
|
## ref: https://hub.docker.com/r/bitnami/spark/tags/
|
|
##
|
|
image:
|
|
registry: docker.io
|
|
repository: bitnami/spark
|
|
tag: 2.4.5-debian-10-r16
|
|
## Specify a imagePullPolicy
|
|
## Defaults to 'Always' if image tag is 'latest', else set to 'IfNotPresent'
|
|
## ref: http://kubernetes.io/docs/user-guide/images/#pre-pulling-images
|
|
##
|
|
pullPolicy: IfNotPresent
|
|
|
|
## Pull secret for this image
|
|
# pullSecrets:
|
|
# - myRegistryKeySecretName
|
|
|
|
## Set to true if you would like to see extra information on logs
|
|
## It turns BASH and NAMI debugging in minideb
|
|
## ref: https://github.com/bitnami/minideb-extras/#turn-on-bash-debugging
|
|
debug: false
|
|
|
|
## String to partially override spark.fullname template (will maintain the release name)
|
|
##
|
|
# nameOverride:
|
|
|
|
## String to fully override spark.fullname template
|
|
##
|
|
# fullnameOverride:
|
|
|
|
## Spark master specific configuration
|
|
##
|
|
master:
|
|
## Set a custom configuration by using an existing configMap with the configuration file.
|
|
##
|
|
# configurationConfigMap:
|
|
|
|
## Spark container ports
|
|
##
|
|
webPort: 8080
|
|
clusterPort: 7077
|
|
|
|
## Set the master daemon memory limit.
|
|
##
|
|
# daemonMemoryLimit:
|
|
|
|
## Use a string to set the config options for in the form "-Dx=y"
|
|
##
|
|
# configOptions:
|
|
|
|
## An array to add extra env vars
|
|
## For example:
|
|
## extraEnvVars:
|
|
## - name: SPARK_DAEMON_JAVA_OPTS
|
|
## value: -Dx=y
|
|
##
|
|
# extraEnvVars:
|
|
|
|
## Kubernetes Security Context
|
|
## https://kubernetes.io/docs/tasks/configure-pod-container/security-context/
|
|
##
|
|
securityContext:
|
|
enabled: true
|
|
fsGroup: 1001
|
|
runAsUser: 1001
|
|
|
|
## Node labels for pod assignment
|
|
## Ref: https://kubernetes.io/docs/user-guide/node-selection/
|
|
##
|
|
nodeSelector: {}
|
|
|
|
## Tolerations for pod assignment
|
|
## Ref: https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/
|
|
##
|
|
tolerations: []
|
|
|
|
## Affinity for pod assignment
|
|
## Ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#affinity-and-anti-affinity
|
|
##
|
|
affinity: {}
|
|
|
|
## Configure resource requests and limits
|
|
## ref: http://kubernetes.io/docs/user-guide/compute-resources/
|
|
##
|
|
resources:
|
|
# We usually recommend not to specify default resources and to leave this as a conscious
|
|
# choice for the user. This also increases chances charts run on environments with little
|
|
# resources, such as Minikube. If you do want to specify resources, uncomment the following
|
|
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
|
|
limits: {}
|
|
# cpu: 250m
|
|
# memory: 256Mi
|
|
requests: {}
|
|
# cpu: 250m
|
|
# memory: 256Mi
|
|
|
|
## Configure liveness and readiness probes
|
|
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-probes/#configure-probes)
|
|
##
|
|
livenessProbe:
|
|
enabled: true
|
|
initialDelaySeconds: 180
|
|
periodSeconds: 20
|
|
timeoutSeconds: 5
|
|
failureThreshold: 6
|
|
successThreshold: 1
|
|
readinessProbe:
|
|
enabled: true
|
|
initialDelaySeconds: 30
|
|
periodSeconds: 10
|
|
timeoutSeconds: 5
|
|
failureThreshold: 6
|
|
successThreshold: 1
|
|
|
|
## Spark worker specific configuration
|
|
##
|
|
worker:
|
|
## Set a custom configuration by using an existing configMap with the configuration file.
|
|
##
|
|
# configurationConfigMap:
|
|
|
|
## Spark container ports
|
|
##
|
|
webPort: 8081
|
|
# clusterPort:
|
|
|
|
## Set the daemonMemoryLimit as the daemon max memory
|
|
##
|
|
# daemonMemoryLimit:
|
|
|
|
## Set the worker memory limit
|
|
##
|
|
# memoryLimit:
|
|
|
|
## Set the maximun number of cores
|
|
##
|
|
# coreLimit:
|
|
|
|
## Working directory for the application
|
|
##
|
|
# dir:
|
|
|
|
## Options for the JVM as "-Dx=y"
|
|
##
|
|
# javaOptions:
|
|
|
|
## Configuraion options in the form "-Dx=y"
|
|
##
|
|
# configOptions:
|
|
|
|
## An array to add extra env vars
|
|
## For example:
|
|
## extraEnvVars:
|
|
## - name: SPARK_DAEMON_JAVA_OPTS
|
|
## value: -Dx=y
|
|
# extraEnvVars:
|
|
|
|
## Number of spark workers (will be the min number when autoscaling is enabled)
|
|
##
|
|
replicaCount: 2
|
|
|
|
## Kubernetes Security Context
|
|
## https://kubernetes.io/docs/tasks/configure-pod-container/security-context/
|
|
##
|
|
securityContext:
|
|
enabled: true
|
|
fsGroup: 1001
|
|
runAsUser: 1001
|
|
|
|
## Node labels for pod assignment
|
|
## Ref: https://kubernetes.io/docs/user-guide/node-selection/
|
|
##
|
|
nodeSelector: {}
|
|
|
|
## Tolerations for pod assignment
|
|
## Ref: https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/
|
|
##
|
|
tolerations: []
|
|
|
|
## Affinity for pod assignment
|
|
## Ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#affinity-and-anti-affinity
|
|
##
|
|
affinity: {}
|
|
|
|
## Configure resource requests and limits
|
|
## ref: http://kubernetes.io/docs/user-guide/compute-resources/
|
|
##
|
|
resources:
|
|
# We usually recommend not to specify default resources and to leave this as a conscious
|
|
# choice for the user. This also increases chances charts run on environments with little
|
|
# resources, such as Minikube. If you do want to specify resources, uncomment the following
|
|
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
|
|
limits: {}
|
|
# cpu: 250m
|
|
# memory: 256Mi
|
|
requests: {}
|
|
# cpu: 250m
|
|
# memory: 256Mi
|
|
|
|
## Configure liveness and readiness probes
|
|
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-probes/#configure-probes)
|
|
##
|
|
livenessProbe:
|
|
enabled: true
|
|
initialDelaySeconds: 180
|
|
periodSeconds: 20
|
|
timeoutSeconds: 5
|
|
failureThreshold: 6
|
|
successThreshold: 1
|
|
readinessProbe:
|
|
enabled: true
|
|
initialDelaySeconds: 30
|
|
periodSeconds: 10
|
|
timeoutSeconds: 5
|
|
failureThreshold: 6
|
|
successThreshold: 1
|
|
|
|
## Array to add extra volumes
|
|
##
|
|
## extraVolumes:
|
|
## Array to add extra mounts (normally used with extraVolumes)
|
|
##
|
|
## extraVolumeMounts: {}
|
|
## Autoscaling parameters
|
|
##
|
|
autoscaling:
|
|
## Enable replica autoscaling depending on CPU
|
|
enabled: false
|
|
CpuTargetPercentage: 50
|
|
## Max number of workers when using autoscaling
|
|
replicasMax: 5
|
|
|
|
## Security configuration
|
|
##
|
|
security:
|
|
## Name of the secret that contains all the passwords. This is optional, by default random passwords are generated.
|
|
##
|
|
# passwordsSecretName:
|
|
|
|
## RPC configuration
|
|
##
|
|
rpc:
|
|
authenticationEnabled: false
|
|
encryptionEnabled: false
|
|
|
|
## Enables local storage encryption
|
|
##
|
|
storageEncryptionEnabled: false
|
|
|
|
## Name of the secret that contains the certificates
|
|
## It should contains two keys called "spark-keystore.jks" and "spark-truststore.jks" with the files in JKS format.
|
|
##
|
|
# certificatesSecretName:
|
|
|
|
## SSL configuration
|
|
##
|
|
ssl:
|
|
enabled: false
|
|
needClientAuth: false
|
|
protocol: TLSv1.2
|
|
|
|
## Service parameters
|
|
##
|
|
service:
|
|
## Kubernetes service type
|
|
##
|
|
type: ClusterIP
|
|
|
|
## Cluster Service port
|
|
##
|
|
clusterPort: 7077
|
|
|
|
## Web Service port
|
|
##
|
|
webPort: 80
|
|
|
|
## Specify the nodePort(s) value(s) for the LoadBalancer and NodePort service types.
|
|
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport
|
|
##
|
|
nodePorts:
|
|
cluster: ""
|
|
web: ""
|
|
|
|
## Set the LoadBalancer service type to internal only.
|
|
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#internal-load-balancer
|
|
##
|
|
# loadBalancerIP:
|
|
|
|
## Provide any additional annotations which may be required. This can be used to
|
|
## set the LoadBalancer service type to internal only.
|
|
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#internal-load-balancer
|
|
##
|
|
annotations: {}
|
|
|
|
## Ingress paramaters
|
|
##
|
|
ingress:
|
|
## Set to true to enable ingress record generation
|
|
##
|
|
enabled: false
|
|
|
|
## Set this to true in order to add the corresponding annotations for cert-manager
|
|
##
|
|
certManager: false
|
|
|
|
## If certManager is set to true, annotation kubernetes.io/tls-acme: "true" will automatically be set
|
|
##
|
|
annotations: {}
|
|
|
|
## The list of hostnames to be covered with this ingress record.
|
|
## Most likely this will be just one host, but in the event more hosts are needed, this is an array
|
|
##
|
|
hosts:
|
|
- name: spark.local
|
|
path: /
|
|
tls:
|
|
- hosts:
|
|
- spark.domain.com
|
|
secretName: spark.local-tls
|