Files
charts/bitnami/zookeeper/values.yaml
Bitnami Containers be5e16f11f [bitnami/zookeeper] Release 7.4.9 updating components versions
Signed-off-by: Bitnami Containers <containers@bitnami.com>
2021-10-25 08:38:58 +00:00

724 lines
28 KiB
YAML

## @section Global parameters
## Global Docker image parameters
## Please, note that this will override the image parameters, including dependencies, configured to use the global value
## Current available global Docker image parameters: imageRegistry, imagePullSecrets and storageClass
##
## @param global.imageRegistry Global Docker image registry
## @param global.imagePullSecrets Global Docker registry secret names as an array
## @param global.storageClass Global StorageClass for Persistent Volume(s)
##
global:
imageRegistry: ""
## E.g.
## imagePullSecrets:
## - myRegistryKeySecretName
##
imagePullSecrets: []
storageClass: ""
## @section Common parameters
##
## @param nameOverride String to partially override common.names.fullname template (will maintain the release name)
##
nameOverride: ""
## @param fullnameOverride String to fully override common.names.fullname template
##
fullnameOverride: ""
## @param clusterDomain Kubernetes Cluster Domain
##
clusterDomain: cluster.local
## @param extraDeploy Extra objects to deploy (value evaluated as a template)
##
extraDeploy: []
## @param commonLabels Add labels to all the deployed resources
##
commonLabels: {}
## @param commonAnnotations Add annotations to all the deployed resources
##
commonAnnotations: {}
## Enable diagnostic mode in the deployment
##
diagnosticMode:
## @param diagnosticMode.enabled Enable diagnostic mode (all probes will be disabled and the command will be overridden)
##
enabled: false
## @param diagnosticMode.command Command to override all containers in the deployment
##
command:
- sleep
## @param diagnosticMode.args Args to override all containers in the deployment
##
args:
- infinity
## @section Zookeeper chart parameters
##
## Bitnami Zookeeper image version
## ref: https://hub.docker.com/r/bitnami/zookeeper/tags/
## @param image.registry ZooKeeper image registry
## @param image.repository ZooKeeper image repository
## @param image.tag ZooKeeper Image tag (immutable tags are recommended)
## @param image.pullPolicy ZooKeeper image pull policy
## @param image.pullSecrets Specify docker-registry secret names as an array
## @param image.debug Specify if debug values should be set
##
image:
registry: docker.io
repository: bitnami/zookeeper
tag: 3.7.0-debian-10-r185
## Specify a imagePullPolicy
## Defaults to 'Always' if image tag is 'latest', else set to 'IfNotPresent'
## ref: http://kubernetes.io/docs/user-guide/images/#pre-pulling-images
##
pullPolicy: IfNotPresent
## Optionally specify an array of imagePullSecrets.
## Secrets must be manually created in the namespace.
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry/
## Example:
## pullSecrets:
## - myRegistryKeySecretName
##
pullSecrets: []
## Set to true if you would like to see extra information on logs
## It turns BASH and/or NAMI debugging in the image
##
debug: false
## @param tickTime Basic time unit in milliseconds used by ZooKeeper for heartbeats
##
tickTime: 2000
## @param initLimit ZooKeeper uses to limit the length of time the ZooKeeper servers in quorum have to connect to a leader
##
initLimit: 10
## @param syncLimit How far out of date a server can be from a leader
##
syncLimit: 5
## @param preAllocSize Block size for transaction log file
##
preAllocSize: 65536
## @param snapCount The number of transactions recorded in the transaction log before a snapshot can be taken (and the transaction log rolled)
##
snapCount: 100000
## @param maxClientCnxns Limits the number of concurrent connections that a single client may make to a single member of the ZooKeeper ensemble
##
maxClientCnxns: 60
## @param fourlwCommandsWhitelist A list of comma separated Four Letter Words commands to use
##
fourlwCommandsWhitelist: srvr, mntr, ruok
## @param listenOnAllIPs Allow Zookeeper to listen for connections from its peers on all available IP addresses
##
listenOnAllIPs: false
## @param allowAnonymousLogin Allow to accept connections from unauthenticated users
##
allowAnonymousLogin: true
autopurge:
## @param autopurge.snapRetainCount Retains the snapRetainCount most recent snapshots and the corresponding transaction logs and deletes the rest
##
snapRetainCount: 3
## @param autopurge.purgeInterval The time interval in hours for which the purge task has to be triggered
## Set to a positive integer (1 and above) to enable the auto purging
##
purgeInterval: 0
## @param maxSessionTimeout Maximum session timeout in milliseconds that the server will allow the client to negotiate
## Defaults to 20 times the tickTime
##
maxSessionTimeout: 40000
auth:
## @param auth.existingSecret Use existing secret (ignores previous password)
##
existingSecret: ""
## @param auth.enabled Enable Zookeeper auth. It uses SASL/Digest-MD5
##
enabled: false
## @param auth.clientUser User that will use ZooKeeper clients to auth
##
clientUser: ""
## @param auth.clientPassword Password that will use ZooKeeper clients to auth
##
clientPassword: ""
## @param auth.serverUsers Comma, semicolon or whitespace separated list of user to be created
## Specify them as a string, for example: "user1,user2,admin"
##
serverUsers: ""
## @param auth.serverPasswords Comma, semicolon or whitespace separated list of passwords to assign to users when created
## Specify them as a string, for example: "pass4user1, pass4user2, pass4admin"
##
serverPasswords: ""
## @param heapSize Size in MB for the Java Heap options (Xmx and XMs)
## This env var is ignored if Xmx an Xms are configured via JVMFLAGS
##
heapSize: 1024
## @param logLevel Log level for the Zookeeper server. ERROR by default
## Have in mind if you set it to INFO or WARN the ReadinessProve will produce a lot of logs
##
logLevel: ERROR
## @param dataLogDir Data log directory. Specifying this option will direct zookeeper to write the transaction log to the dataLogDir rather than the dataDir.
## This allows a dedicated log device to be used, and helps avoid competition between logging and snaphots.
## Example:
## dataLogDir: /bitnami/zookeeper/dataLog
##
dataLogDir: ""
## @param jvmFlags Default JVMFLAGS for the ZooKeeper process
##
jvmFlags: ""
## @param config Configure ZooKeeper with a custom zoo.cfg file
##
config: ""
## @param namespaceOverride Namespace for ZooKeeper resources
##
namespaceOverride: ""
## @param hostAliases Deployment pod host aliases
## https://kubernetes.io/docs/concepts/services-networking/add-entries-to-pod-etc-hosts-with-host-aliases/
##
hostAliases: []
## @section Statefulset parameters
##
## Extra volumes and extra volume mounts allows you to mount other volumes
## Example Use Cases: mount certificates to enable tls
## @param extraVolumes Extra volumes
## e.g:
## extraVolumes:
## - name: zookeeper-keystore
## secret:
## defaultMode: 288
## secretName: zookeeper-keystore
## - name: zookeeper-trustsore
## secret:
## defaultMode: 288
## secretName: zookeeper-truststore
##
extraVolumes: []
## @param extraVolumeMounts Mount extra volume(s)
## e.g:
## - name: zookeeper-keystore
## mountPath: /certs/keystore
## readOnly: true
## - name: zookeeper-truststore
## mountPath: /certs/truststore
## readOnly: true
##
extraVolumeMounts: []
## @param updateStrategy StatefulSet controller supports automated updates. There are two valid update strategies: `RollingUpdate` and `OnDelete`
## ref: https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#updating-statefulsets
##
updateStrategy: RollingUpdate
## Limits the number of pods of the replicated application that are down simultaneously from voluntary disruptions
## The PDB will only be created if replicaCount is greater than 1
## ref: https://kubernetes.io/docs/concepts/workloads/pods/disruptions
## @param podDisruptionBudget.maxUnavailable Max number of pods down simultaneously
##
podDisruptionBudget:
maxUnavailable: 1
## @param rollingUpdatePartition Partition update strategy
## https://kubernetes.io/docs/concepts/workloads/controllers/statefulset/#partitions
##
rollingUpdatePartition: ""
## @param podManagementPolicy StatefulSet controller supports relax its ordering guarantees while preserving its uniqueness and identity guarantees. There are two valid pod management policies: `OrderedReady` and `Parallel`
## ref: https://kubernetes.io/docs/tutorials/stateful-application/basic-stateful-set/#pod-management-policy
##
podManagementPolicy: Parallel
## @param replicaCount Number of ZooKeeper nodes
##
replicaCount: 1
## @param containerPort ZooKeeper port
##
containerPort: 2181
## @param tlsContainerPort ZooKeeper TLS port
##
tlsContainerPort: 3181
## @param followerContainerPort ZooKeeper follower port
##
followerContainerPort: 2888
## @param electionContainerPort ZooKeeper election port
##
electionContainerPort: 3888
## @param minServerId Minimal SERVER_ID value, nodes increment their IDs respectively
## Servers increment their ID starting at this minimal value.
## E.g., with `minServerId=10` and 3 replicas, server IDs will be 10, 11, 12 for z-0, z-1 and z-2 respectively.
##
minServerId: 1
## Zookeeper Pod Security Context
## @param securityContext.enabled Enable security context (ZooKeeper master pod)
## @param securityContext.fsGroup Group ID for the container (ZooKeeper master pod)
## @param securityContext.runAsUser User ID for the container (ZooKeeper master pod)
##
securityContext:
enabled: true
fsGroup: 1001
runAsUser: 1001
## @param initContainers Extra init container to add to the statefulset
## Example:
## initContainers:
## - name: your-image-name
## image: your-image
## imagePullPolicy: Always
## ports:
## - name: portname
## containerPort: 1234
##
initContainers: []
## @param podAffinityPreset Pod affinity preset. Ignored if `affinity` is set. Allowed values: `soft` or `hard`
## ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#inter-pod-affinity-and-anti-affinity
##
podAffinityPreset: ""
## @param podAntiAffinityPreset Pod anti-affinity preset. Ignored if `affinity` is set. Allowed values: `soft` or `hard`
## Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#inter-pod-affinity-and-anti-affinity
##
podAntiAffinityPreset: soft
## Node affinity preset
## Ref: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#node-affinity
##
nodeAffinityPreset:
## @param nodeAffinityPreset.type Node affinity preset type. Ignored if `affinity` is set. Allowed values: `soft` or `hard`
##
type: ""
## @param nodeAffinityPreset.key Node label key to match Ignored if `affinity` is set.
## E.g.
## key: "kubernetes.io/e2e-az-name"
##
key: ""
## @param nodeAffinityPreset.values Node label values to match. Ignored if `affinity` is set.
## E.g.
## values:
## - e2e-az1
## - e2e-az2
##
values: []
## @param affinity Affinity for pod assignment
## Ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#affinity-and-anti-affinity
## Note: podAffinityPreset, podAntiAffinityPreset, and nodeAffinityPreset will be ignored when it's set
##
affinity: {}
## @param nodeSelector Node labels for pod assignment
## Ref: https://kubernetes.io/docs/user-guide/node-selection/
##
nodeSelector: {}
## @param tolerations Tolerations for pod assignment
## Ref: https://kubernetes.io/docs/concepts/configuration/taint-and-toleration/
##
tolerations: []
## @param topologySpreadConstraints Topology Spread Constraints for pod assignment spread across your cluster among failure-domains. Evaluated as a template
## Ref: https://kubernetes.io/docs/concepts/workloads/pods/pod-topology-spread-constraints/#spread-constraints-for-pods
##
topologySpreadConstraints: {}
## @param podLabels ZooKeeper pod labels
## Ref: https://kubernetes.io/docs/concepts/overview/working-with-objects/labels/
##
podLabels: {}
## @param podAnnotations ZooKeeper Pod annotations
## Ref: https://kubernetes.io/docs/concepts/overview/working-with-objects/annotations/
##
podAnnotations: {}
## @param priorityClassName Name of the existing priority class to be used by ZooKeeper pods, priority class needs to be created beforehand
## Ref: https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/
##
priorityClassName: ""
## @param schedulerName Kubernetes pod scheduler registry
## https://kubernetes.io/docs/tasks/administer-cluster/configure-multiple-schedulers/
##
schedulerName: ""
## Configure resource requests and limits
## ref: http://kubernetes.io/docs/user-guide/compute-resources/
## @param resources.requests [object] The requested resources for the container
##
resources:
requests:
memory: 256Mi
cpu: 250m
## Configure extra options for liveness probe
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-probes/#configure-probes
## @param livenessProbe.enabled Enable livenessProbe
## @param livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe
## @param livenessProbe.periodSeconds Period seconds for livenessProbe
## @param livenessProbe.timeoutSeconds Timeout seconds for livenessProbe
## @param livenessProbe.failureThreshold Failure threshold for livenessProbe
## @param livenessProbe.successThreshold Success threshold for livenessProbe
## @param livenessProbe.probeCommandTimeout Probe command timeout for livenessProbe
##
livenessProbe:
enabled: true
initialDelaySeconds: 30
periodSeconds: 10
timeoutSeconds: 5
failureThreshold: 6
successThreshold: 1
probeCommandTimeout: 2
## Configure extra options for readiness probe
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/configure-liveness-readiness-probes/#configure-probes
## @param readinessProbe.enabled Enable readinessProbe
## @param readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe
## @param readinessProbe.periodSeconds Period seconds for readinessProbe
## @param readinessProbe.timeoutSeconds Timeout seconds for readinessProbe
## @param readinessProbe.failureThreshold Failure threshold for readinessProbe
## @param readinessProbe.successThreshold Success threshold for readinessProbe
## @param readinessProbe.probeCommandTimeout Probe command timeout for readinessProbe
##
readinessProbe:
enabled: true
initialDelaySeconds: 5
periodSeconds: 10
timeoutSeconds: 5
failureThreshold: 6
successThreshold: 1
probeCommandTimeout: 2
## @param customLivenessProbe Override default liveness probe
##
customLivenessProbe: {}
## @param customReadinessProbe Override default readiness probe
##
customReadinessProbe: {}
## @section Traffic Exposure parameters
##
## Kubernetes configuration
## For minikube, set this to NodePort, elsewhere use LoadBalancer
##
service:
## @param service.type Kubernetes Service type
##
type: ClusterIP
## @param service.loadBalancerIP Load balancer IP for the Zookeper Service (optional, cloud specific)
## ref: http://kubernetes.io/docs/user-guide/services/#type-loadbalancer
##
loadBalancerIP: ""
## @param service.port ZooKeeper port
##
port: 2181
## @param service.followerPort ZooKeeper follower port
##
followerPort: 2888
## @param service.electionPort ZooKeeper election port
##
electionPort: 3888
## @param service.nodePorts [object] Specify the nodePort value for the LoadBalancer and NodePort service types.
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport
##
nodePorts:
client: ""
clientTls: ""
## @param service.publishNotReadyAddresses If the ZooKeeper headless service should publish DNS records for not ready pods
##
publishNotReadyAddresses: true
## @param service.tlsClientPort Service port for tls client connections
## Previously service.tls.service_port
##
tlsClientPort: 3181
## @param service.disableBaseClientPort Remove client port from service definitions.
##
disableBaseClientPort: false
## @param service.annotations Annotations for the Service
## Previously service.tls.disable_base_client_port
##
annotations: {}
## @param service.headless.annotations Annotations for the Headless Service
##
headless:
annotations: {}
## Service account for Zookeeper to use.
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/
##
serviceAccount:
## @param serviceAccount.create Enable creation of ServiceAccount for Zookeeper pod
##
create: false
## @param serviceAccount.name The name of the ServiceAccount to use.
## If not set and create is true, a name is generated using the common.names.fullname template
##
name: ""
## @param serviceAccount.automountServiceAccountToken Allows auto mount of ServiceAccountToken on the serviceAccount created
## Can be set to false if pods using this serviceAccount do not need to use K8s API
##
automountServiceAccountToken: true
## Network policies
## Ref: https://kubernetes.io/docs/concepts/services-networking/network-policies/
##
## @param sidecars Extra containers to the pod
## For example:
## sidecars:
## - name: your-image-name
## image: your-image
## imagePullPolicy: Always
## ports:
## - name: portname
## containerPort: 1234
sidecars: []
networkPolicy:
## @param networkPolicy.enabled Specifies whether a NetworkPolicy should be created
##
enabled: false
## @param networkPolicy.allowExternal Don't require client label for connections
## When set to false, only pods with the correct client label will have network access to the port Redis&trade; is
## listening on. When true, zookeeper accept connections from any source (with the correct destination port).
##
allowExternal: true
## @section Persistence parameters
##
## Zookeeper data Persistent Volume Storage Class
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
## If undefined (the default) or set to null, no storageClassName spec is
## set, choosing the default provisioner. (gp2 on AWS, standard on
## GKE, AWS & OpenStack)
##
persistence:
## @param persistence.existingClaim Provide an existing `PersistentVolumeClaim`
## If defined, PVC must be created manually before volume will be bound
## The value is evaluated as a template
##
existingClaim: ""
## @param persistence.enabled Enable Zookeeper data persistence using PVC
##
enabled: true
## @param persistence.storageClass PVC Storage Class for ZooKeeper data volume
##
storageClass: ""
## @param persistence.accessModes PVC Access modes
##
accessModes:
- ReadWriteOnce
## @param persistence.size PVC Storage Request for ZooKeeper data volume
##
size: 8Gi
## @param persistence.annotations Annotations for the PVC
##
annotations: {}
## @param persistence.selector Selector to match an existing Persistent Volume for Zookeeper's data PVC
## If set, the PVC can't have a PV dynamically provisioned for it
## E.g.
## selector:
## matchLabels:
## app: my-app
##
selector: {}
dataLogDir:
## @param persistence.dataLogDir.size PVC Storage Request for ZooKeeper's Data log directory
##
size: 8Gi
## @param persistence.dataLogDir.existingClaim Provide an existing `PersistentVolumeClaim` for Zookeeper's Data log directory
## If defined, PVC must be created manually before volume will be bound
## The value is evaluated as a template
##
existingClaim: ""
## @param persistence.dataLogDir.selector Selector to match an existing Persistent Volume for Zookeeper's Data log PVC
## If set, the PVC can't have a PV dynamically provisioned for it
## E.g.
## selector:
## matchLabels:
## app: my-app
##
selector: {}
## @section Volume Permissions parameters
##
## Init containers parameters:
## volumePermissions: Change the owner and group of the persistent volume mountpoint to runAsUser:fsGroup values from the securityContext section.
##
volumePermissions:
## @param volumePermissions.enabled Enable init container that changes the owner and group of the persistent volume(s) mountpoint to `runAsUser:fsGroup`
##
enabled: false
## @param volumePermissions.image.registry Init container volume-permissions image registry
## @param volumePermissions.image.repository Init container volume-permissions image repository
## @param volumePermissions.image.tag Init container volume-permissions image tag (immutable tags are recommended)
## @param volumePermissions.image.pullPolicy Init container volume-permissions image pull policy
## @param volumePermissions.image.pullSecrets Init container volume-permissions image pull secrets
##
image:
registry: docker.io
repository: bitnami/bitnami-shell
tag: 10-debian-10-r230
pullPolicy: IfNotPresent
## Optionally specify an array of imagePullSecrets.
## Secrets must be manually created in the namespace.
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry/
## Example:
## pullSecrets:
## - myRegistryKeySecretName
##
pullSecrets: []
## @param volumePermissions.resources Init container resource requests/limit
##
resources: {}
## @section Metrics parameters
##
## Zookeeper Prometheus Exporter configuration
##
metrics:
## @param metrics.enabled Enable prometheus to access zookeeper metrics endpoint
##
enabled: false
## @param metrics.containerPort Zookeeper Prometheus Exporter container port
##
containerPort: 9141
## Service configuration
##
service:
## @param metrics.service.type Zookeeper Prometheus Exporter service type
##
type: ClusterIP
## @param metrics.service.port Prometheus metrics service port
##
port: 9141
## @param metrics.service.annotations [object] Annotations for the Zookeeper to auto-discover the metrics endpoint
##
annotations:
prometheus.io/scrape: "true"
prometheus.io/port: "{{ .Values.metrics.service.port }}"
prometheus.io/path: "/metrics"
## Prometheus Operator ServiceMonitor configuration
##
serviceMonitor:
## @param metrics.serviceMonitor.enabled if `true`, creates a Prometheus Operator ServiceMonitor (also requires `metrics.enabled` to be `true`)
##
enabled: false
## @param metrics.serviceMonitor.namespace Namespace for the ServiceMonitor Resource (defaults to the Release Namespace)
##
namespace: ""
## @param metrics.serviceMonitor.interval Interval at which metrics should be scraped.
## ref: https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#endpoint
## e.g:
## interval: 10s
##
interval: ""
## @param metrics.serviceMonitor.scrapeTimeout Timeout after which the scrape is ended
## ref: https://github.com/coreos/prometheus-operator/blob/master/Documentation/api.md#endpoint
## e.g:
## scrapeTimeout: 10s
##
scrapeTimeout: ""
## @param metrics.serviceMonitor.selector Prometheus instance selector labels
## ref: https://github.com/bitnami/charts/tree/master/bitnami/prometheus-operator#prometheus-configuration
## e.g:
## selector:
## prometheus: my-prometheus
##
selector: {}
## Prometheus Operator PrometheusRule configuration
##
prometheusRule:
## @param metrics.prometheusRule.enabled if `true`, creates a Prometheus Operator PrometheusRule (also requires `metrics.enabled` to be `true` and `metrics.prometheusRule.rules`)
##
enabled: false
## @param metrics.prometheusRule.namespace Namespace for the PrometheusRule Resource (defaults to the Release Namespace)
##
namespace: ""
## @param metrics.prometheusRule.selector Prometheus instance selector labels
## ref: https://github.com/bitnami/charts/tree/master/bitnami/prometheus-operator#prometheus-configuration
## e.g:
## selector:
## prometheus: my-prometheus
##
selector: {}
## @param metrics.prometheusRule.rules Prometheus Rule definitions
## - alert: ZookeeperSyncedFollowers
## annotations:
## message: The number of synced followers for the leader node in Zookeeper deployment my-release is less than 2. This usually means that some of the Zookeeper nodes aren't communicating properly. If it doesn't resolve itself you can try killing the pods (one by one).
## expr: max(synced_followers{service="my-release-metrics"}) < 2
## for: 5m
## labels:
## severity: critical
## - alert: ZookeeperOutstandingRequests
## annotations:
## message: The number of outstanding requests for Zookeeper pod {{ $labels.pod }} is greater than 10. This can indicate a performance issue with the Pod or cluster a whole.
## expr: outstanding_requests{service="my-release-metrics"} > 10
## for: 5m
## labels:
## severity: critical
##
rules: []
## @section TLS/SSL parameters
##
## Enable SSL/TLS encryption
##
tls:
client:
## @param tls.client.enabled Enable TLS for client connections
##
enabled: false
## @param tls.client.autoGenerated Generate automatically self-signed TLS certificates for Zookeeper client communications
## Currently only supports PEM certificates
##
autoGenerated: false
## @param tls.client.existingSecret Name of the existing secret containing the TLS certificates for Zookeper client communications
##
existingSecret: ""
## @param tls.client.keystorePath Location of the KeyStore file used for Client connections
##
keystorePath: /opt/bitnami/zookeeper/config/certs/client/zookeeper.keystore.jks
## @param tls.client.truststorePath Location of the TrustStore file used for Client connections
##
truststorePath: /opt/bitnami/zookeeper/config/certs/client/zookeeper.truststore.jks
## @param tls.client.passwordsSecretName Existing secret containing Keystore and truststore passwords
##
passwordsSecretName: ""
## @param tls.client.keystorePassword Password to access KeyStore if needed
##
keystorePassword: ""
## @param tls.client.truststorePassword Password to access TrustStore if needed
##
truststorePassword: ""
quorum:
## @param tls.quorum.enabled Enable TLS for quorum protocol
##
enabled: false
## @param tls.quorum.autoGenerated Create self-signed TLS certificates. Currently only supports PEM certificates.
##
autoGenerated: false
## @param tls.quorum.existingSecret Name of the existing secret containing the TLS certificates for Zookeper quorum protocol
##
existingSecret: ""
## @param tls.quorum.keystorePath Location of the KeyStore file used for Quorum protocol
##
keystorePath: /opt/bitnami/zookeeper/config/certs/quorum/zookeeper.keystore.jks
## @param tls.quorum.truststorePath Location of the TrustStore file used for Quorum protocol
##
truststorePath: /opt/bitnami/zookeeper/config/certs/quorum/zookeeper.truststore.jks
## @param tls.quorum.passwordsSecretName Existing secret containing Keystore and truststore passwords
##
passwordsSecretName: ""
## @param tls.quorum.keystorePassword Password to access KeyStore if needed
##
keystorePassword: ""
## @param tls.quorum.truststorePassword Password to access TrustStore if needed
##
truststorePassword: ""
## We usually recommend not to specify default resources and to leave this as a conscious
## choice for the user. This also increases chances charts run on environments with little
## resources, such as Minikube. If you do want to specify resources, uncomment the following
## lines, adjust them as necessary, and remove the curly braces after 'resources:'.
## @param tls.resources.limits The resources limits for the TLS init container
## @param tls.resources.requests The requested resources for the TLS init container
##
resources:
## Example:
## limits:
## cpu: 100m
## memory: 128Mi
##
limits: {}
## Examples:
## requests:
## cpu: 100m
## memory: 128Mi
##
requests: {}