Files
charts/bitnami/influxdb
Bartlomiej Wulff 1e8e275b85 [bitnami/influxdb] Fix Backup (influxdb-backup-dummy-container) (#6404)
* Update cronjob-backup.yaml

Fix wrong container image declaration for influxdb-backup-dummy-container

* Update Chart.yaml

Bump the chart version to 2.2.3
2021-05-19 08:48:30 +02:00
..
2021-04-29 09:55:33 +02:00

InfluxDBTM

InfluxDBTM is an open source time-series database designed to handle large write and read loads in real-time.

Disclaimer: The respective trademarks mentioned in the offering are owned by the respective companies. We do not provide a commercial license for any of these products. This listing has an open-source license. InfluxDBTM and InfluxDB RelayTM are run and maintained by InfluxData, which is a completely separate project from Bitnami.

TL;DR

$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/influxdb

Introduction

This chart bootstraps a influxdb deployment on a Kubernetes cluster using the Helm package manager.

Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters.

Prerequisites

  • Kubernetes 1.12+
  • Helm 3.1.0
  • PV provisioner support in the underlying infrastructure
  • ReadWriteMany volumes for deployment scaling

Installing the Chart

To install the chart with the release name my-release:

$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/influxdb

These commands deploy influxdb on the Kubernetes cluster in the default configuration. The Parameters section lists the parameters that can be configured during installation.

Tip

: List all releases using helm list

Uninstalling the Chart

To uninstall/delete the my-release statefulset:

$ helm delete my-release

The command removes all the Kubernetes components associated with the chart and deletes the release. Use the option --purge to delete all history too.

Parameters

The following tables lists the configurable parameters of the InfluxDBTM chart and their default values.

Parameter Description Default
global.imageRegistry Global Docker image registry nil
global.imagePullSecrets Global Docker registry secret names as an array [] (does not add image pull secrets to deployed pods)
global.storageClass Global storage class for dynamic provisioning nil

Common parameters

Parameter Description Default
nameOverride String to partially override influxdb.fullname template with a string (will prepend the release name) nil
fullnameOverride String to fully override influxdb.fullname template with a string nil
commonLabels Labels to add to all deployed objects {}
commonAnnotations Annotations to add to all deployed objects {}
clusterDomain Default Kubernetes cluster domain cluster.local
extraDeploy Array of extra objects to deploy with the release [] (evaluated as a template)
kubeVersion Force target Kubernetes version (using Helm capabilities if not set) nil
networkPolicy.enabled Enable NetworkPolicy false
networkPolicy.allowExternal Don't require client label for connections true
persistence.enabled Enable data persistence true
persistence.existingClaim Use a existing PVC which must be created manually before bound nil
persistence.storageClass Specify the storageClass used to provision the volume nil
persistence.accessModes Access mode of data volume ["ReadWriteOnce"]
persistence.size Size of data volume 8Gi

InfluxDBTM parameters

Parameter Description Default
image.registry InfluxDBTM image registry docker.io
image.repository InfluxDBTM image name bitnami/influxdb
image.tag InfluxDBTM image tag {TAG_NAME}
image.pullPolicy InfluxDBTM image pull policy IfNotPresent
image.pullSecrets Specify docker-registry secret names as an array [] (does not add image pull secrets to deployed pods)
image.debug Specify if debug logs should be enabled false
architecture InfluxDBTM architecture (standalone or high-availability) standalone
auth.enabled Enable/disable authentication (Variable to keep compatibility with InfluxDBTM v1, in v2 it will be ignored) true
auth.admin.username InfluxDBTM admin user name admin
auth.admin.password InfluxDBTM admin user's password nil
auth.admin.token InfluxDBTM admin user's token. Only valid with InfluxDBTM v2 nil
auth.admin.org InfluxDBTM admin user's org. Only valid with InfluxDBTM v2 primary
auth.admin.bucket InfluxDBTM admin user's bucket. Only valid with InfluxDBTM v2 primary
auth.admin.usePasswordFile Mount admin user's password as file instead of environment variable false
auth.createUserToken Whether to create tokens for the different users. Take into account these tokens are going to be created by CLI randomly and they will not be accessible from a secret. See more influxdb 2.0 auth ref false
auth.user.username Name for InfluxDBTM user with 'admin' privileges on the bucket specified at auth.user.bucket and auth.user.org or auth.admin.org nil
auth.user.password InfluxDBTM password for user.name user nil
auth.user.usePasswordFile Mount user.name user's password as file instead of environment variable nil
auth.user.bucket Bucket to be created on first run my_database
auth.user.org Org to be created on first run my_database
auth.readUser.username Name for InfluxDBTM user with 'read' privileges on the bucket specified at auth.user.bucket nil
auth.readUser.password InfluxDBTM password for auth.readUser.username user nil
auth.readUser.usePasswordFile Mount auth.readUser.username user's password as file instead of environment variable nil
auth.writeUser.username Name for InfluxDBTM user with 'read' privileges on the bucket specified at auth.user.bucket nil
auth.writeUser.password InfluxDBTM password for auth.writeUser.username user nil
auth.writeUser.usePasswordFile Mount auth.writeUser.username user's password as file instead of environment variable nil
auth.existingSecret Name of existing Secret object with InfluxDBTM credentials (auth.admin.password, auth.user.password, auth.readUser.password, and auth.writeUser.password will be ignored and picked up from this secret) nil
influxdb.configuration Specify content for influxdb.conf nil (do not create influxdb.conf)
influxdb.existingConfiguration Name of existing ConfigMap object with the InfluxDBTM configuration (influxdb.configuration will be ignored). nil
influxdb.initdbScripts Dictionary of initdb scripts nil
influxdb.initdbScriptsCM Name of existing ConfigMap object with the initdb scripts (influxdb.initdbScripts will be ignored). nil
influxdb.initdbScriptsSecret Secret with initdb scripts that contain sensitive information (Note: can be used with initdbScriptsConfigMap or initdbScripts) nil
influxdb.extraEnvVars Array containing extra env vars to configure InfluxDBTM nil
influxdb.replicaCount The number of InfluxDBTM replicas to deploy 1
influxdb.podAffinityPreset InfluxDBTM Pod affinity preset. Ignored if affinity is set. Allowed values: soft or hard ""
influxdb.podAntiAffinityPreset InfluxDBTM Pod anti-affinity preset. Ignored if affinity is set. Allowed values: soft or hard ""
influxdb.nodeAffinityPreset.type InfluxDBTM Node affinity preset type. Ignored if affinity is set. Allowed values: soft or hard ""
influxdb.nodeAffinityPreset.key InfluxDBTM Node label key to match Ignored if affinity is set. ""
influxdb.nodeAffinityPreset.values InfluxDBTM Node label values to match. Ignored if affinity is set. []
influxdb.affinity InfluxDBTM Affinity for pod assignment {} (evaluated as a template)
influxdb.nodeSelector InfluxDBTM Node labels for pod assignment {} (evaluated as a template)
influxdb.tolerations InfluxDBTM Tolerations for pod assignment [] (evaluated as a template)
influxdb.extraVolumeMounts Array of extra volume mounts to be added to the container (evaluated as template). Normally used with extraVolumes. [] (evaluated as a template)
influxdb.extraVolumes Array of extra volumes to be added to the deployment (evaluated as template). Requires setting extraVolumeMounts [] (evaluated as a template)
influxdb.podManagementPolicy Pod Management Policy [OrderedReady or Parallel] OrderedReady
influxdb.securityContext.enabled Enable security context for InfluxDBTM true
influxdb.securityContext.fsGroup Group ID for the InfluxDBTM filesystem 1001
influxdb.securityContext.runAsUser User ID for the InfluxDBTM container 1001
influxdb.resources The [resources] to allocate for container {}
influxdb.livenessProbe Liveness probe configuration for InfluxDBTM Check values.yaml file
influxdb.readinessProbe Readiness probe configuration for InfluxDBTM Check values.yaml file
influxdb.customLivenessProbe Override default liveness probe nil
influxdb.customReadinessProbe Override default readiness probe nil
influxdb.containerPorts.http InfluxDBTM container HTTP port 8086
influxdb.containerPorts.rpc InfluxDBTM container RPC port 8088
influxdb.service.type Kubernetes service type (ClusterIP, NodePort or LoadBalancer) ClusterIP
influxdb.service.port InfluxDBTM HTTP port 8086
influxdb.service.rpcPort InfluxDBTM RPC port 8088
influxdb.service.nodePorts.http Kubernetes HTTP node port ""
influxdb.service.nodePorts.rpc Kubernetes RPC node port ""
influxdb.service.annotations Annotations for InfluxDBTM service {}
influxdb.service.loadBalancerIP loadBalancerIP if service type is LoadBalancer nil
influxdb.service.loadBalancerSourceRanges Address that are allowed when service is LoadBalancer []
influxdb.service.clusterIP Static clusterIP or None for headless services nil
influxdb.service.sessionAffinity Session affinity for the InfluxDBTM service nil
influxdb.service.sessionAffinityConfig Additional settings for the sessionAffinity {}

InfluxDB RelayTM parameters

Parameter Description Default
relay.image.registry InfluxDB RelayTM image registry docker.io
relay.image.repository InfluxDB RelayTM image name bitnami/influxdb-relay
relay.image.tag InfluxDB RelayTM image tag {TAG_NAME}
relay.image.pullPolicy InfluxDB RelayTM image pull policy IfNotPresent
relay.image.pullSecrets Specify docker-registry secret names as an array [] (does not add image pull secrets to deployed pods)
relay.configuration Specify content for relay.toml Check values.yaml file
relay.existingConfiguration Name of existing ConfigMap object with the InfluxDB RelayTM configuration (relay.configuration will be ignored) nil
relay.replicaCount The number of InfluxDB RelayTM replicas to deploy 1
relay.podAffinityPreset InfluxDB RelayTM Pod affinity preset. Ignored if affinity is set. Allowed values: soft or hard ""
relay.podAntiAffinityPreset InfluxDB RelayTM Pod anti-affinity preset. Ignored if affinity is set. Allowed values: soft or hard ""
relay.nodeAffinityPreset.type InfluxDB RelayTM Node affinity preset type. Ignored if affinity is set. Allowed values: soft or hard ""
relay.nodeAffinityPreset.key InfluxDB RelayTM Node label key to match Ignored if affinity is set. ""
relay.nodeAffinityPreset.values InfluxDB RelayTM Node label values to match. Ignored if affinity is set. []
relay.affinity InfluxDB RelayTM Affinity for pod assignment {} (evaluated as a template)
relay.nodeSelector InfluxDB RelayTM Node labels for pod assignment {} (evaluated as a template)
relay.tolerations InfluxDB RelayTM Tolerations for pod assignment [] (evaluated as a template)
relay.extraVolumeMounts Array of extra volume mounts to be added to the container (evaluated as template). Normally used with extraVolumes. [] (evaluated as a template)
relay.extraVolumes Array of extra volumes to be added to the deployment (evaluated as template). Requires setting extraVolumeMounts [] (evaluated as a template)
relay.securityContext.enabled Enable security context for InfluxDB RelayTM true
relay.securityContext.fsGroup Group ID for the InfluxDB RelayTM filesystem 1001
relay.securityContext.runAsUser User ID for the InfluxDB RelayTM container 1001
relay.resources The [resources] to allocate for container {}
relay.livenessProbe Liveness probe configuration for InfluxDB RelayTM Check values.yaml file
relay.readinessProbe Readiness probe configuration for InfluxDB RelayTM Check values.yaml file
relay.customLivenessProbe Override default liveness probe nil
relay.customReadinessProbe Override default readiness probe nil
relay.containerPorts.http InfluxDB RelayTM container HTTP port 9096
relay.service.type Kubernetes service type (ClusterIP, NodePort or LoadBalancer) ClusterIP
relay.service.port InfluxDB RelayTM HTTP port 9096
relay.service.nodePort Kubernetes HTTP node port ""
relay.service.annotations Annotations for InfluxDB RelayTM service {}
relay.service.loadBalancerIP loadBalancerIP if service type is LoadBalancer nil
relay.service.loadBalancerSourceRanges Address that are allowed when service is LoadBalancer []
relay.service.clusterIP Static clusterIP or None for headless services nil

InfluxDB CollectdTM parameters

Parameter Description Default
collectd.enabled InfluxDB CollectdTM service enable false
collectd.service.type Kubernetes service type (ClusterIP, NodePort or LoadBalancer) ClusterIP
collectd.service.port InfluxDB CollectdTM UDP port (should match with corresponding port in influxdb.conf) 25826
collectd.service.nodePort Kubernetes HTTP node port ""
collectd.service.annotations Annotations for InfluxDB CollectdTM service {}
collectd.service.loadBalancerIP loadBalancerIP if service type is LoadBalancer nil
collectd.service.loadBalancerSourceRanges Address that are allowed when service is LoadBalancer []
collectd.service.clusterIP Static clusterIP or None for headless services nil

Exposing parameters

Parameter Description Default
ingress.enabled Enable ingress controller resource false
ingress.certManager Add annotations for cert-manager false
ingress.hostname Default host for the ingress resource phabricator.local
ingress.apiVersion Force Ingress API version (automatically detected if not set) nil
ingress.path Ingress path /
ingress.pathType Ingress path type ImplementationSpecific
ingress.tls Create TLS Secret false
ingress.annotations Ingress annotations [] (evaluated as a template)
ingress.extraHosts[0].name Additional hostnames to be covered nil
ingress.extraHosts[0].path Additional hostnames to be covered nil
ingress.extraPaths Additional arbitrary path/backend objects nil
ingress.extraTls[0].hosts[0] TLS configuration for additional hostnames to be covered nil
ingress.extraTls[0].secretName TLS configuration for additional hostnames to be covered nil
ingress.secrets[0].name TLS Secret Name nil
ingress.secrets[0].certificate TLS Secret Certificate nil
ingress.secrets[0].key TLS Secret Key nil

Metrics parameters

Parameter Description Default
metrics.enabled Enable the export of Prometheus metrics false
metrics.service.type Kubernetes service type (ClusterIP, NodePort or LoadBalancer) ClusterIP
metrics.service.port InfluxDBTM Prometheus port 9122
metrics.service.nodePort Kubernetes HTTP node port ""
metrics.service.annotations Annotations for Prometheus metrics service Check values.yaml file
metrics.service.loadBalancerIP loadBalancerIP if service type is LoadBalancer nil
metrics.service.loadBalancerSourceRanges Address that are allowed when service is LoadBalancer []
metrics.service.clusterIP Static clusterIP or None for headless services nil
metrics.serviceMonitor.enabled if true, creates a Prometheus Operator ServiceMonitor (also requires metrics.enabled to be true) false
metrics.serviceMonitor.namespace Namespace in which Prometheus is running nil
metrics.serviceMonitor.interval Interval at which metrics should be scraped. nil (Prometheus Operator default value)
metrics.serviceMonitor.scrapeTimeout Timeout after which the scrape is ended nil (Prometheus Operator default value)
metrics.serviceMonitor.selector Prometheus instance selector labels nil

Volume permissions parameters

Parameter Description Default
volumePermissions.enabled Enable init container that changes the owner and group of the persistent volume mountpoint to runAsUser:fsGroup false
volumePermissions.image.registry Init container volume-permissions image registry docker.io
volumePermissions.image.repository Init container volume-permissions image name bitnami/bitnami-shell
volumePermissions.image.tag Init container volume-permissions image tag "10"
volumePermissions.image.pullPolicy Init container volume-permissions image pull policy Always
volumePermissions.image.pullSecrets Specify docker-registry secret names as an array [] (does not add image pull secrets to deployed pods)
volumePermissions.securityContext.runAsUser User ID for the init container (when facing issues in OpenShift or uid unknown, try value "auto") 0

InfluxDBTM backup parameters

Parameter Description Default
backup.enabled enable InfluxDBTM backup false
backup.directory directory where backups are stored in "/backups"
backup.retentionDays retention time in days for backups (older backups are deleted) 10
backup.cronjob.schedule crontab style time schedule for backup execution "0 2 * * *"
backup.cronjob.historyLimit cronjob historylimit 1
backup.cronjob.annotations backup pod annotations {}
backup.uploadProviders.google.enabled enable upload to google storage bucket false
backup.uploadProviders.google.secret json secret with serviceaccount data to access Google storage bucket ""
backup.uploadProviders.google.secretKey service account secret key name "key.json"
backup.uploadProviders.google.existingSecret Name of existing secret object with Google serviceaccount json credentials ""
backup.uploadProviders.google.bucketName google storage bucket name name "gs://bucket/influxdb"
backup.uploadProviders.google.image.registry Google Cloud SDK image registry docker.io
backup.uploadProviders.google.image.repository Google Cloud SDK image name bitnami/google-cloud-sdk
backup.uploadProviders.google.image.tag Google Cloud SDK image tag {TAG_NAME}
backup.uploadProviders.google.image.pullPolicy Google Cloud SDK image pull policy IfNotPresent
backup.uploadProviders.google.image.pullSecrets Specify docker-registry secret names as an array [] (does not add image pull secrets to deployed pods)
backup.uploadProviders.azure.enabled enable upload to azure storage container false
backup.uploadProviders.azure.secret secret with credentials to access Azure storage ""
backup.uploadProviders.azure.secretKey service account secret key name "connection-string"
backup.uploadProviders.azure.existingSecret Name of existing secret object ""
backup.uploadProviders.azure.containerName destination container "influxdb-container"
backup.uploadProviders.azure.image.registry Azure CLI image registry docker.io
backup.uploadProviders.azure.image.repository Azure CLI image name bitnami/azure-cli
backup.uploadProviders.azure.image.tag Azure CLI image tag {TAG_NAME}
backup.uploadProviders.azure.image.pullPolicy Azure CLI image pull policy IfNotPresent
backup.uploadProviders.azure.image.pullSecrets Specify docker-registry secret names as an array [] (does not add image pull secrets to deployed pods)

Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,

$ helm install my-release \
  --set auth.admin.username=admin-user bitnami/influxdb

The above command sets the InfluxDBTM admin user to admin-user.

NOTE: Once this chart is deployed, it is not possible to change the application's access credentials, such as usernames or passwords, using Helm. To change these application credentials after deployment, delete any persistent volumes (PVs) used by the chart and re-deploy it, or use the application's built-in administrative tools if available.

Alternatively, a YAML file that specifies the values for the parameters can be provided while installing the chart. For example,

$ helm install my-release -f values.yaml bitnami/influxdb

Tip

: You can use the default values.yaml

Configuration and installation details

Rolling VS Immutable tags

It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.

Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.

Standalone vs High Availability architecture

You can install the InfluxDBTM chart with two different architecture setups: "standalone" or "high-availability", you can use the architecture parameter:

architecture="standalone"
architecture="high-availability"

The standalone architecture installs a deployment with one InfluxDBTM server (it cannot be scaled):

               ┌──────────────────┐
               │     Ingress      │
               │    Controller    │
               └────────┬─────────┘
                        |
                        │ /query
                        │ /write
                        ▼
                ┌────────────────┐
                │  InfluxDB(TM)  │
                |      svc       │
                └───────┬────────┘
                        │
                        ▼
                 ┌──────────────┐
                 │ InfluxDB(TM) │
                 │    Server    │
                 │     Pod      │
                 └──────────────┘

The high availability install a statefulset with N InfluxDBTM servers and M InfluxDB RelayTM instances:

                   ┌──────────────────┐
                   │     Ingress      │
                   │    Controller    │
                   └───────┬─┬────────┘
                           │ │
                           │ │
              ┌────────────┘ └─────────────┐
              │                            │
              │ /write              /query │
              ▼                            ▼
      ┌────────────────────┐  ┌────────────────────┐
      │ InfluxDB Relay(TM) │  │    InfluxDB(TM)    │
      │          svc       │  │         svc        │
      └───────────┬─┬──────┘  └─────┬─────┬────────┘
      ┌────────── │─|───────────────|─────│───────┐
      |           │ |               |     │       ▼
┌─────┴────────┐  │ |               |     │  ┌──────────────┐
│   InfluxDB   │  │ |               |     │  │ InfluxDB(TM) │
│  Relay(TM)   │◀─┘ |               |     └─▶│    Server    │
│     Pod      │    │               │        │     Pod      │
└─────┬────────┘    │               │        └──────────────┘
      |             │               │           ▲
      └─────────────│───────────────│───────┐   |
                    │               │       |   |
  ┌──────────────── │───────────────│───────────┘
  |                 │               │       |
  |                 │               │       ▼
┌─┴─────────────┐   │               │   ┌──────────────┐
│    InfluxDB   │   │               │   │ InfluxDB(TM) │
│   Relay(TM)   │◀──┘               └──▶│  Server      │
│      Pod      │                       │   Pod        │
└─────┬─────────┘                       └──────────────┘
      |                                   ▲
      └───────────────────────────────────┘

When using the high-availability architecture, it is recommended to configure sticky sessions using --set influxdb.service.sessionAffinity="ClientIP" or configuring the IngressController accordingly.

Configure the way how to expose InfluxDBTM

  • Ingress: The ingress controller must be installed in the Kubernetes cluster. Set ingress.enabled=true to expose InfluxDBTM through Ingress.
  • ClusterIP: Exposes the service on a cluster-internal IP. Choosing this value makes the service only reachable from within the cluster. Set influxdb.service.type=ClusterIP to choose this service type.
  • NodePort: Exposes the service on each Node's IP at a static port (the NodePort). Youll be able to contact the NodePort service, from outside the cluster, by requesting NodeIP:NodePort. Set influxdb.service.type=NodePort to choose this service type.
  • LoadBalancer: Exposes the service externally using a cloud provider's load balancer. Set influxdb.service.type=LoadBalancer to choose this service type.

Using custom configuration

This helm chart supports to customize the whole configuration file.

Add your custom configuration file to "files/conf" in your working directory. This file will be mounted as a configMap to the containers and it will be used for configuring InfluxDBTM.

Alternatively, you can specify the InfluxDBTM configuration using the influxdb.configuration parameter.

In addition to these options, you can also set an external ConfigMap with all the configuration files. This is done by setting the influxdb.existingConfiguration parameter. Note that this will override the two previous options.

Adding extra environment variables

In case you want to add extra environment variables, you can use the influxdb.extraEnvVars property.

extraEnvVars:
  - name: INFLUXDB_DATA_QUERY_LOG_ENABLED
    value: "true"

Initialize a fresh instance

The Bitnami InfluxDBTM image allows you to use your custom scripts to initialize a fresh instance. In order to execute the scripts, they must be located inside the chart folder files/docker-entrypoint-initdb.d so they can be consumed as a ConfigMap.

Alternatively, you can specify custom scripts using the influxdb.initdbScripts parameter.

In addition to these options, you can also set an external ConfigMap with all the initialization scripts. This is done by setting the influxdb.initdbScriptsCM parameter. Note that this will override the two previous options. parameter.

The allowed extensions are .sh, and .txt.

Setting Pod's affinity

This chart allows you to set your custom affinity using the XXX.affinity parameter(s). Find more information about Pod's affinity in the kubernetes documentation.

As an alternative, you can use of the preset configurations for pod affinity, pod anti-affinity, and node affinity available at the bitnami/common chart. To do so, set the XXX.podAffinityPreset, XXX.podAntiAffinityPreset, or XXX.nodeAffinityPreset parameters.

Persistence

The data is persisted by default using PVC(s). You can disable the persistence setting the persistence.enabled parameter to false. A default StorageClass is needed in the Kubernetes cluster to dynamically provision the volumes. Specify another StorageClass in the persistence.storageClass or set persistence.existingClaim if you have already existing persistent volumes to use.

Adjust permissions of persistent volume mountpoint

As the images run as non-root by default, it is necessary to adjust the ownership of the persistent volumes so that the containers can write data into it.

By default, the chart is configured to use Kubernetes Security Context to automatically change the ownership of the volume. However, this feature does not work in all Kubernetes distributions. As an alternative, this chart supports using an initContainer to change the ownership of the volume before mounting it in the final destination.

You can enable this initContainer by setting volumePermissions.enabled to true. There are K8s distribution, such as OpenShift, where you can dynamically define the UID to run this initContainer. To do so, set the volumePermissions.securityContext.runAsUser to auto.

Troubleshooting

Find more information about how to deal with common errors related to Bitnamis Helm charts in this troubleshooting guide.

Upgrade

It's necessary to specify the existing passwords while performing an upgrade to ensure the secrets are not updated with invalid randomly generated passwords. Remember to specify the existing values of the auth.admin.password, user.pwd, auth.readUser.password and auth.writeUser.password parameters when upgrading the chart:

$ helm upgrade my-release bitnami/influxdb \
    --set auth.admin.password=[ADMIN_USER_PASSWORD] \
    --set auth.user.password=[USER_PASSWORD] \
    --set auth.readUser.password=[READ_USER_PASSWORD] \
    --set auth.writeUser.password=[WRITE_USER_PASSWORD]

Note: you need to substitute the placeholders [ADMIN_USER_PASSWORD], [USER_PASSWORD], [READ_USER_PASSWORD], and [WRITE_USER_PASSWORD] with the values obtained from instructions in the installation notes.

Upgrading

To 2.0.0

This version adds support to InfluxDBTM +2.0, since this version the chart is only verified to work with InfluxDBTM +2.0 bitnami images. However, you can use images for versions ~1.x.x taking into account the chart may need some modification to run with them.

Installing InfluxDBTM v1 in chart v2.

$ helm install bitnami/influxdb --set image.tag=1.8.3-debian-10-r88

As a consecuece some breaking changes have been included in this version.

  • Authentication values has been moved to auth.<property>....
  • We keep auth.readUser and auth.writeUser user options in order to be compatible with InfluxDBTM v1. If you are using InfluxDBTM 2.0, please, use the CLI to create user and tokens through initDb scripts at influxdb.initdbScripts or the UI due to we are not able to automacally provide a token for them to be used. See more InfluxDBTM 2.0 auth
  • InfluxDBTM 2.0 has removed database concept, now it is called Bucket so the property database has been also moved to auth.user.bucket.
  • Removed support for files/docker-entrypoint-initdb.d/*.{sh,txt}, please use .influxdb.initdbScripts or .Values.influxdb.initdbScriptsCM instead.
  • Removed support for files/conf/influxdb.conf, please use .influxdb.configuration or .influxdb.existingConfiguration instead.
  • Removed support for files/conf/relay.toml, please use .Values.relay.configuration or .Values.relay.existingConfiguration instead.
  • ingress.hosts parameter has been replaced by ingress.hostname and ingress.extraHosts that may give us a bit more flexibility.

Migrating form InfluxDBTM v1 to v2

Since this release could mean lot of concepts changes, we strongly recommend to not do it direcly using the chart upgrade. Please, read more info in their upgrade guide.

We actually recommend to backup all the data form a previous helm release, install new release using latest version of the chart and images and then restore data following their guides.

Upgrading the chart form 1.x.x to 2.x.x using InfluxDBTM v1 images.

NOTE: Please, create a backup of your database before running any of those actions.

Having an already existing chart release called influxdb and deployed like

$ helm install influxdb bitnami/influxdb
Export secrets and required values to update
$ export INFLUXDB_ADMIN_PASSWORD=$(kubectl get secret --namespace default influxdb -o jsonpath="{.data.admin-user-password}" | base64 --decode)
Upgrade the chart release

NOTE: Please remember to migrate all the values to its new path following the above notes, e.g: adminUser.pwd -> auth.admin.password.

$ helm upgrade influxdb bitnami/influxdb --set image.tag=1.8.3-debian-10-r99 \
  --set auth.admin.password=${INFLUXDB_ADMIN_PASSWORD}

To 1.1.0

This version introduces bitnami/common, a library chart as a dependency. More documentation about this new utility could be found here. Please, make sure that you have updated the chart dependencies before executing any upgrade.

To 1.0.0

On November 13, 2020, Helm v2 support was formally finished, this major version is the result of the required changes applied to the Helm Chart to be able to incorporate the different features added in Helm v3 and to be consistent with the Helm project itself regarding the Helm v2 EOL.

What changes were introduced in this major version?

  • Previous versions of this Helm Chart use apiVersion: v1 (installable by both Helm 2 and 3), this Helm Chart was updated to apiVersion: v2 (installable by Helm 3 only). Here you can find more information about the apiVersion field.
  • The different fields present in the Chart.yaml file has been ordered alphabetically in a homogeneous way for all the Bitnami Helm Charts

Considerations when upgrading to this version

  • If you want to upgrade to this version from a previous one installed with Helm v3, you shouldn't face any issues
  • If you want to upgrade to this version using Helm v2, this scenario is not supported as this version doesn't support Helm v2 anymore
  • If you installed the previous version with Helm v2 and wants to upgrade to this version with Helm v3, please refer to the official Helm documentation about migrating from Helm v2 to v3

Useful links