# Fluentd
[Fluentd](https://www.fluentd.org/) is an open source data collector, which lets you unify the data collection and consumption for a better use and understanding of data.
## TL;DR
```console
$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/fluentd
```
## Introduction
This chart bootstraps a [Fluentd](https://github.com/bitnami/bitnami-docker-fluentd) deployment on a [Kubernetes](http://kubernetes.io) cluster using the [Helm](https://helm.sh) package manager.
Bitnami charts can be used with [Kubeapps](https://kubeapps.com/) for deployment and management of Helm Charts in clusters.
## Prerequisites
- Kubernetes 1.12+
- Helm 3.1.0
- PV provisioner support in the underlying infrastructure
> Note: Please, note that the forwarder runs the container as root by default setting the `forwarder.securityContext.runAsUser` to `0` (_root_ user)
## Installing the Chart
To install the chart with the release name `my-release`:
```console
$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/fluentd
```
These commands deploy Fluentd on the Kubernetes cluster in the default configuration. The [Parameters](#parameters) section lists the parameters that can be configured during installation.
> **Tip**: List all releases using `helm list`
## Uninstalling the Chart
To uninstall/delete the `my-release` resources:
```console
$ helm delete my-release
```
The command removes all the Kubernetes components associated with the chart and deletes the release. Use the option `--purge` to delete all history too.
## Parameters
The following tables lists the configurable parameters of the fluentd chart and their default values.
| Parameter | Description | Default |
|-------------------------------------------------|----------------------------------------------------------------------------------------------------------------|---------------------------------------------------------------------------------------------------------|
| `global.imageRegistry` | Global Docker image registry | `nil` |
| `global.imagePullSecrets` | Global Docker registry secret names as an array | `[]` (does not add image pull secrets to deployed pods) |
| `global.storageClass` | Global storage class for dynamic provisioning | `nil` |
| `image.registry` | Fluentd image registry | `docker.io` |
| `image.repository` | Fluentd image name | `bitnami/fluentd` |
| `image.tag` | Fluentd image tag | `{TAG_NAME}` |
| `image.pullPolicy` | Fluentd image pull policy | `IfNotPresent` |
| `image.pullSecrets` | Specify docker-registry secret names as an array | `[]` (does not add image pull secrets to deployed pods) |
| `kubeVersion` | Force target Kubernetes version (using Helm capabilities if not set) | `nil` |
| `nameOverride` | String to partially override fluentd.fullname template with a string (will prepend the release name) | `nil` |
| `fullnameOverride` | String to fully override fluentd.fullname template with a string | `nil` |
| `clusterDomain` | Kubernetes DNS domain name to use | `cluster.local` |
| `extraDeploy` | Array of extra objects to deploy with the release | `[]` (evaluated as a template) |
| `forwarder.enabled` | Enable Fluentd forwarder | `true` |
| `forwarder.daemonUser` | Fluentd forwarder daemon system user | `root` |
| `forwarder.daemonGroup` | Fluentd forwarder daemon system group | `root` |
| `forwarder.securityContext.enabled` | Enable security context for forwarder pods | `true` |
| `forwarder.securityContext.fsGroup` | Group ID for forwarder's containers filesystem | `0` |
| `forwarder.securityContext.runAsUser` | User ID for forwarder's containers | `0` |
| `forwarder.securityContext.runAsGroup` | Group ID for forwarder's containers | `0` |
| `forwarder.securityContext.*` | Other pod security context to be included as-is in pod spec | `{}` |
| `forwarder.containerSecurityContext.enabled` | Enable security context for the forwarder container | `true` |
| `forwarder.containerSecurityContext.*` | Other container security context to be included as-is in pod spec | `{ "privileged" : false, "allowPrivilegeEscalation": false, "capabilities": { "drop": ["ALL"] } }` |
| `forwarder.terminationGracePeriodSeconds` | Duration in seconds the pod needs to terminate gracefully | 30 |
| `forwarder.configFile` | Name of the config file that will be used by Fluentd at launch under the `/opt/bitnami/fluentd/conf` directory | `fluentd.conf` |
| `forwarder.configMap` | Name of the config map that contains the Fluentd configuration files | `nil` |
| `forwarder.configMapFiles` | Files to be added to be config map. Ignored if `forwarder.configMap` is set | `Check values.yaml` |
| `forwarder.hostAliases` | Add deployment host aliases | `[]` |
| `forwarder.extraArgs` | Extra arguments for the Fluentd command line | `nil` |
| `forwarder.priorityClassName` | Set Pods Priority Class | `nil` |
| `forwarder.extraEnv` | Extra environment variables to pass to the container | `[]` |
| `forwarder.containerPorts` | Ports the forwarder containers will listen on | `Check values.yaml` |
| `forwarder.service.type` | Kubernetes service type (`ClusterIP`, `NodePort`, or `LoadBalancer`) for the forwarders | `ClusterIP` |
| `forwarder.service.ports` | Array containing the forwarder service ports | `Check values.yaml file` |
| `forwarder.service.loadBalancerIP` | loadBalancerIP if service type is `LoadBalancer` | `nil` |
| `forwarder.service.loadBalancerSourceRanges` | Addresses that are allowed when service is LoadBalancer | `[]` |
| `forwarder.service.clusterIP` | Static clusterIP or None for headless services | `nil` |
| `forwarder.service.annotations` | Annotations for the forwarder service | `{}` |
| `forwarder.persistence.enabled` | Enable persistence volume for the forwarder | `false` |
| `forwarder.persistence.hostPath.path` | Directory from the host node's filesystem to mount as hostPath volume for persistence. | `false` |
| `forwarder.livenessProbe.enabled` | Enable liveness probes for the forwarder | `true` |
| `forwarder.livenessProbe.httpGet.path` | Path to access the liveness probes for the forwarder | `/fluentd.healthcheck?json=%7B%22ping%22%3A+%22pong%22%7D` |
| `forwarder.livenessProbe.httpGet.port` | Name or number of the port to access on the liveness probes for the forwarder | `http` |
| `forwarder.livenessProbe.initialDelaySeconds` | Delay before liveness probe is initiated | `60` |
| `forwarder.livenessProbe.periodSeconds` | How often to perform the probe | `10` |
| `forwarder.livenessProbe.timeoutSeconds` | When the probe times out | `5` |
| `forwarder.livenessProbe.failureThreshold` | Minimum consecutive failures for the probe to be considered failed after having succeeded. | `6` |
| `forwarder.livenessProbe.successThreshold` | Minimum consecutive successes for the probe to be considered successful after having failed. | `1` |
| `forwarder.readinessProbe.enabled` | Enable readiness probes for the forwarder | `true` |
| `forwarder.readinessProbe.initialDelaySeconds` | Delay before readiness probe is initiated | `5` |
| `forwarder.readinessProbe.periodSeconds` | How often to perform the probe | `10` |
| `forwarder.readinessProbe.timeoutSeconds` | When the probe times out | `5` |
| `forwarder.readinessProbe.failureThreshold` | Minimum consecutive failures for the probe to be considered failed after having succeeded. | `6` |
| `forwarder.readinessProbe.successThreshold` | Minimum consecutive successes for the probe to be considered successful after having failed. | `1` |
| `forwarder.updateStrategy` | Update strategy for the forwarder DaemonSet | `RollingUpdate` |
| `forwarder.resources` | Configure resource requests and limits | `nil` |
| `forwarder.podAffinityPreset` | Forwarder Pod affinity preset. Ignored if `affinity` is set. Allowed values: `soft` or `hard` | `""` |
| `forwarder.podAntiAffinityPreset` | Forwarder Pod anti-affinity preset. Ignored if `affinity` is set. Allowed values: `soft` or `hard` | `""` |
| `forwarder.nodeAffinityPreset.type` | Forwarder Node affinity preset type. Ignored if `affinity` is set. Allowed values: `soft` or `hard` | `""` |
| `forwarder.nodeAffinityPreset.key` | Forwarder Node label key to match Ignored if `affinity` is set. | `""` |
| `forwarder.nodeAffinityPreset.values` | Forwarder Node label values to match. Ignored if `affinity` is set. | `[]` |
| `forwarder.affinity` | Forwarder Affinity for pod assignment | `{}` (evaluated as a template) |
| `forwarder.nodeSelector` | Forwarder Node labels for pod assignment | `{}` (evaluated as a template) |
| `forwarder.tolerations` | Forwarder Tolerations for pod assignment | `[]` (evaluated as a template) |
| `forwarder.podAnnotations` | Pod annotations | `{}` |
| `forwarder.serviceAccount.create` | Specify whether a ServiceAccount should be created. | `true` |
| `forwarder.serviceAccount.name` | The name of the ServiceAccount to create | Generated using the `fluentd.fullname` template |
| `forwarder.serviceAccount.annotations` | Additional Service Account annotations (evaluated as a template) | `{}` |
| `forwarder.rbac.create` | Specify whether RBAC resources should be created and used, allowing the get, watch and list of pods/namespaces | `true` |
| `forwarder.rbac.pspEnabled` | Specify whether the bundled Pod Security Policy should be created and bound with RBAC | `false` |
| `forwarder.initContainers` | Additional init containers to add to the pods | `[]` |
| `forwarder.sidecars` | Add additional containers to the pods | `[]` |
| `forwarder.extraVolumes` | Extra volumes | `nil` |
| `forwarder.extraVolumeMounts` | Mount extra volume(s) | `nil` |
| `aggregator.enabled` | Enable Fluentd aggregator | `true` |
| `aggregator.replicaCount` | Number of aggregator pods to deploy in the Stateful Set | `2` |
| `aggregator.securityContext.enabled` | Enable security context for aggregator pods | `true` |
| `aggregator.securityContext.fsGroup` | Group ID for aggregator's containers filesystem | `1001` |
| `aggregator.securityContext.runAsUser` | User ID for aggregator's containers | `1001` |
| `aggregator.securityContext.runAsGroup` | Group ID for aggregator's containers | `1001` |
| `aggregator.securityContext.*` | Other pod security context to be included as-is in pod spec | `{}` |
| `aggregator.containerSecurityContext.enabled` | Enable security context for the aggregator container | `true` |
| `aggregator.containerSecurityContext.*` | Other container security context to be included as-is in pod spec | `{ "privileged" : false, "allowPrivilegeEscalation": false, "capabilities": { "drop": ["ALL"] } }` |
| `aggregator.terminationGracePeriodSeconds` | Duration in seconds the pod needs to terminate gracefully | 30 |
| `aggregator.configFile` | Name of the config file that will be used by Fluentd at launch under the `/opt/bitnami/fluentd/conf` directory | `fluentd.conf` |
| `aggregator.configMap` | Name of the config map that contains the Fluentd configuration files | `nil` |
| `aggregator.configMapFiles` | Files to be added to be config map. Ignored if `aggregator.configMap` is set | `Check values.yaml` |
| `aggregator.port` | Kubernetes Service port - Fluentd transport port for the aggregators | `24224` |
| `aggregator.extraArgs` | Extra arguments for the Fluentd command line | `nil` |
| `aggregator.extraEnv` | Extra environment variables to pass to the container | `[]` |
| `aggregator.containerPorts` | Ports the aggregator containers will listen on | `Check values.yaml` |
| `aggregator.hostAliases` | Add deployment host aliases | `[]` |
| `aggregator.service.type` | Kubernetes service type (`ClusterIP`, `NodePort`, or `LoadBalancer`) for the aggregators | `ClusterIP` |
| `aggregator.service.ports` | Array containing the aggregator service ports | `Check values.yaml file` |
| `aggregator.service.loadBalancerIP` | loadBalancerIP if service type is `LoadBalancer` | `nil` |
| `aggregator.service.loadBalancerSourceRanges` | Addresses that are allowed when service is LoadBalancer | `[]` |
| `aggregator.service.clusterIP` | Static clusterIP or None for headless services | `nil` |
| `aggregator.service.annotations` | Annotations for the aggregator service | `{}` |
| `aggregator.ingress.enabled` | Enable ingress controller resource | `false` |
| `aggregator.ingress.certManager` | Add annotations for cert-manager | `false` |
| `aggregator.ingress.hostname` | Default host for the ingress resource | `wordpress.local` |
| `aggregator.ingress.path` | Default path for the ingress resource | `/` |
| `aggregator.ingress.pathType` | How the path matching is interpreted | `ImplementationSpecific` |
| `aggregator.ingress.tls` | Create TLS Secret | `false` |
| `aggregator.ingress.annotations` | Ingress annotations | `[]` (evaluated as a template) |
| `aggregator.ingress.extraHosts[0].name` | Additional hostnames to be covered | `nil` |
| `aggregator.ingress.extraHosts[0].path` | Additional hostnames to be covered | `nil` |
| `aggregator.ingress.extraPaths` | Additional arbitrary path/backend objects | `nil` |
| `aggregator.ingress.extraTls[0].hosts[0]` | TLS configuration for additional hostnames to be covered | `nil` |
| `aggregator.ingress.extraTls[0].secretName` | TLS configuration for additional hostnames to be covered | `nil` |
| `aggregator.ingress.secrets[0].name` | TLS Secret Name | `nil` |
| `aggregator.ingress.secrets[0].certificate` | TLS Secret Certificate | `nil` |
| `aggregator.ingress.secrets[0].key` | TLS Secret Key | `nil` |
| `aggregator.persistence.enabled` | Enable persistence volume for the aggregator | `false` |
| `aggregator.persistence.storageClass` | Persistent Volume storage class | `nil` |
| `aggregator.persistence.accessMode` | Persistent Volume access mode | `ReadWriteOnce` |
| `aggregator.persistence.size` | Persistent Volume size | `10Gi` |
| `aggregator.livenessProbe.enabled` | Enable liveness probes for the aggregator | `true` |
| `aggregator.livenessProbe.httpGet.path` | Path to access the liveness probes for the aggregator | `/fluentd.healthcheck?json=%7B%22ping%22%3A+%22pong%22%7D` |
| `aggregator.livenessProbe.httpGet.port` | Name or number of the port to access on the liveness probes for the aggregator | `http` |
| `aggregator.livenessProbe.initialDelaySeconds` | Delay before liveness probe is initiated | `60` |
| `aggregator.livenessProbe.periodSeconds` | How often to perform the probe | `10` |
| `aggregator.livenessProbe.timeoutSeconds` | When the probe times out | `5` |
| `aggregator.livenessProbe.failureThreshold` | Minimum consecutive failures for the probe to be considered failed after having succeeded. | `6` |
| `aggregator.livenessProbe.successThreshold` | Minimum consecutive successes for the probe to be considered successful after having failed. | `1` |
| `aggregator.readinessProbe.enabled` | Enable readiness probes for the aggregator | `true` |
| `aggregator.readinessProbe.initialDelaySeconds` | Delay before readiness probe is initiated | `5` |
| `aggregator.readinessProbe.periodSeconds` | How often to perform the probe | `10` |
| `aggregator.readinessProbe.timeoutSeconds` | When the probe times out | `5` |
| `aggregator.readinessProbe.failureThreshold` | Minimum consecutive failures for the probe to be considered failed after having succeeded. | `6` |
| `aggregator.readinessProbe.successThreshold` | Minimum consecutive successes for the probe to be considered successful after having failed. | `1` |
| `aggregator.updateStrategy` | Update strategy for the aggregator DaemonSet | `RollingUpdate` |
| `aggregator.resources` | Configure resource requests and limits | `nil` |
| `aggregator.podAffinityPreset` | Aggregator Pod affinity preset. Ignored if `affinity` is set. Allowed values: `soft` or `hard` | `""` |
| `aggregator.podAntiAffinityPreset` | Aggregator Pod anti-affinity preset. Ignored if `affinity` is set. Allowed values: `soft` or `hard` | `soft` |
| `aggregator.nodeAffinityPreset.type` | Aggregator Node affinity preset type. Ignored if `affinity` is set. Allowed values: `soft` or `hard` | `""` |
| `aggregator.nodeAffinityPreset.key` | Aggregator Node label key to match Ignored if `affinity` is set. | `""` |
| `aggregator.nodeAffinityPreset.values` | Aggregator Node label values to match. Ignored if `affinity` is set. | `[]` |
| `aggregator.affinity` | Aggregator Affinity for pod assignment | `{}` (evaluated as a template) |
| `aggregator.nodeSelector` | Aggregator Node labels for pod assignment | `{}` (evaluated as a template) |
| `aggregator.tolerations` | Aggregator Tolerations for pod assignment | `[]` (evaluated as a template) |
| `aggregator.podAnnotations` | Pod annotations | `{}` |
| `aggregator.serviceAccount.create` | Specify whether a ServiceAccount should be created. | `false` |
| `aggregator.serviceAccount.name` | The name of the ServiceAccount to create | Generated using the `fluentd.fullname` template |
| `aggregator.serviceAccount.annotations` | Additional Service Account annotations (evaluated as a template) | `{}` |
| `aggregator.autoscaling.enabled` | Create an Horizontal Pod Autoscaler | `false` |
| `aggregator.autoscaling.minReplicas` | Minimum number of replicas for the HPA | `2` |
| `aggregator.autoscaling.maxReplicas` | Maximum number of replicas for the HPA | `5` |
| `aggregator.autoscaling.metrics` | Metrics for the HPA to manage the scaling | `Check values.yaml` |
| `aggregator.initContainers` | Additional init containers to add to the pods | `[]` |
| `aggregator.sidecars` | Add additional containers to the pods | `[]` |
| `aggregator.extraVolumes` | Extra volumes | `nil` |
| `aggregator.extraVolumeMounts` | Mount extra volume(s) | `nil` |
| `metrics.enabled` | Enable the export of Prometheus metrics | `nil` |
| `metrics.service.type` | Prometheus metrics service type | `ClusterIP` |
| `metrics.service.loadBalancerIP` | Load Balancer IP if the Prometheus metrics server type is `LoadBalancer` | `nil` |
| `metrics.service.port` | Prometheus metrics service port | `24231` |
| `metrics.service.annotations` | Annotations for Prometheus metrics service | `{ prometheus.io/scrape: "true", prometheus.io/port: "80", prometheus.io/path: "_prometheus/metrics" }` |
| `metrics.serviceMonitor.enabled` | if `true`, creates a Prometheus Operator ServiceMonitor (also requires `metrics.enabled` to be `true`) | `false` |
| `metrics.serviceMonitor.namespace` | Namespace in which Prometheus is running | `nil` |
| `metrics.serviceMonitor.interval` | Interval at which metrics should be scraped. | `nil` (Prometheus Operator default value) |
| `metrics.serviceMonitor.scrapeTimeout` | Timeout after which the scrape is ended | `nil` (Prometheus Operator default value) |
| `metrics.serviceMonitor.selector` | Prometheus instance selector labels | `nil` |
| `metrics.serviceMonitor.labels` | ServiceMonitor extra labels | `nil` |
| `metrics.serviceMonitor.annotations` | ServiceMonitor annotations | `nil` |
| `tls.enabled` | Enable the addition of TLS certificates | `false` |
| `tls.caCertificate` | Ca certificate | Certificate Authority (CA) bundle content |
| `tls.certificate` | Server certificate | Server certificate content |
| `tls.key` | Server Key | Server private key content |
| `tls.existingSecret` | Existing secret with certificate content | `nil` |
Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`. For example,
```console
$ helm install my-release \
--set aggregator.port=24444 bitnami/fluentd
```
The above command sets the aggregators to listen on port 24444.
Alternatively, a YAML file that specifies the values for the parameters can be provided while installing the chart. For example,
```console
$ helm install my-release -f values.yaml bitnami/fluentd
```
> **Tip**: You can use the default [values.yaml](values.yaml)
## Configuration and installation details
### [Rolling VS Immutable tags](https://docs.bitnami.com/containers/how-to/understand-rolling-tags-containers/)
It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.
Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.
### Forwarding the logs to another service
By default, the aggregators in this chart will send the processed logs to the standard output. However, a common practice is to send them to another service, like Elasticsearch, instead. This can be achieved with this Helm Chart by mounting your own configuration files. For example:
**configmap.yaml**
```yaml
apiVersion: v1
kind: ConfigMap
metadata:
name: elasticsearch-output
data:
fluentd.conf: |
# Prometheus Exporter Plugin
# input plugin that exports metrics
@type prometheus
port 24231
# input plugin that collects metrics from MonitorAgent
@type prometheus_monitor
host ${hostname}
# input plugin that collects metrics for output plugin
@type prometheus_output_monitor
host ${hostname}
# Ignore fluentd own events
@type null
# TCP input to receive logs from the forwarders
@type forward
bind 0.0.0.0
port 24224
# HTTP input for the liveness and readiness probes
@type http
bind 0.0.0.0
port 9880
# Throw the healthcheck to the standard output instead of forwarding it
@type stdout
# Send the logs to the standard output
@type elasticsearch
include_tag_key true
host "#{ENV['ELASTICSEARCH_HOST']}"
port "#{ENV['ELASTICSEARCH_PORT']}"
logstash_format true
@type file
path /opt/bitnami/fluentd/logs/buffers/logs.buffer
flush_thread_count 2
flush_interval 5s
```
As an example, using the above configmap, you should specify the required parameters when upgrading or installing the chart:
```console
aggregator.configMap=elasticsearch-output
aggregator.extraEnv[0].name=ELASTICSEARCH_HOST
aggregator.extraEnv[0].value=your-ip-here
aggregator.extraEnv[1].name=ELASTICSEARCH_PORT
aggregator.extraEnv[1].value=your-port-here
```
### Forwarder Security Context & Policy
By default, the **forwarder** `DaemonSet` from this chart **runs as the `root` user**, within the `root` group, assigning `root` file system permissions. This is different to the default behaviour of most Bitnami Helm charts where we [prefer to work with non-root containers](https://docs.bitnami.com/tutorials/work-with-non-root-containers/).
The default behaviour is to run as `root` because:
- the forwarder needs to mount `hostPath` volumes from the underlying node to read Docker container (& potentially other) logs
- in many Kubernetes node distributions, these log files are not readable by anyone other than `root`
- `fsGroup` doesn't work with `hostPath` volumes to allow the process to run non-root with alternate file system permissions
Since we would like the chart to work out-of-the-box for as many users as possible, the `forwarder` thus runs as root by default. You can read more about the motivation for this at [#1905](https://github.com/bitnami/charts/issues/1905) and [#2323](https://github.com/bitnami/charts/pull/2323), however you should be aware of this, and the risks of running root containers in general.
If you enable the forwarder's [bundled PodSecurityPolicy](templates/forwarder-psp.yaml) with `forwarder.rbac.pspEnabled=true` it will allow the pod to run as `root` by default, while ensuring as many other privileges as possible are dropped.
#### Running as non-root
You can run as the `fluentd` user/group (non-root) with the below overrides if:
- you have control of the `hostPath` filesystem permissions on your nodes sufficient to allow the fluentd user to read from them
- don't need to write to the `hostPath`s
Note that if you have enabled the [bundled PodSecurityPolicy](templates/forwarder-psp.yaml), it will adapt to the Chart values overrides.
```yaml
forwarder:
daemonUser: fluentd
daemonGroup: fluentd
securityContext:
runAsUser: 1001
runAsGroup: 1001
fsGroup: 1001
```
#### Pod Security Policy & Custom `hostPath`s
Mounting additional `hostPath`s is sometimes required to deal with `/var/lib` being symlinked on some Kubernetes environments. If you need to do so, the [bundled PodSecurityPolicy](templates/forwarder-psp.yaml) will likely not meet your needs, as it whitelists only the standard `hostPath`s.
### Setting Pod's affinity
This chart allows you to set your custom affinity using the `XXX.affinity` parameter(s). Find more information about Pod's affinity in the [kubernetes documentation](https://kubernetes.io/docs/concepts/configuration/assign-pod-node/#affinity-and-anti-affinity).
As an alternative, you can use of the preset configurations for pod affinity, pod anti-affinity, and node affinity available at the [bitnami/common](https://github.com/bitnami/charts/tree/master/bitnami/common#affinities) chart. To do so, set the `XXX.podAffinityPreset`, `XXX.podAntiAffinityPreset`, or `XXX.nodeAffinityPreset` parameters.
## Troubleshooting
Find more information about how to deal with common errors related to Bitnami’s Helm charts in [this troubleshooting guide](https://docs.bitnami.com/general/how-to/troubleshoot-helm-chart-issues).
## Upgrading
### To 3.1.0
This version also introduces `bitnami/common`, a [library chart](https://helm.sh/docs/topics/library_charts/#helm) as a dependency. More documentation about this new utility could be found [here](https://github.com/bitnami/charts/tree/master/bitnami/common#bitnami-common-library-chart). Please, make sure that you have updated the chart dependencies before executing any upgrade.
### To 3.0.0
[On November 13, 2020, Helm v2 support was formally finished](https://github.com/helm/charts#status-of-the-project), this major version is the result of the required changes applied to the Helm Chart to be able to incorporate the different features added in Helm v3 and to be consistent with the Helm project itself regarding the Helm v2 EOL.
**What changes were introduced in this major version?**
- Previous versions of this Helm Chart use `apiVersion: v1` (installable by both Helm 2 and 3), this Helm Chart was updated to `apiVersion: v2` (installable by Helm 3 only). [Here](https://helm.sh/docs/topics/charts/#the-apiversion-field) you can find more information about the `apiVersion` field.
- The different fields present in the *Chart.yaml* file has been ordered alphabetically in a homogeneous way for all the Bitnami Helm Charts
**Considerations when upgrading to this version**
- If you want to upgrade to this version from a previous one installed with Helm v3, you shouldn't face any issues
- If you want to upgrade to this version using Helm v2, this scenario is not supported as this version doesn't support Helm v2 anymore
- If you installed the previous version with Helm v2 and wants to upgrade to this version with Helm v3, please refer to the [official Helm documentation](https://helm.sh/docs/topics/v2_v3_migration/#migration-use-cases) about migrating from Helm v2 to v3
**Useful links**
- https://docs.bitnami.com/tutorials/resolve-helm2-helm3-post-migration-issues/
- https://helm.sh/docs/topics/v2_v3_migration/
- https://helm.sh/blog/migrate-from-helm-v2-to-helm-v3/
### To 2.0.0
This version introduces the ability to create/customise a `ServiceAccount` to be used by the **aggregator**, making it possible to target the aggregator with [`PodSecurityPolicy`](https://kubernetes.io/docs/concepts/policy/pod-security-policy/) independent of the **forwarder**'s `ServiceAccount`.
The **forwarder** previously used the below top-level values to configure its own `ServiceAccount`, which have been moved under the `forwarder.` prefix to avoid confusion, and only created if `forwarder.enabled=true`. There is no functional change as a result of this, and if you did not override the defaults for `serviceAccount` or `rbac`, this change does not require any action from you.
If you are overriding the default values from the `1.x` chart, the chart will fail installation with your old overrides and warn you of the necessary changes.
```yaml
# before - 1.x
serviceAccount:
create: true
name: my-custom-service-account
annotations:
my-custom-annotation: my-custom-annotation-value
rbac:
create: true
# after - 2.x
forwarder:
# ...
serviceAccount:
create: true
name: my-custom-service-account
annotations:
my-custom-annotation: my-custom-annotation-value
rbac:
create: true
```
### 1.0.0
In this version of the chart the Fluentd forwarder daemon system user will be root by default. This is done to ensure that mounted host paths are readable by the forwarder. For more context, check this [support case](https://github.com/bitnami/charts/issues/1905).
No issues are expected in the upgrade process. However, please ensure that you add extra security measures in your cluster as you will be running root containers. If you want the daemon to be run as a user different from root, you can change the `forwarder.daemonUser` and `forwarder.daemonGroup` values. In this case make sure that the user you choose has sufficient permissions to read log files under `/var/lib/docker/containers` directory.