Files
charts/bitnami/orangehrm

OrangeHRM

OrangeHRM is a free HR management system that offers a wealth of modules to suit the needs of your business. This widely-used system is feature-rich, intuitive and provides an essential HR management platform along with free documentation and access to a broad community of users.

TL;DR

$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/orangehrm

Introduction

This chart bootstraps a OrangeHRM deployment on a Kubernetes cluster using the Helm package manager.

It also packages the Bitnami MariaDB chart which is required for bootstrapping a MariaDB deployment for the database requirements of the OrangeHRM application.

Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters. This chart has been tested to work with NGINX Ingress, cert-manager, fluentd and Prometheus on top of the BKPR.

Prerequisites

  • Kubernetes 1.12+
  • Helm 3.1.0
  • PV provisioner support in the underlying infrastructure
  • ReadWriteMany volumes for deployment scaling

Installing the Chart

To install the chart with the release name my-release:

$ helm install my-release bitnami/orangehrm

The command deploys OrangeHRM on the Kubernetes cluster in the default configuration. The Parameters section lists the parameters that can be configured during installation.

Tip

: List all releases using helm list

Uninstalling the Chart

To uninstall/delete the my-release deployment:

$ helm delete my-release

The command removes all the Kubernetes components associated with the chart and deletes the release.

Parameters

Global parameters

Name Description Value
global.imageRegistry Global Docker image registry ""
global.imagePullSecrets Global Docker registry secret names as an array []
global.storageClass Global StorageClass for Persistent Volume(s) ""

Common parameters

Name Description Value
kubeVersion Force target Kubernetes version (using Helm capabilities if not set) ""
nameOverride String to partially override orangehrm.fullname template (will maintain the release name) ""
fullnameOverride String to fully override orangehrm.fullname template ""
extraDeploy Array of extra objects to deploy with the release (evaluated as a template) []
commonAnnotations Common annotations to add to all OrangeHRM resources (sub-charts are not considered). Evaluated as a template {}
commonLabels Common labels to add to all OrangeHRM resources (sub-charts are not considered). Evaluated as a template {}

OrangeHRM parameters

Name Description Value
image.registry OrangeHRM image registry docker.io
image.repository OrangeHRM image repository bitnami/orangehrm
image.tag OrangeHRM Image tag (immutable tags are recommended) 4.8.0-0-debian-10-r181
image.pullPolicy OrangeHRM image pull policy IfNotPresent
image.pullSecrets Specify docker-registry secret names as an array []
image.debug Specify if debug logs should be enabled false
hostAliases Deployment pod host aliases []
replicaCount Number of OrangeHRM Pods to run (requires ReadWriteMany PVC support) 1
orangehrmSkipInstall Skip OrangeHRM installation wizard. Useful for migrations and restoring from SQL dump false
orangehrmUsername User of the application admin
orangehrmPassword Application password ""
orangehrmEnforcePasswordStrength Whether the OrangeHRM password validation should use strong or medium level true
allowEmptyPassword Allow DB blank passwords true
command Override default container command (useful when using custom images) []
args Override default container args (useful when using custom images) []
updateStrategy.type Update strategy - only really applicable for deployments with RWO PVs attached RollingUpdate
extraEnvVars An array to add extra env vars []
extraEnvVarsCM ConfigMap with extra environment variables ""
extraEnvVarsSecret Secret with extra environment variables ""
extraVolumes Array of extra volumes to be added to the deployment. Requires setting extraVolumeMounts []
extraVolumeMounts Array of extra volume mounts to be added to the container. Normally used with extraVolumes []
initContainers Extra init containers to add to the deployment []
sidecars Extra sidecar containers to add to the deployment []
tolerations Tolerations for pod assignment []
existingSecret Use existing secret for the application password ""
smtpHost SMTP host ""
smtpPort SMTP port ""
smtpUser SMTP user ""
smtpPassword SMTP password. Ignored if smtpExistingSecret is set ""
smtpProtocol SMTP Protocol (options: ssl, none) ""
smtpExistingSecret Use an existing secret for the SMTP Password ""
containerPorts Container ports {}
sessionAffinity Control where client requests go, to the same pod or round-robin None
persistence.enabled Enable persistence using PVC true
persistence.storageClass OrangeHRM Data Persistent Volume Storage Class ""
persistence.accessMode PVC Access Mode for OrangeHRM volume ReadWriteOnce
persistence.size PVC Storage Request for OrangeHRM volume 8Gi
persistence.existingClaim A manually managed Persistent Volume Claim ""
persistence.hostPath If defined, the orangehrm-data volume will mount to the specified hostPath ""
persistence.subPath volumeMount subPath, use it for compatibility with previous versions of the chart orangehrm
podAffinityPreset Pod affinity preset. Ignored if affinity is set. Allowed values: soft or hard ""
podAntiAffinityPreset Pod anti-affinity preset. Ignored if affinity is set. Allowed values: soft or hard soft
nodeAffinityPreset.type Node affinity preset type. Ignored if affinity is set. Allowed values: soft or hard ""
nodeAffinityPreset.key Node label key to match Ignored if affinity is set. ""
nodeAffinityPreset.values Node label values to match. Ignored if affinity is set. []
affinity Affinity for pod assignment {}
nodeSelector Node labels for pod assignment. Evaluated as a template. {}
resources.requests The requested resources for the container {}
podSecurityContext.enabled Enable OrangeHRM pods' Security Context true
podSecurityContext.fsGroup OrangeHRM pods' group ID 1001
containerSecurityContext.enabled Enable OrangeHRM containers' Security Context true
containerSecurityContext.runAsUser OrangeHRM containers' Security Context 1001
livenessProbe.enabled Enable livenessProbe true
livenessProbe.path Request path for livenessProbe /symfony/web/index.php
livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 120
livenessProbe.periodSeconds Period seconds for livenessProbe 10
livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 5
livenessProbe.failureThreshold Failure threshold for livenessProbe 6
livenessProbe.successThreshold Success threshold for livenessProbe 1
readinessProbe.enabled Enable readinessProbe true
readinessProbe.path Request path for readinessProbe /symfony/web/index.php
readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 30
readinessProbe.periodSeconds Period seconds for readinessProbe 5
readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 3
readinessProbe.failureThreshold Failure threshold for readinessProbe 6
readinessProbe.successThreshold Success threshold for readinessProbe 1
customLivenessProbe Override default liveness probe {}
customReadinessProbe Override default readiness probe {}
lifecycleHooks LifecycleHooks for the container to automate configuration before or after startup {}
podAnnotations Pod annotations {}
podLabels Pod extra labels {}

Traffic Exposure Parameters

Name Description Value
service.type Kubernetes Service type LoadBalancer
service.port Service HTTP port 80
service.httpsPort Service HTTPS port 443
service.httpsTargetPort Service Target HTTPS port https
service.clusterIP OrangeHRM service cluster IP ""
service.loadBalancerSourceRanges Control hosts connecting to "LoadBalancer" only []
service.loadBalancerIP Load balancer IP for the OrangeHRM Service (optional, cloud specific) ""
service.nodePorts.http Kubernetes HTTP node port ""
service.nodePorts.https Kubernetes HTTPS node port ""
service.externalTrafficPolicy Enable client source IP preservation Cluster
service.annotations Provide any additional annotations that may be required (evaluated as a template) {}
service.extraPorts Extra ports to expose in the service (normally used with the sidecar value) []
ingress.enabled Set to true to enable ingress record generation false
ingress.pathType Ingress path type ImplementationSpecific
ingress.apiVersion Override API Version (automatically detected if not set) ""
ingress.hostname Default host for the ingress resource orangehrm.local
ingress.path The Path to OrangeHRM. You may need to set this to '/*' in order to use this with ALB ingress controllers. /
ingress.annotations Additional annotations for the Ingress resource. To enable certificate autogeneration, place here your cert-manager annotations. {}
ingress.tls Enable TLS configuration for the hostname defined at ingress.hostname parameter false
ingress.extraHosts The list of additional hostnames to be covered with this ingress record. []
ingress.extraPaths Any additional arbitrary paths that may need to be added to the ingress under the main host. []
ingress.extraTls The tls configuration for additional hostnames to be covered with this ingress record. []
ingress.secrets If you're providing your own certificates, please use this to add the certificates as secrets []

Database parameters

Name Description Value
mariadb.enabled Whether to deploy a MariaDB server to satisfy the applications database requirements true
mariadb.architecture MariaDB architecture. Allowed values: standalone or replication standalone
mariadb.auth.rootPassword Password for the MariaDB root user ""
mariadb.auth.database Database name to create bitnami_orangehrm
mariadb.auth.username Database user to create bn_orangehrm
mariadb.auth.password Password for the database ""
mariadb.primary.persistence.enabled Enable database persistence using PVC true
mariadb.primary.persistence.storageClass MariaDB primary persistent volume storage Class ""
mariadb.primary.persistence.accessModes Database Persistent Volume Access Modes ["ReadWriteOnce"]
mariadb.primary.persistence.size Database Persistent Volume Size 8Gi
mariadb.primary.persistence.hostPath Set path in case you want to use local host path volumes (not recommended in production) ""
mariadb.primary.persistence.existingClaim Name of an existing PersistentVolumeClaim for MariaDB primary replicas ""
externalDatabase.host Host of the existing database ""
externalDatabase.port Port of the existing database 3306
externalDatabase.user Existing username in the external db bn_orangehrm
externalDatabase.password Password for the above username ""
externalDatabase.database Name of the existing database bitnami_orangehrm

Volume Permissions parameters

Name Description Value
volumePermissions.enabled Enable init container that changes volume permissions in the data directory (for cases where the default k8s runAsUser and fsUser values do not work) false
volumePermissions.image.registry Init container volume-permissions image registry docker.io
volumePermissions.image.repository Init container volume-permissions image repository bitnami/bitnami-shell
volumePermissions.image.tag Init container volume-permissions image tag (immutable tags are recommended) 10-debian-10-r215
volumePermissions.image.pullPolicy Init container volume-permissions image pull policy IfNotPresent
volumePermissions.image.pullSecrets Specify docker-registry secret names as an array []
volumePermissions.resources.limits The resources limits for the container {}
volumePermissions.resources.requests The requested resources for the container {}

Metrics parameters

Name Description Value
metrics.enabled Start a side-car prometheus exporter false
metrics.image.registry Apache exporter image registry docker.io
metrics.image.repository Apache exporter image repository bitnami/apache-exporter
metrics.image.tag Apache exporter image tag (immutable tags are recommended) 0.10.1-debian-10-r16
metrics.image.pullPolicy Apache exporter image pull policy IfNotPresent
metrics.image.pullSecrets Specify docker-registry secret names as an array []
metrics.resources Metrics exporter resource requests and limits {}
metrics.podAnnotations Metrics exporter pod Annotation and Labels {}

Certificate injection parameters

Name Description Value
certificates.customCertificate.certificateSecret Secret containing the certificate and key to add ""
certificates.customCertificate.chainSecret.name Name of the secret containing the certificate chain ""
certificates.customCertificate.chainSecret.key Key of the certificate chain file inside the secret ""
certificates.customCertificate.certificateLocation Location in the container to store the certificate /etc/ssl/certs/ssl-cert-snakeoil.pem
certificates.customCertificate.keyLocation Location in the container to store the private key /etc/ssl/private/ssl-cert-snakeoil.key
certificates.customCertificate.chainLocation Location in the container to store the certificate chain /etc/ssl/certs/mychain.pem
certificates.customCAs Defines a list of secrets to import into the container trust store []
certificates.command Override default container command (useful when using custom images) []
certificates.args Override default container args (useful when using custom images) []
certificates.extraEnvVars Container sidecar extra environment variables []
certificates.extraEnvVarsCM ConfigMap with extra environment variables ""
certificates.extraEnvVarsSecret Secret with extra environment variables ""
certificates.image.registry Container sidecar registry docker.io
certificates.image.repository Container sidecar image repository bitnami/bitnami-shell
certificates.image.tag Container sidecar image tag (immutable tags are recommended) 10-debian-10-r215
certificates.image.pullPolicy Container sidecar image pull policy IfNotPresent
certificates.image.pullSecrets Container sidecar image pull secrets []

The above parameters map to the env variables defined in bitnami/orangehrm. For more information please refer to the bitnami/orangehrm image documentation.

Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,

$ helm install my-release \
  --set orangehrmUsername=admin,orangehrmPassword=password,mariadb.auth.rootPassword=secretpassword \
    bitnami/orangehrm

The above command sets the OrangeHRM administrator account username and password to admin and password respectively. Additionally, it sets the MariaDB root user password to secretpassword.

NOTE: Once this chart is deployed, it is not possible to change the application's access credentials, such as usernames or passwords, using Helm. To change these application credentials after deployment, delete any persistent volumes (PVs) used by the chart and re-deploy it, or use the application's built-in administrative tools if available.

Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,

$ helm install my-release -f values.yaml bitnami/orangehrm

Tip

: You can use the default values.yaml

Configuration and installation details

Rolling VS Immutable tags

It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.

Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.

Image

The image parameter allows specifying which image will be pulled for the chart.

Private registry

If you configure the image value to one in a private registry, you will need to specify an image pull secret.

  1. Manually create image pull secret(s) in the namespace. See this YAML example reference. Consult your image registry's documentation about getting the appropriate secret.
  2. Note that the imagePullSecrets configuration value cannot currently be passed to helm using the --set parameter, so you must supply these using a values.yaml file, such as:
imagePullSecrets:
  - name: SECRET_NAME
  1. Install the chart

Setting Pod's affinity

This chart allows you to set your custom affinity using the affinity parameter. Find more information about Pod's affinity in the kubernetes documentation.

As an alternative, you can use of the preset configurations for pod affinity, pod anti-affinity, and node affinity available at the bitnami/common chart. To do so, set the podAffinityPreset, podAntiAffinityPreset, or nodeAffinityPreset parameters.

TLS Secrets

This chart will facilitate the creation of TLS secrets for use with the ingress controller, however, this is not required. There are three common use cases:

  • Helm generates/manages certificate secrets
  • User generates/manages certificates separately
  • An additional tool (like kube-lego) manages the secrets for the application

In the first two cases, one will need a certificate and a key. We would expect them to look like this:

  • certificate files should look like (and there can be more than one certificate if there is a certificate chain)
-----BEGIN CERTIFICATE-----
MIID6TCCAtGgAwIBAgIJAIaCwivkeB5EMA0GCSqGSIb3DQEBCwUAMFYxCzAJBgNV
...
jScrvkiBO65F46KioCL9h5tDvomdU1aqpI/CBzhvZn1c0ZTf87tGQR8NK7v7
-----END CERTIFICATE-----
  • keys should look like:
-----BEGIN RSA PRIVATE KEY-----
MIIEogIBAAKCAQEAvLYcyu8f3skuRyUgeeNpeDvYBCDcgq+LsWap6zbX5f8oLqp4
...
wrj2wDbCDCFmfqnSJ+dKI3vFLlEz44sAV8jX/kd4Y6ZTQhlLbYc=
-----END RSA PRIVATE KEY-----

If you are going to use Helm to manage the certificates, please copy these values into the certificate and key values for a given ingress.secrets entry.

If you are going to manage TLS secrets outside of Helm, please know that you can create a TLS secret (named orangehrm.local-tls for example).

Please see this example for more information.

Ingress-terminated https

In cases where HTTPS/TLS is terminated on the ingress, you may run into an issue where non-https liveness and readiness probes result in a 302 (redirect from HTTP to HTTPS) and are interpreted by Kubernetes as not-live/not-ready. (See Kubernetes issue #47893 on GitHub for further details about 302 not being interpreted as "successful".) To work around this problem, use livenessProbeHeaders and readinessProbeHeaders to pass the same headers that your ingress would pass in order to get an HTTP 200 status result. For example (where the following is in a --values-referenced file):

livenessProbeHeaders:
  - name: X-Forwarded-Proto
    value: https
readinessProbeHeaders:
  - name: X-Forwarded-Proto
    value: https

Any number of name/value pairs may be specified; they are all copied into the liveness or readiness probe definition.

Persistence

The Bitnami OrangeHRM image stores the OrangeHRM configurations at the /bitnami/orangehrm path of the container.

Persistent Volume Claims are used to keep the data across deployments. This is known to work in GCE, AWS, and minikube. See the Parameters section to configure the PVC or to disable persistence.

Existing PersistentVolumeClaim

  1. Create the PersistentVolume
  2. Create the PersistentVolumeClaim
  3. Install the chart
$ helm install my-release --set persistence.existingClaim=PVC_NAME bitnami/orangehrm

Host path

System compatibility

  • The local filesystem accessibility to a container in a pod with hostPath has been tested on OSX/MacOS with xhyve, and Linux with VirtualBox.
  • Windows has not been tested with the supported VM drivers. Minikube does however officially support Mounting Host Folders per pod. Or you may manually sync your container whenever host files are changed with tools like docker-sync or docker-bg-sync.

Mounting steps

  1. The specified hostPath directory must already exist (create one if it does not).

  2. Install the chart

    $ helm install my-release --set persistence.hostPath=/PATH/TO/HOST/MOUNT bitnami/orangehrm
    

    This will mount the orangehrm-data volume into the hostPath directory. The site data will be persisted if the mount path contains valid data, else the site data will be initialized at first launch.

  3. Because the container cannot control the host machine's directory permissions, you must set the OrangeHRM file directory permissions yourself and disable or clear OrangeHRM cache.

Troubleshooting

Find more information about how to deal with common errors related to Bitnamis Helm charts in this troubleshooting guide.

Upgrading

To 10.0.0

This version standardizes the way of defining Ingress rules. When configuring a single hostname for the Ingress rule, set the ingress.hostname value. When defining more than one, set the ingress.extraHosts array. Apart from this case, no issues are expected to appear when upgrading.

To 9.0.0

The Bitnami OrangeHRM image was updated to support and enable the "non-root" user approach

If you want to continue to run the container image as the root user, you need to set podSecurityContext.enabled=false and containerSecurity.context.enabled=false.

This upgrade also adapts the chart to the latest Bitnami good practices. Check the Parameters section for more information.

To upgrade from 8.x.x, backup OrangeHRM data and the previous MariaDB databases, install a new OrangeHRM chart and import the backups and data, ensuring the 1001 user has the appropriate permissions on the migrated volume. As an alternative, you can use helm upgrade if you specify volumePermissions.enabled=true and persistence.subPath="".

To 8.0.0

In this major there were two main changes introduced:

  1. Adaptation to Helm v2 EOL
  2. Updated MariaDB dependency version

Please read the update notes carefully.

1. Adaptation to Helm v2 EOL

On November 13, 2020, Helm v2 support was formally finished, this major version is the result of the required changes applied to the Helm Chart to be able to incorporate the different features added in Helm v3 and to be consistent with the Helm project itself regarding the Helm v2 EOL.

What changes were introduced in this major version?

  • Previous versions of this Helm Chart use apiVersion: v1 (installable by both Helm 2 and 3), this Helm Chart was updated to apiVersion: v2 (installable by Helm 3 only). Here you can find more information about the apiVersion field.
  • Move dependency information from the requirements.yaml to the Chart.yaml
  • After running helm dependency update, a Chart.lock file is generated containing the same structure used in the previous requirements.lock
  • The different fields present in the Chart.yaml file has been ordered alphabetically in a homogeneous way for all the Bitnami Helm Charts

Considerations when upgrading to this version

  • If you want to upgrade to this version from a previous one installed with Helm v3, you shouldn't face any issues
  • If you want to upgrade to this version using Helm v2, this scenario is not supported as this version doesn't support Helm v2 anymore
  • If you installed the previous version with Helm v2 and wants to upgrade to this version with Helm v3, please refer to the official Helm documentation about migrating from Helm v2 to v3

Useful links

2. Updated MariaDB dependency version

In this major the MariaDB dependency version was also bumped to a new major version that introduces several incompatilibites. Therefore, backwards compatibility is not guaranteed unless an external database is used. Check MariaDB Upgrading Notes for more information.

To upgrade to 8.0.0, it should be done reusing the PVCs used to hold both the MariaDB and OrangeHRM data on your previous release. To do so, follow the instructions below (the following example assumes that the release name is orangehrm):

NOTE: Please, create a backup of your database before running any of those actions. The steps below would be only valid if your application (e.g. any plugins or custom code) is compatible with MariaDB 10.5.x

Obtain the credentials and the names of the PVCs used to hold both the MariaDB and OrangeHRM data on your current release:

export ORANGEHRM_PASSWORD=$(kubectl get secret --namespace default orangehrm -o jsonpath="{.data.orangehrm-password}" | base64 --decode)
export MARIADB_ROOT_PASSWORD=$(kubectl get secret --namespace default orangehrm-mariadb -o jsonpath="{.data.mariadb-root-password}" | base64 --decode)
export MARIADB_PASSWORD=$(kubectl get secret --namespace default orangehrm-mariadb -o jsonpath="{.data.mariadb-password}" | base64 --decode)
export MARIADB_PVC=$(kubectl get pvc -l app=mariadb,component=master,release=orangehrm -o jsonpath="{.items[0].metadata.name}")

Delete the OrangeHRM deployment and delete the MariaDB statefulset. Notice the option --cascade=false in the latter.

$ kubectl delete deployments.apps orangehrm

$ kubectl delete statefulsets.apps orangehrm-mariadb --cascade=false

Now the upgrade works:

$ helm upgrade orangehrm bitnami/orangehrm --set mariadb.primary.persistence.existingClaim=$MARIADB_PVC --set mariadb.auth.rootPassword=$MARIADB_ROOT_PASSWORD --set mariadb.auth.password=$MARIADB_PASSWORD --set orangehrmPassword=$ORANGEHRM_PASSWORD

You will have to delete the existing MariaDB pod and the new statefulset is going to create a new one

$ kubectl delete pod osclass-mariadb-0

Finally, you should see the lines below in MariaDB container logs:

$ kubectl logs $(kubectl get pods -l app.kubernetes.io/instance=orangehrm,app.kubernetes.io/name=mariadb,app.kubernetes.io/component=primary -o jsonpath="{.items[0].metadata.name}")
...
mariadb 12:13:24.98 INFO  ==> Using persisted data
mariadb 12:13:25.01 INFO  ==> Running mysql_upgrade
...

To 7.0.0

Helm performs a lookup for the object based on its group (apps), version (v1), and kind (Deployment). Also known as its GroupVersionKind, or GVK. Changing the GVK is considered a compatibility breaker from Kubernetes' point of view, so you cannot "upgrade" those objects to the new GVK in-place. Earlier versions of Helm 3 did not perform the lookup correctly which has since been fixed to match the spec.

In this version the apiVersion of the deployment resources is updated to apps/v1 in tune with the api's deprecated, resulting in compatibility breakage.

This major version signifies this change.

To 3.0.0

Backwards compatibility is not guaranteed unless you modify the labels used on the chart's deployments. Use the workaround below to upgrade from versions previous to 3.0.0. The following example assumes that the release name is orangehrm:

$ kubectl patch deployment orangehrm-orangehrm --type=json -p='[{"op": "remove", "path": "/spec/selector/matchLabels/chart"}]'
$ kubectl delete statefulset orangehrm-mariadb --cascade=false