* Fix default values in parse * Fix default values in phabricator * Fix default values in phpbb * Fix default values in phpmyadmin * Fix default values in postgresql * Fix default values in postgresql-ha * Fix default values in prestashop * Fix default values in pytorch * Fix default values in rabbitmq * Fix default values in oauth2-proxy * Update bitnami/oauth2-proxy/values.yaml * Update bitnami/rabbitmq/values.yaml * Update bitnami/oauth2-proxy/values.yaml * Update README.md * Update README.md
Phabricator
Phabricator is a collection of open source web applications that help software companies build better software. Phabricator is built by developers for developers. Every feature is optimized around developer efficiency for however you like to work. Code Quality starts with an effective collaboration between team members.
TL;DR
$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/phabricator
Introduction
This chart bootstraps a Phabricator deployment on a Kubernetes cluster using the Helm package manager.
It also packages the Bitnami MariaDB chart which is required for bootstrapping a MariaDB deployment for the database requirements of the Phabricator application.
Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters. This chart has been tested to work with NGINX Ingress, cert-manager, fluentd and Prometheus on top of the BKPR.
Prerequisites
- Kubernetes 1.12+
- Helm 3.1.0
- PV provisioner support in the underlying infrastructure
- ReadWriteMany volumes for deployment scaling
Installing the Chart
To install the chart with the release name my-release:
$ helm install my-release bitnami/phabricator
The command deploys Phabricator on the Kubernetes cluster in the default configuration. The Parameters section lists the parameters that can be configured during installation.
Tip
: List all releases using
helm list
Uninstalling the Chart
To uninstall/delete the my-release deployment:
$ helm delete my-release
The command removes all the Kubernetes components associated with the chart and deletes the release.
Parameters
Global parameters
| Name | Description | Value |
|---|---|---|
global.imageRegistry |
Global Docker image registry | "" |
global.imagePullSecrets |
Global Docker registry secret names as an array | [] |
global.storageClass |
Global StorageClass for Persistent Volume(s) | "" |
Common parameters
| Name | Description | Value |
|---|---|---|
nameOverride |
String to partially override common.names.fullname template (will maintain the release name) | "" |
fullnameOverride |
String to fully override common.names.fullname template | "" |
commonLabels |
Add labels to all the deployed resources | {} |
commonAnnotations |
Add annotations to all the deployed resources | {} |
kubeVersion |
Force target Kubernetes version (using Helm capabilities if not set) | "" |
clusterDomain |
Kubernetes Cluster Domain | cluster.local |
extraDeploy |
Extra objects to deploy (value evaluated as a template) | [] |
Phabricator parameters
| Name | Description | Value |
|---|---|---|
image.registry |
Phabricator image registry | docker.io |
image.repository |
Phabricator image repository | bitnami/phabricator |
image.tag |
Phabricator image tag (immutable tags are recommended) | 2021.26.0-debian-10-r11 |
image.pullPolicy |
Image pull policy | IfNotPresent |
image.pullSecrets |
Specify docker-registry secret names as an array | [] |
image.debug |
Specify if debug logs should be enabled | false |
phabricatorUsername |
Phabricator admin user | user |
phabricatorPassword |
Admin user password | "" |
phabricatorEmail |
Admin user email | user@example.com |
phabricatorFirstName |
Admin user first name | First Name |
phabricatorLastName |
Admin user last name | Last Name |
phabricatorHost |
Phabricator host to create application URLs | "" |
phabricatorAlternateFileDomain |
Phabricator alternate domain to upload files | "" |
phabricatorEnableHttps |
Configure Phabricator to build application URLs using https | true |
phabricatorUseLFS |
Configure Phabricator to use GIT Large File Storage (LFS) | false |
phabricatorGitSSH |
Configure a self-hosted GIT repository with SSH authentication | false |
phabricatorEnablePygments |
Enable syntax highlighting using Pygments | true |
phabricatorSkipInstall |
Skip the initial bootstrapping for the application | false |
smtpHost |
SMTP mail delivery host | "" |
smtpPort |
SMTP mail delivery port | "" |
smtpUser |
SMTP mail delivery user | "" |
smtpPassword |
SMTP mail delivery password | "" |
smtpProtocol |
SMTP mail delivery protocol [ssl, tls] |
"" |
command |
Override default container command (useful when using custom images) | [] |
args |
Override default container args (useful when using custom images) | [] |
extraEnvVars |
An array to add extra env vars | [] |
extraEnvVarsCM |
Name of existing ConfigMap containing extra environment variables | "" |
extraEnvVarsSecret |
Name of existing Secret containing extra environment variables | "" |
Phabricator deployment parameters
| Name | Description | Value |
|---|---|---|
updateStrategy.type |
StrategyType | RollingUpdate |
containerPorts |
Phabricator container ports to open | {} |
podSecurityContext.enabled |
Enable Phabricator pods' Security Context | true |
podSecurityContext.fsGroup |
Group ID for the volumes of the pod | 1001 |
containerSecurityContext.enabled |
Enable Phabricator containers' SecurityContext | true |
containerSecurityContext.runAsUser |
User ID for the container | 1001 |
resources.limits |
The resources limits for the container | {} |
resources.requests |
The requested resources for the container | {} |
livenessProbe.enabled |
Enable livenessProbe | true |
livenessProbe.path |
Request path for livenessProbe | /auth/ |
livenessProbe.initialDelaySeconds |
Initial delay seconds for livenessProbe | 180 |
livenessProbe.periodSeconds |
Period seconds for livenessProbe | 10 |
livenessProbe.timeoutSeconds |
Timeout seconds for livenessProbe | 5 |
livenessProbe.failureThreshold |
Failure threshold for livenessProbe | 6 |
livenessProbe.successThreshold |
Success threshold for livenessProbe | 1 |
readinessProbe.enabled |
Enable readinessProbe | true |
readinessProbe.path |
Request path for readinessProbe | /auth/ |
readinessProbe.initialDelaySeconds |
Initial delay seconds for readinessProbe | 30 |
readinessProbe.periodSeconds |
Period seconds for readinessProbe | 10 |
readinessProbe.timeoutSeconds |
Timeout seconds for readinessProbe | 5 |
readinessProbe.failureThreshold |
Failure threshold for readinessProbe | 6 |
readinessProbe.successThreshold |
Success threshold for readinessProbe | 1 |
startupProbe.enabled |
Enable startupProbe | false |
startupProbe.path |
Request path for startupProbe | /auth/ |
startupProbe.initialDelaySeconds |
Initial delay seconds for startupProbe | 0 |
startupProbe.periodSeconds |
Period seconds for startupProbe | 10 |
startupProbe.timeoutSeconds |
Timeout seconds for startupProbe | 3 |
startupProbe.failureThreshold |
Failure threshold for startupProbe | 20 |
startupProbe.successThreshold |
Success threshold for startupProbe | 1 |
customLivenessProbe |
Override default liveness probe | {} |
customReadinessProbe |
Override default readiness probe | {} |
customStartupProbe |
Custom Startup probe for Phabricator | {} |
podLabels |
Extra labels for Phabricator pods | {} |
podAnnotations |
Annotations for Phabricator pods | {} |
podAffinityPreset |
Pod affinity preset. Ignored if affinity is set. Allowed values: soft or hard |
"" |
podAntiAffinityPreset |
Pod anti-affinity preset. Ignored if affinity is set. Allowed values: soft or hard |
soft |
nodeAffinityPreset.type |
Node affinity preset type. Ignored if affinity is set. Allowed values: soft or hard |
"" |
nodeAffinityPreset.key |
Node label key to match. Ignored if affinity is set. |
"" |
nodeAffinityPreset.values |
Node label values to match. Ignored if affinity is set. |
[] |
affinity |
Affinity for pod assignment. Evaluated as a template. | {} |
nodeSelector |
Node labels for pod assignment. Evaluated as a template. | {} |
tolerations |
Tolerations for pod assignment. Evaluated as a template. | [] |
extraVolumes |
Optionally specify extra list of additional volumes for Phabricator pods | [] |
extraVolumeMounts |
Optionally specify extra list of additional volumeMounts for Phabricator container(s) | [] |
initContainers |
Add init containers to the Phabricator pods. | [] |
sidecars |
Add sidecars to the Phabricator pods | [] |
persistence.enabled |
Enable persistence using PVC | true |
persistence.storageClass |
PVC Storage Class for Phabricator volume | "" |
persistence.accessMode |
PVC Access Mode for Phabricator volume | ReadWriteOnce |
persistence.size |
PVC Storage Request for Phabricator volume | 8Gi |
persistence.existingClaim |
An Existing PVC name for Phabricator volume | "" |
persistence.hostPath |
Host mount path for Phabricator volume | "" |
Traffic Exposure Parameters
| Name | Description | Value |
|---|---|---|
service.type |
Kubernetes Service type | LoadBalancer |
service.port |
Service HTTP port | 80 |
service.httpsPort |
Service HTTP port | 443 |
service.sshPort |
Service SSH port | 22 |
service.nodePorts.http |
Kubernetes HTTP node port | "" |
service.nodePorts.https |
Kubernetes HTTPS node port | "" |
service.nodePorts.ssh |
Kubernetes SSH node port | "" |
service.loadBalancerIP |
loadBalancerIP for the Phabricator Service |
"" |
service.externalTrafficPolicy |
Enable client source IP preservation | Cluster |
service.annotations |
Service annotations | {} |
ingress.enabled |
Set to true to enable ingress record generation | false |
ingress.tls |
Enable TLS configuration for the hostname defined at ingress.hostname parameter | false |
ingress.certManager |
Set this to true in order to add the corresponding annotations for cert-manager | false |
ingress.pathType |
Ingress path type | ImplementationSpecific |
ingress.apiVersion |
Force Ingress API version (automatically detected if not set) | "" |
ingress.hostname |
Default host for the ingress resource | phabricator.local |
ingress.path |
The Path to Phabricator. You may need to set this to '/*' in order to use this with ALB ingress controllers. | ImplementationSpecific |
ingress.annotations |
Ingress annotations | {} |
ingress.extraHosts |
The list of additional hostnames to be covered with this ingress record. | [] |
ingress.extraPaths |
Any additional arbitrary paths that may need to be added to the ingress under the main host. | [] |
ingress.extraTls |
The tls configuration for additional hostnames to be covered with this ingress record. | [] |
ingress.secrets |
If you're providing your own certificates, please use this to add the certificates as secrets | [] |
Database parameters
| Name | Description | Value |
|---|---|---|
mariadb.enabled |
Whether to use the MariaDB chart | true |
mariadb.architecture |
MariaDB architecture. Allowed values: standalone or replication |
standalone |
mariadb.auth.rootPassword |
Password for the MariaDB root user |
"" |
mariadb.primary.configuration |
MariaDB Primary configuration to be injected as ConfigMap | "" |
mariadb.primary.extraFlags |
Additional command line flags | --local-infile=0 |
mariadb.primary.persistence.enabled |
Enable database persistence using PVC | true |
mariadb.primary.persistence.storageClass |
mariadb data Persistent Volume Storage Class | "" |
mariadb.primary.persistence.accessModes |
Database Persistent Volume Access Modes | [] |
mariadb.primary.persistence.size |
Database Persistent Volume Size | 8Gi |
mariadb.primary.persistence.hostPath |
Host mount path for MariaDB volume in case you want to use local host path volumes (not recommended in production) | "" |
mariadb.primary.persistence.existingClaim |
Enable persistence using an existing PVC | "" |
externalDatabase.existingSecret |
Name of the database existing Secret Object | "" |
externalDatabase.host |
Host of the existing database | localhost |
externalDatabase.port |
Port of the existing database | 3306 |
externalDatabase.rootUser |
Username in the external db with root privileges | root |
externalDatabase.rootPassword |
Password for the above username | "" |
Volume Permissions parameters
| Name | Description | Value |
|---|---|---|
volumePermissions.enabled |
Enable init container that changes volume permissions in the data directory ( | false |
volumePermissions.image.registry |
Init container volume-permissions image registry | docker.io |
volumePermissions.image.repository |
Init container volume-permissions image repository | bitnami/bitnami-shell |
volumePermissions.image.tag |
Init container volume-permissions image tag (immutable tags are recommended) | 10-debian-10-r127 |
volumePermissions.image.pullPolicy |
Init container volume-permissions image pull policy | Always |
volumePermissions.image.pullSecrets |
Specify docker-registry secret names as an array | [] |
volumePermissions.resources.limits |
The resources limits for the container | {} |
volumePermissions.resources.requests |
The requested resources for the container | {} |
Metrics parameters
| Name | Description | Value |
|---|---|---|
metrics.enabled |
Start a side-car prometheus exporter | false |
metrics.image.registry |
Apache exporter image registry | docker.io |
metrics.image.repository |
Apache exporter image repository | bitnami/apache-exporter |
metrics.image.tag |
Apache exporter image tag (immutable tags are recommended) | 0.9.0-debian-10-r24 |
metrics.image.pullPolicy |
Image pull policy | IfNotPresent |
metrics.image.pullSecrets |
Specify docker-registry secret names as an array | [] |
metrics.podAnnotations |
Additional annotations for Metrics exporter pod | {} |
metrics.resources.limits |
The resources limits for the container | {} |
metrics.resources.requests |
The requested resources for the container | {} |
The above parameters map to the env variables defined in bitnami/phabricator. For more information please refer to the bitnami/phabricator image documentation.
Note
:
For Phabricator to function correctly, you should specify the
phabricatorHostparameter to specify the FQDN (recommended) or the public IP address of the Phabricator service.Optionally, you can specify the
phabricatorLoadBalancerIPparameter to assign a reserved IP address to the Phabricator service of the chart. However please note that this feature is only available on a few cloud providers (f.e. GKE).To reserve a public IP address on GKE:
$ gcloud compute addresses create phabricator-public-ipThe reserved IP address can be associated to the Phabricator service by specifying it as the value of the
phabricatorLoadBalancerIPparameter while installing the chart.
Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,
$ helm install my-release \
--set phabricatorUsername=admin,phabricatorPassword=password,mariadb.auth.rootPassword=secretpassword \
bitnami/phabricator
The above command sets the Phabricator administrator account username and password to admin and password respectively. Additionally, it sets the MariaDB root user password to secretpassword.
NOTE: Once this chart is deployed, it is not possible to change the application's access credentials, such as usernames or passwords, using Helm. To change these application credentials after deployment, delete any persistent volumes (PVs) used by the chart and re-deploy it, or use the application's built-in administrative tools if available.
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,
$ helm install my-release -f values.yaml bitnami/phabricator
Tip
: You can use the default values.yaml
Configuration and installation details
Rolling VS Immutable tags
It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.
Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.
Using an external database
Sometimes you may want to have Phabricator connect to an external database rather than installing one inside your cluster, e.g. to use a managed database service, or use run a single database server for all your applications. To do this, the chart allows you to specify credentials for an external database under the externalDatabase parameter. You should also disable the MariaDB installation with the mariadb.enabled option. For example with the following parameters:
mariadb.enabled=false
externalDatabase.host=myexternalhost
externalDatabase.port=3306
externalDatabase.rootUser=myuser
externalDatabase.rootPassword=mypassword
Note also if you disable MariaDB per above you MUST supply values for the externalDatabase connection.
Ingress
This chart provides support for ingress resources. If you have an ingress controller installed on your cluster, such as nginx-ingress-controller or contour you can utilize the ingress controller to serve your application.
To enable ingress integration, please set ingress.enabled to true.
Hosts
Most likely you will only want to have one hostname that maps to this Phabricator installation. If that's your case, the property ingress.hostname will set it. However, it is possible to have more than one host. To facilitate this, the ingress.extraHosts object can be specified as an array. You can also use ingress.extraTLS to add the TLS configuration for extra hosts.
For each host indicated at ingress.extraHosts, please indicate a name, path, and any annotations that you may want the ingress controller to know about.
For annotations, please see this document. Not all annotations are supported by all ingress controllers, but this document does a good job of indicating which annotation is supported by many popular ingress controllers.
TLS Secrets
This chart will facilitate the creation of TLS secrets for use with the ingress controller, however, this is not required. There are three common use cases:
- Helm generates/manages certificate secrets.
- User generates/manages certificates separately.
- An additional tool (like cert-manager) manages the secrets for the application.
In the first two cases, it's needed a certificate and a key. We would expect them to look like this:
-
certificate files should look like (and there can be more than one certificate if there is a certificate chain)
-----BEGIN CERTIFICATE----- MIID6TCCAtGgAwIBAgIJAIaCwivkeB5EMA0GCSqGSIb3DQEBCwUAMFYxCzAJBgNV ... jScrvkiBO65F46KioCL9h5tDvomdU1aqpI/CBzhvZn1c0ZTf87tGQR8NK7v7 -----END CERTIFICATE----- -
keys should look like:
-----BEGIN RSA PRIVATE KEY----- MIIEogIBAAKCAQEAvLYcyu8f3skuRyUgeeNpeDvYBCDcgq+LsWap6zbX5f8oLqp4 ... wrj2wDbCDCFmfqnSJ+dKI3vFLlEz44sAV8jX/kd4Y6ZTQhlLbYc= -----END RSA PRIVATE KEY-----
If you are going to use Helm to manage the certificates, please copy these values into the certificate and key values for a given ingress.secrets entry.
If you are going to manage TLS secrets outside of Helm, please know that you can create a TLS secret (named phabricator.local-tls for example).
Adding extra environment variables
In case you want to add extra environment variables (useful for advanced operations like custom init scripts), you can use the extraEnvVars property.
extraEnvVars:
- name: LOG_LEVEL
value: DEBUG
Alternatively, you can use a ConfigMap or a Secret with the environment variables. To do so, use the extraEnvVarsCM or the extraEnvVarsSecret values.
Sidecars and Init Containers
If you have a need for additional containers to run within the same pod as the Phabricator app (e.g. an additional metrics or logging exporter), you can do so via the sidecars config parameter. Simply define your container according to the Kubernetes container spec.
sidecars:
- name: your-image-name
image: your-image
imagePullPolicy: Always
ports:
- name: portname
containerPort: 1234
Similarly, you can add extra init containers using the initContainers parameter.
initContainers:
- name: your-image-name
image: your-image
imagePullPolicy: Always
ports:
- name: portname
containerPort: 1234
Deploying extra resources
There are cases where you may want to deploy extra objects, such a ConfigMap containing your app's configuration or some extra deployment with a micro service used by your app. For covering this case, the chart allows adding the full specification of other objects using the extraDeploy parameter.
Setting Pod's affinity
This chart allows you to set your custom affinity using the affinity paremeter. Find more infomation about Pod's affinity in the kubernetes documentation.
As an alternative, you can use of the preset configurations for pod affinity, pod anti-affinity, and node affinity available at the bitnami/common chart. To do so, set the podAffinityPreset, podAntiAffinityPreset, or nodeAffinityPreset parameters.
Persistence
The Bitnami Phabricator image stores the Phabricator data and configurations at the /bitnami/phabricator path of the container.
Persistent Volume Claims are used to keep the data across deployments. There is a known issue in Kubernetes Clusters with EBS in different availability zones. Ensure your cluster is configured properly to create Volumes in the same availability zone where the nodes are running. Kuberentes 1.12 solved this issue with the Volume Binding Mode.
See the Parameters section to configure the PVC or to disable persistence.
Existing PersistentVolumeClaim
-
Create the PersistentVolume
-
Create the PersistentVolumeClaim
-
Install the chart
$ helm install my-release --set persistence.existingClaim=PVC_NAME bitnami/phabricator
Host path
System compatibility
- The local filesystem accessibility to a container in a pod with
hostPathhas been tested on OSX/MacOS with xhyve, and Linux with VirtualBox. - Windows has not been tested with the supported VM drivers. Minikube does however officially support Mounting Host Folders per pod. Or you may manually sync your container whenever host files are changed with tools like docker-sync or docker-bg-sync.
Mounting steps
-
The specified
hostPathdirectory must already exist (create one if it does not). -
Install the chart
$ helm install my-release --set persistence.hostPath=/PATH/TO/HOST/MOUNT bitnami/phabricatorThis will mount the
phabricator-datavolume into thehostPathdirectory. The site data will be persisted if the mount path contains valid data, else the site data will be initialized at first launch. -
Because the container cannot control the host machine's directory permissions, you must set the Phabricator file directory permissions yourself.
Troubleshooting
Find more information about how to deal with common errors related to Bitnami’s Helm charts in this troubleshooting guide.
Upgrading
To 11.0.0
In this major there were two main changes introduced:
- Parameter standarizations
- Migration to non-root
To upgrade to 11.0.0, backup Phabricator data and the previous MariaDB databases, install a new Phabricator chart and import the backups and data, ensuring the 1001 user has the appropriate permissions on the migrated volume.
1. Chart standarizations
This upgrade adapts the chart to the latest Bitnami good practices. Check the Parameters section for more information. In summary:
- New parameters were added.
- Some parameters were renamed or disappeared in favor of new ones in this major version.
- This version also introduces
bitnami/common, a library chart as a dependency. More documentation about this new utility could be found here. Please, make sure that you have updated the chart dependencies before executing any upgrade.
2. Migration of the Phabricator image to non-root
The Bitnami Phabricator image was migrated to a "non-root" user approach. Previously the container ran as the root user and the Apache daemon was started as the daemon while the Phabricator daemons were started as the phabricator user. From now on, both the container and the different daemons run as user 1001.. Consequences:
- The HTTP/HTTPS ports exposed by the container are now
8080/8443instead of80/443. - Backwards compatibility is not guaranteed. Uninstall & install the chart again to obtain the latest version.
You can revert this behavior by setting the parameters containerSecurityContext.runAsUser to root.
To 10.0.0
In this major there were two main changes introduced:
- Adaptation to Helm v2 EOL
- Updated MariaDB dependency version
Please read the update notes carefully.
1. Adaptation to Helm v2 EOL
On November 13, 2020, Helm v2 support was formally finished, this major version is the result of the required changes applied to the Helm Chart to be able to incorporate the different features added in Helm v3 and to be consistent with the Helm project itself regarding the Helm v2 EOL.
What changes were introduced in this major version?
- Previous versions of this Helm Chart use
apiVersion: v1(installable by both Helm 2 and 3), this Helm Chart was updated toapiVersion: v2(installable by Helm 3 only). Here you can find more information about theapiVersionfield. - Move dependency information from the requirements.yaml to the Chart.yaml
- After running
helm dependency update, a Chart.lock file is generated containing the same structure used in the previous requirements.lock - The different fields present in the Chart.yaml file has been ordered alphabetically in a homogeneous way for all the Bitnami Helm Charts
Considerations when upgrading to this version
- If you want to upgrade to this version from a previous one installed with Helm v3, you shouldn't face any issues
- If you want to upgrade to this version using Helm v2, this scenario is not supported as this version doesn't support Helm v2 anymore
- If you installed the previous version with Helm v2 and wants to upgrade to this version with Helm v3, please refer to the official Helm documentation about migrating from Helm v2 to v3
Useful links
- https://docs.bitnami.com/tutorials/resolve-helm2-helm3-post-migration-issues/
- https://helm.sh/docs/topics/v2_v3_migration/
- https://helm.sh/blog/migrate-from-helm-v2-to-helm-v3/
2. Updated MariaDB dependency version
In this major the MariaDB dependency version was also bumped to a new major version that introduces several incompatilibites. Therefore, backwards compatibility is not guaranteed unless an external database is used. Check MariaDB Upgrading Notes for more information.
To upgrade to 10.0.0, it should be done reusing the PVCs used to hold both the MariaDB and Phabricator data on your previous release. To do so, follow the instructions below (the following example assumes that the release name is phabricator and that a rootUser.password was defined for MariaDB in values.yaml when the chart was first installed):
NOTE: Please, create a backup of your database before running any of those actions. The steps below would be only valid if your application (e.g. any plugins or custom code) is compatible with MariaDB 10.5.x
Obtain the credentials and the names of the PVCs used to hold both the MariaDB and Phabricator data on your current release:
export PHABRICATOR_HOST=$(kubectl get svc --namespace default phabricator --template "{{ range (index .status.loadBalancer.ingress 0) }}{{ . }}{{ end }}")
export PHABRICATOR_PASSWORD=$(kubectl get secret --namespace default phabricator -o jsonpath="{.data.phabricator-password}" | base64 --decode)
export MARIADB_ROOT_PASSWORD=$(kubectl get secret --namespace default phabricator-mariadb -o jsonpath="{.data.mariadb-root-password}" | base64 --decode)
export MARIADB_PVC=$(kubectl get pvc -l app=mariadb,component=master,release=phabricator -o jsonpath="{.items[0].metadata.name}")
Delete the Phabricator deployment and delete the MariaDB statefulset. Notice the option --cascade=false in the latter.
$ kubectl delete deployments.apps phabricator
$ kubectl delete statefulsets.apps phabricator-mariadb --cascade=false
Now the upgrade works:
$ helm upgrade phabricator bitnami/phabricator --set mariadb.primary.persistence.existingClaim=$MARIADB_PVC --set mariadb.auth.rootPassword=$MARIADB_ROOT_PASSWORD --set phabricatorPassword=$PHABRICATOR_PASSWORD --set phabricatorHost=$PHABRICATOR_HOST
You will have to delete the existing MariaDB pod and the new statefulset is going to create a new one
$ kubectl delete pod phabricator-mariadb-0
Finally, you should see the lines below in MariaDB container logs:
$ kubectl logs $(kubectl get pods -l app.kubernetes.io/instance=phabricator,app.kubernetes.io/name=mariadb,app.kubernetes.io/component=primary -o jsonpath="{.items[0].metadata.name}")
...
mariadb 12:13:24.98 INFO ==> Using persisted data
mariadb 12:13:25.01 INFO ==> Running mysql_upgrade
...
To 9.0.0
Helm performs a lookup for the object based on its group (apps), version (v1), and kind (Deployment). Also known as its GroupVersionKind, or GVK. Changing the GVK is considered a compatibility breaker from Kubernetes' point of view, so you cannot "upgrade" those objects to the new GVK in-place. Earlier versions of Helm 3 did not perform the lookup correctly which has since been fixed to match the spec.
In https://github.com/helm/charts/pulls/17305 the apiVersion of the deployment resources was updated to apps/v1 in tune with the api's deprecated, resulting in compatibility breakage.
This major version signifies this change.
To 7.0.0
Backwards compatibility is not guaranteed. The following notables changes were included:
- Labels are adapted to follow the Helm charts best practices.
- The parameters
persistence.phabricator.storageClass,persistence.phabricator.accessMode, andpersistence.phabricator.sizeswitch topersistence.storageClass,persistence.accessMode, andpersistence.size. - The way of setting the ingress rules has changed. Instead of using
ingress.pathsandingress.hostsas separate objects, you should now define the rules as objects inside theingress.hostsvalue, for example:
ingress:
hosts:
- name: phabricator.local
path: /
To 3.0.0
Backwards compatibility is not guaranteed unless you modify the labels used on the chart's deployments. Use the workaround below to upgrade from versions previous to 3.0.0. The following example assumes that the release name is opencart:
$ kubectl patch deployment opencart-opencart --type=json -p='[{"op": "remove", "path": "/spec/selector/matchLabels/app"}]'
$ kubectl delete statefulset opencart-mariadb --cascade=false