Files
charts/bitnami/jupyterhub
Rafael Ríos Saavedra 7228f09e64 [bitnami/jupyterhub] Fixes error when using storageclass (#6205)
* [bitnami/jupyterhub] Fixes error when using storageclass

* Updates chart version
2021-04-26 10:08:22 +02:00
..

JupyterHub

JupyterHub is a multi-user version of the notebook designed for companies, classrooms and research labs.

TL;DR

$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/jupyterhub

Introduction

Bitnami charts for Helm are carefully engineered, actively maintained and are the quickest and easiest way to deploy containers on a Kubernetes cluster that are ready to handle production workloads.

This chart bootstraps a JupyterHub Deployment in a Kubernetes cluster using the Helm package manager.

Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters. This Helm chart has been tested on top of Bitnami Kubernetes Production Runtime (BKPR). Deploy BKPR to get automated TLS certificates, logging and monitoring for your applications.

Prerequisites

  • Kubernetes 1.12+
  • Helm 3.1.0
  • PV provisioner support in the underlying infrastructure

Installing the Chart

To install the chart with the release name my-release:

$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/jupyterhub

These commands deploy JupyterHub on the Kubernetes cluster in the default configuration. The Parameters section lists the parameters that can be configured during installation.

Tip

: List all releases using helm list

Uninstalling the Chart

To uninstall/delete the my-release helm release:

$ helm uninstall my-release

The command removes all the Kubernetes components associated with the chart and deletes the release.

Overview of JupyterHub

This chart deploys three basic elements:

  • JupyterHub: Central element of the chart. Manages authentication and is responsible for creating the Jupyter Notebook instances (called Single User instances). As a consequence, the Hub requires special RBAC privileges in order to access the Kubernetes API to create and manage Deployments.
  • Proxy: This is the external endpoint for users. It manages the communication with the Hub and the Single User instances.
  • Image Puller: In order to improve the Single User instance boot time, a DaemonSet object is deployed that pre-pulls all the necessary images to run the Single User Notebooks.

The following diagram shows a deployed release of the chart:

                                                         │
                                                         │
                                                         │
                                                         │
             ┌────────────────┐                          │
             │                │                          │
             │  Image Puller  │◄──────Pull images to─────┘
             │                │         all nodes
             └────────────────┘

    ┌───────────┐           ┌─────────────┐
    │           │           │             │
    │   Proxy   ├──────────►│     Hub     │
    │           │           │             │
    └───────────┘           └─────────────┘

After accessing the hub and creating a Single User instance, the deployment looks as follows:

                                                         │
                                                         │
                                                         │
                                                         │
             ┌────────────────┐                          │
             │                │                          │
             │  Image Puller  │◄──────Pull images to─────┘
             │                │         all nodes
             └────────────────┘

   ┌───────────┐           ┌─────────────┐
   │           │           │             │
   │   Proxy   ├──────────►│     Hub     │
   │           │           │             │
   └────┬──────┘           └───────┬─────┘
        │                          │
        │                          │
        │                          │
        │     ┌─────────────┐      │
        │     │ Single User │      │
        └────►│  Instance   │◄─────┘
              └─────────────┘

For more information, check the official JupyterHub documentation.

Parameters

The following table lists the configurable parameters of the JupyterHub chart and their default values per section/component:

Global parameters

Name Description Value
global.imageRegistry Global Docker image registry nil
global.imagePullSecrets Global Docker registry secret names as an array []

Common parameters

Name Description Value
kubeVersion Override Kubernetes version nil
nameOverride String to partially override common.names.fullname nil
fullnameOverride String to fully override common.names.fullname nil
commonLabels Labels to add to all deployed objects {}
commonAnnotations Annotations to add to all deployed objects {}
extraDeploy Array of extra objects to deploy with the release []

Hub deployment parameters

Name Description Value
hub.image.registry Hub image registry docker.io
hub.image.repository Hub image repository jupyterhub/k8s-hub
hub.image.tag Hub image tag (immutabe tags are recommended) 0.11.1
hub.image.pullPolicy Hub image pull policy IfNotPresent
hub.image.pullSecrets Hub image pull secrets []
hub.startupProbe.enabled Enable startupProbe true
hub.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 10
hub.startupProbe.periodSeconds Period seconds for startupProbe 10
hub.startupProbe.timeoutSeconds Timeout seconds for startupProbe 3
hub.startupProbe.failureThreshold Failure threshold for startupProbe 30
hub.startupProbe.successThreshold Success threshold for startupProbe 1
hub.livenessProbe.enabled Enable livenessProbe true
hub.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 10
hub.livenessProbe.periodSeconds Period seconds for livenessProbe 10
hub.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 3
hub.livenessProbe.failureThreshold Failure threshold for livenessProbe 30
hub.livenessProbe.successThreshold Success threshold for livenessProbe 1
hub.readinessProbe.enabled Enable readinessProbe true
hub.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 10
hub.readinessProbe.periodSeconds Period seconds for readinessProbe 10
hub.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 3
hub.readinessProbe.failureThreshold Failure threshold for readinessProbe 30
hub.readinessProbe.successThreshold Success threshold for readinessProbe 1
hub.baseUrl Hub base URL /
hub.adminUser Hub Dummy authenticator admin user user
hub.password Hub Dummy authenticator password nil
hub.configuration Hub configuration file (to be used by jupyterhub_config.py) {}
hub.containerPort Hub container port 8081
hub.existingConfigmap Configmap with Hub init scripts (replaces the scripts in templates/hub/configmap.yml) nil
hub.existingSecret Secret with hub configuration (replaces the hub.configuration value) and proxy token nil
hub.command Override Hub default command []
hub.args Override Hub default args []
hub.pdb.create Deploy Hub PodDisruptionBudget false
hub.pdb.minAvailable Set minimum available hub instances nil
hub.pdb.maxUnavailable Set maximum available hub instances nil
hub.priorityClassName Hub pod priority class name nil
hub.hostAliases Add deployment host aliases []
hub.resources.limits The resources limits for the Hub container {}
hub.resources.requests The requested resources for the Hub container {}
hub.containerSecurityContext.enabled Enabled Hub containers' Security Context true
hub.containerSecurityContext.runAsUser Set Hub container's Security Context runAsUser 1000
hub.containerSecurityContext.runAsNonRoot Set Hub container's Security Context runAsNonRoot true
hub.podSecurityContext.enabled Enabled Hub pods' Security Context true
hub.podSecurityContext.fsGroup Set Hub pod's Security Context fsGroup 1001
hub.podAffinityPreset Pod affinity preset. Ignored if affinity is set. Allowed values: soft or hard ""
hub.podAntiAffinityPreset Pod anti-affinity preset. Ignored if affinity is set. Allowed values: soft or hard soft
hub.nodeAffinityPreset.type Node affinity preset type. Ignored if affinity is set. Allowed values: soft or hard ""
hub.nodeAffinityPreset.key Node label key to match. Ignored if affinity is set ""
hub.nodeAffinityPreset.values Node label values to match. Ignored if affinity is set []
hub.affinity Affinity for pod assignment {}
hub.nodeSelector Node labels for pod assignment {}
hub.tolerations Tolerations for pod assignment []
hub.podLabels Extra labels for Hub pods {}
hub.podAnnotations Annotations for Hub pods {}
hub.lifecycleHooks Add lifecycle hooks to the Hub deployment {}
hub.customStartupProbe Override default startup probe {}
hub.customLivenessProbe Override default liveness probe {}
hub.customReadinessProbe Override default readiness probe {}
hub.updateStrategy.type Hub deployment update strategy RollingUpdate
hub.extraEnvVars Add extra environment variables to the Hub container []
hub.extraEnvVarsCM Name of existing ConfigMap containing extra env vars nil
hub.extraEnvVarsSecret Name of existing Secret containing extra env vars nil
hub.extraVolumes Optionally specify extra list of additional volumes for Hub pods []
hub.extraVolumeMounts Optionally specify extra list of additional volumeMounts for Hub container(s) []
hub.initContainers Add additional init containers to the Hub pods {}
hub.sidecars Add additional sidecar containers to the Hub pod {}

Hub RBAC parameters

Name Description Value
hub.serviceAccount.create Create Hub service account true
hub.serviceAccount.name Override Hub service account name nil
hub.rbac.create Create RBAC rules for the Hub service account true

Hub Traffic Exposure Parameters

Name Description Value
hub.networkPolicy.enabled Deploy Hub network policies true
hub.networkPolicy.allowInterspaceAccess Allow communication between pods in different namespaces true
hub.networkPolicy.extraIngress Add extra ingress rules to the NetworkPolicy []
hub.networkPolicy.extraEgress Add extra ingress rules to the NetworkPolicy {}
hub.service.type Hub service type ClusterIP
hub.service.port Hub service port 8081
hub.service.loadBalancerIP Hub service LoadBalancer IP nil
hub.service.loadBalancerSourceRanges loadBalancerIP source ranges for the Service []
hub.service.nodePorts.http NodePort for the HTTP endpoint ""
hub.service.externalTrafficPolicy External traffic policy for the service Cluster

Proxy deployment parameters

Name Description Value
proxy.image.registry Proxy image registry docker.io
proxy.image.repository Proxy image repository bitnami/configurable-http-proxy
proxy.image.tag Proxy image tag (immutabe tags are recommended) 4.3.1-debian-10-r0
proxy.image.pullPolicy Proxy image pull policy IfNotPresent
proxy.image.pullSecrets Proxy image pull secrets []
proxy.image.debug Activate verbose output false
proxy.startupProbe.enabled Enable startupProbe true
proxy.startupProbe.initialDelaySeconds Initial delay seconds for startupProbe 10
proxy.startupProbe.periodSeconds Period seconds for startupProbe 10
proxy.startupProbe.timeoutSeconds Timeout seconds for startupProbe 3
proxy.startupProbe.failureThreshold Failure threshold for startupProbe 30
proxy.startupProbe.successThreshold Success threshold for startupProbe 1
proxy.livenessProbe.enabled Enable livenessProbe true
proxy.livenessProbe.initialDelaySeconds Initial delay seconds for livenessProbe 10
proxy.livenessProbe.periodSeconds Period seconds for livenessProbe 10
proxy.livenessProbe.timeoutSeconds Timeout seconds for livenessProbe 3
proxy.livenessProbe.failureThreshold Failure threshold for livenessProbe 30
proxy.livenessProbe.successThreshold Success threshold for livenessProbe 1
proxy.readinessProbe.enabled Enable readinessProbe true
proxy.readinessProbe.initialDelaySeconds Initial delay seconds for readinessProbe 10
proxy.readinessProbe.periodSeconds Period seconds for readinessProbe 10
proxy.readinessProbe.timeoutSeconds Timeout seconds for readinessProbe 3
proxy.readinessProbe.failureThreshold Failure threshold for readinessProbe 30
proxy.readinessProbe.successThreshold Success threshold for readinessProbe 1
proxy.command Override Proxy default command []
proxy.args Override Proxy default args []
proxy.secretToken Proxy secret token (used for communication with the Hub) nil
proxy.hostAliases Add deployment host aliases []
proxy.pdb.create Deploy Proxy PodDisruptionBudget false
proxy.pdb.minAvailable Set minimum available proxy instances nil
proxy.pdb.maxUnavailable Set maximum available proxy instances nil
proxy.containerPort.api Proxy api container port 8001
proxy.containerPort.http Proxy http container port 8000
proxy.priorityClassName Proxy pod priority class name nil
proxy.resources.limits The resources limits for the Proxy container {}
proxy.resources.requests The requested resources for the Proxy container {}
proxy.containerSecurityContext.enabled Enabled Proxy containers' Security Context true
proxy.containerSecurityContext.runAsUser Set Proxy container's Security Context runAsUser 1001
proxy.containerSecurityContext.runAsNonRoot Set Proxy container's Security Context runAsNonRoot true
proxy.podSecurityContext.enabled Enabled Proxy pods' Security Context true
proxy.podSecurityContext.fsGroup Set Proxy pod's Security Context fsGroup 1001
proxy.podAffinityPreset Pod affinity preset. Ignored if affinity is set. Allowed values: soft or hard ""
proxy.podAntiAffinityPreset Pod anti-affinity preset. Ignored if affinity is set. Allowed values: soft or hard soft
proxy.nodeAffinityPreset.type Node affinity preset type. Ignored if affinity is set. Allowed values: soft or hard ""
proxy.nodeAffinityPreset.key Node label key to match. Ignored if affinity is set ""
proxy.nodeAffinityPreset.values Node label values to match. Ignored if affinity is set []
proxy.affinity Affinity for pod assignment {}
proxy.nodeSelector Node labels for pod assignment {}
proxy.tolerations Tolerations for pod assignment []
proxy.podLabels Extra labels for Proxy pods {}
proxy.podAnnotations Annotations for Proxy pods {}
proxy.lifecycleHooks Add lifecycle hooks to the Proxy deployment {}
proxy.customStartupProbe Override default startup probe {}
proxy.customLivenessProbe Override default liveness probe {}
proxy.customReadinessProbe Override default readiness probe {}
proxy.updateStrategy.type Proxy deployment update strategy RollingUpdate
proxy.extraEnvVars Add extra environment variables to the Proxy container []
proxy.extraEnvVarsCM Name of existing ConfigMap containing extra env vars nil
proxy.extraEnvVarsSecret Name of existing Secret containing extra env vars nil
proxy.extraVolumes Optionally specify extra list of additional volumes for Proxy pods []
proxy.extraVolumeMounts Optionally specify extra list of additional volumeMounts for Proxy container(s) []
proxy.initContainers Add additional init containers to the Proxy pods {}
proxy.sidecars Add additional sidecar containers to the Proxy pod {}

Proxy Traffic Exposure Parameters

Name Description Value
proxy.networkPolicy.enabled Deploy Proxy network policies true
proxy.networkPolicy.allowInterspaceAccess Allow communication between pods in different namespaces true
proxy.networkPolicy.extraIngress Add extra ingress rules to the NetworkPolicy {}
proxy.networkPolicy.extraEgress Add extra ingress rules to the NetworkPolicy nil
proxy.service.api.type API service type ClusterIP
proxy.service.api.port API service port 8001
proxy.service.api.loadBalancerIP API service LoadBalancer IP nil
proxy.service.api.loadBalancerSourceRanges loadBalancerIP source ranges for the Service []
proxy.service.api.nodePorts.http NodePort for the HTTP endpoint ""
proxy.service.api.externalTrafficPolicy External traffic policy for the service Cluster
proxy.service.public.type Public service type LoadBalancer
proxy.service.public.port Public service port 80
proxy.service.public.loadBalancerIP Public service LoadBalancer IP nil
proxy.service.public.loadBalancerSourceRanges loadBalancerIP source ranges for the Service []
proxy.service.public.nodePorts.http NodePort for the HTTP endpoint ""
proxy.service.public.externalTrafficPolicy External traffic policy for the service Cluster
proxy.ingress.enabled Enable ingress for the Public service false
proxy.ingress.path Path to the Proxy pod. /
proxy.ingress.pathType Ingress path type ImplementationSpecific
proxy.ingress.certManager Add cert-manager annotations to the Ingress object false
proxy.ingress.hostname Set ingress rule hostname jupyterhub.local
proxy.ingress.annotations Add annotations to the Ingress object {}
proxy.ingress.tls Enable ingress tls configuration for the hostname defined at proxy.ingress.hostname false
proxy.ingress.extraHosts Add extra hosts to the ingress rule []
proxy.ingress.extraTls Add extra tls configuration for additional hostnames []
proxy.ingress.extraPaths Add extra paths to the ingress rule []
proxy.ingress.secrets Add extra secrets for the tls configuration []

Image puller deployment parameters

Name Description Value
imagePuller.enabled Deploy ImagePuller daemonset true
imagePuller.command Override ImagePuller default command []
imagePuller.args Override ImagePuller default args []
imagePuller.hostAliases Add deployment host aliases []
imagePuller.resources.limits The resources limits for the ImagePuller container {}
imagePuller.resources.requests The resources request for the ImagePuller container {}
imagePuller.containerSecurityContext.enabled Enabled ImagePuller containers' Security Context true
imagePuller.containerSecurityContext.runAsUser Set ImagePuller container's Security Context runAsUser 1001
imagePuller.containerSecurityContext.runAsNonRoot Set ImagePuller container's Security Context runAsNonRoot true
imagePuller.podSecurityContext.enabled Enabled ImagePuller pods' Security Context true
imagePuller.podSecurityContext.fsGroup Set ImagePuller pod's Security Context fsGroup 1001
imagePuller.podAffinityPreset Pod affinity preset. Ignored if affinity is set. Allowed values: soft or hard ""
imagePuller.podAntiAffinityPreset Pod anti-affinity preset. Ignored if affinity is set. Allowed values: soft or hard soft
imagePuller.nodeAffinityPreset.type Node affinity preset type. Ignored if affinity is set. Allowed values: soft or hard ""
imagePuller.nodeAffinityPreset.key Node label key to match. Ignored if affinity is set ""
imagePuller.nodeAffinityPreset.values Node label values to match. Ignored if affinity is set []
imagePuller.affinity Affinity for pod assignment {}
imagePuller.nodeSelector Node labels for pod assignment {}
imagePuller.tolerations Tolerations for pod assignment []
imagePuller.podLabels Extra labels for ImagePuller pods {}
imagePuller.podAnnotations Annotations for ImagePuller pods {}
imagePuller.priorityClassName ImagePuller pod priority class name ""
imagePuller.lifecycleHooks Add lifecycle hooks to the ImagePuller deployment {}
imagePuller.customStartupProbe Override default startup probe {}
imagePuller.customLivenessProbe Override default liveness probe {}
imagePuller.customReadinessProbe Override default readiness probe {}
imagePuller.updateStrategy.type ImagePuller deployment update strategy RollingUpdate
imagePuller.extraEnvVars Add extra environment variables to the ImagePuller container []
imagePuller.extraEnvVarsCM Name of existing ConfigMap containing extra env vars nil
imagePuller.extraEnvVarsSecret Name of existing Secret containing extra env vars nil
imagePuller.extraVolumes Optionally specify extra list of additional volumes for ImagePuller pods []
imagePuller.extraVolumeMounts Optionally specify extra list of additional volumeMounts for ImagePuller container(s) []
imagePuller.initContainers Add additional init containers to the ImagePuller pods {}
imagePuller.sidecars Add additional sidecar containers to the ImagePuller pod {}

Singleuser deployment parameters

Name Description Value
singleuser.image.registry Single User image registry docker.io
singleuser.image.repository Single User image repository bitnami/jupyter-base-notebook
singleuser.image.tag Single User image tag (immutabe tags are recommended) 1.3.0-debian-10-r2
singleuser.image.pullPolicy Single User image pull policy IfNotPresent
singleuser.image.pullSecrets Single User image pull secrets []
singleuser.command Override Single User default command []
singleuser.tolerations Tolerations for pod assignment []
singleuser.containerPort Single User container port 8888
singleuser.notebookDir Notebook directory (it will be the same as the PVC volume mount) /opt/bitnami/jupyterhub-singleuser
singleuser.resources.limits The resources limits for the Single User container {}
singleuser.resources.requests The requested resources for the Single User container {}
singleuser.containerSecurityContext.enabled Enabled Single User containers' Security Context true
singleuser.containerSecurityContext.runAsUser Set Single User container's Security Context runAsUser 1001
singleuser.podSecurityContext.enabled Enabled Single User pods' Security Context true
singleuser.podSecurityContext.fsGroup Set Single User pod's Security Context fsGroup 1001
singleuser.nodeSelector Node labels for pod assignment {}
singleuser.podLabels Extra labels for Single User pods {}
singleuser.podAnnotations Annotations for Single User pods {}
singleuser.priorityClassName Single User pod priority class name nil
singleuser.lifecycleHooks Add lifecycle hooks to the Single User deployment {}
singleuser.extraEnvVars Add extra environment variables to the Single User container []
singleuser.extraVolumes Optionally specify extra list of additional volumes for Single User pods []
singleuser.extraVolumeMounts Optionally specify extra list of additional volumeMounts for Single User container(s) []
singleuser.initContainers Add additional init containers to the Single User pods {}
singleuser.sidecars Add additional sidecar containers to the Single User pod {}

Single User RBAC parameters

Name Description Value
singleuser.serviceAccount.create Create Single User service account true
singleuser.serviceAccount.name Override Single User service account name nil

Single User Persistence parameters

Name Description Value
singleuser.persistence.enabled Enable persistent volume creation on Single User instances true
singleuser.persistence.storageClass Persistent Volumes storage class ""
singleuser.persistence.accessModes[0] Persistent Volumes access modes undefined
singleuser.persistence.size Persistent Volumes size 10Gi

Traffic exposure parameters

Name Description Value
singleuser.networkPolicy.enabled Deploy Single User network policies true
singleuser.networkPolicy.allowInterspaceAccess Allow communication between pods in different namespaces true
singleuser.networkPolicy.allowCloudMetadataAccess Allow Single User pods to access Cloud Metada endpoints false
singleuser.networkPolicy.extraIngress Add extra ingress rules to the NetworkPolicy nil
singleuser.networkPolicy.extraEgress Add extra ingress rules to the NetworkPolicy nil

Auxiliary image parameters

Name Description Value
auxiliaryImage.registry Auxiliary image registry docker.io
auxiliaryImage.repository Auxiliary image repository bitnami/bitnami-shell
auxiliaryImage.tag Auxiliary image tag (immutabe tags are recommended) 10
auxiliaryImage.pullPolicy Auxiliary image pull policy Always
auxiliaryImage.pullSecrets Auxiliary image pull secrets []

External Database settings

Name Description Value
externalDatabase.host Host of an external PostgreSQL instance to connect (only if postgresql.enabled=false) nil
externalDatabase.user User of an external PostgreSQL instance to connect (only if postgresql.enabled=false) postgres
externalDatabase.password Password of an external PostgreSQL instance to connect (only if postgresql.enabled=false) ""
externalDatabase.existingSecret Secret containing the password of an external PostgreSQL instance to connect (only if postgresql.enabled=false) ""
externalDatabase.database Database inside an external PostgreSQL to connect (only if postgresql.enabled=false) jupyterhub
externalDatabase.port Port of an external PostgreSQL to connect (only if postgresql.enabled=false) 5432

PostgreSQL subchart settings

Name Description Value
postgresql.enabled Deploy PostgreSQL subchart true
postgresql.nameOverride Override name of the PostgreSQL chart nil
postgresql.existingSecret Existing secret containing the password of the PostgreSQL chart nil
postgresql.postgresqlPassword Password for the postgres user of the PostgreSQL chart (auto-generated if not set) ""
postgresql.postgresqlUsername Username to create when deploying the PostgreSQL chart bn_jupyterhub
postgresql.postgresqlDatabase Database to create when deploying the PostgreSQL chart bitnami_jupyterhub
postgresql.service.port PostgreSQL service port 5432
postgresql.persistence.enabled Use PVCs when deploying the PostgreSQL chart true
postgresql.persistence.existingClaim Use an existing PVC when deploying the PostgreSQL chart nil
postgresql.persistence.storageClass storageClass of the created PVCs nil
postgresql.persistence.accessMode Access mode of the created PVCs ReadWriteOnce
postgresql.persistence.size Size of the created PVCs 8Gi

Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,

$ helm install my-release \
  --set proxy.livenessProbe.successThreshold=5 \
    bitnami/jupyterhub

The above command sets the proxy.livenessProbe.successThreshold to 5.

Alternatively, a YAML file that specifies the values for the parameters can be provided while installing the chart. For example,

$ helm install my-release -f values.yaml bitnami/jupyterhub

Configuration and installation details

Rolling VS Immutable tags

It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.

Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.

Authenticator

The chart configures the Hub DummyAuthenticator by default, with the password set in the hub.password (auto-generated if not set) and user as the administrator user. In order to change the authentication mechanism change the hub.config.JupyterHub section inside the hub.configuration value. The following example sets the NativeAuthenticator authenticator, and configures an admin user called test.

hub:
  configuration: |
    ...
    hub:
      config:
        JupyterHub:
          admin_access: true
          authenticator_class: nativeauthenticator.NativeAuthenticator
          Authenticator:
            admin_users:
              - test
    ...

When deploying, you will need to Sign Up to set the password for the test user. For more information on Authenticators, check the official JupyterHub documentation.

Configuring the Single User instances

As mentioned in the section Overview of JupyterHub, the Hub is responsible for deploying the Single User instances. The configuration of these instances is passed to the Hub instance via the hub.configuration value. In order to make the chart follow Bitnami standards and to ease the generation of this configuration file, the chart has a singleuser section, which is then used for generating the hub.configuration value. The hub.configuration value can be easily overridden via modifying its default value or by providing a secret in the hub.existingSecret value. In that case, all the settings in the singleuser section will be ignored.

All the settings specified in the hub.configuration value are consumed by the jupyter_config.py script available in the templates/hub/configmap.yaml file. This script can be changed by providing a ConfigMap in the hub.existingConfigmap value. Find in the official JupyterHub documentation more examples of the jupyter_config.py script.

Restricting traffic using NetworkPolicies

The Bitnami JupyterHub chart enables NetworkPolicies by default. These restricts the communication between the three main components: the Proxy, the Hub and the Single User instances. There are two elements that were left open on purpose:

  • Ingress access to the Proxy instance HTTP port: by default, it is open to any IP, as it is the entry point to the JupyterHub instance. This behavior can be changed by tweaking the proxy.networkPolicy.extraIngress value.
  • Hub egress access: As the Hub requires access to the Kubernetes API, the Hub can access to any IP by default (depending on the Kubernetes platform, the Service IP ranges can vary and so there is no easy way to detect the Kubernetes API internal IP). This behavior can be changed by tweaking the hub.networkPolicy.extraEgress value.

Sidecars and Init Containers

If you have a need for additional containers to run within the same pod as the JupyterHub container, the Proxy container or the Single User container (e.g. an additional metrics or logging exporter), you can do so via the proxy.sidecars, hub.sidecars or singleuser.sidecars config parameter. Simply define your container according to the Kubernetes container spec.

hub:
  sidecars:
    - name: your-image-name
      image: your-image
      imagePullPolicy: Always
      ports:
        - name: portname
        containerPort: 1234

Similarly, you can add extra init containers using the hub.initContainers, proxy.initContainers and singleuser.initContainers parameters.

hub:
  initContainers:
    - name: your-image-name
      image: your-image
      imagePullPolicy: Always
      ports:
        - name: portname
          containerPort: 1234

Ingress

This chart provides support for ingress resources for the JupyterHub proxy component. If you have an ingress controller installed on your cluster, such as nginx-ingress-controller or contour you can utilize the ingress controller to serve your application.

To enable ingress integration, please set proxy.ingress.enabled to true.

Hosts

Most likely you will only want to have one hostname that maps to this JupyterHub installation. If that's your case, the property proxy.ingress.hostname will set it. However, it is possible to have more than one host. To facilitate this, the proxy.ingress.extraHosts object can be specified as an array. You can also use proxy.ingress.extraTLS to add the TLS configuration for extra hosts.

For each host indicated at proxy.ingress.extraHosts, please indicate a name, path, and any annotations that you may want the ingress controller to know about.

For annotations, please see this document. Not all annotations are supported by all ingress controllers, but this document does a good job of indicating which annotation is supported by many popular ingress controllers.

TLS Secrets

This chart will facilitate the creation of TLS secrets for use with the ingress controller, however, this is not required. There are four common use cases:

  • Helm generates/manages certificate secrets based on the parameters.
  • User generates/manages certificates separately.
  • Helm creates self-signed certificates and generates/manages certificate secrets.
  • An additional tool (like cert-manager) manages the secrets for the application.

In the first two cases, it's needed a certificate and a key. Files are expected in PEM format.

  • If you are going to use Helm to manage the certificates based on the parameters, please copy these values into the certificate and key values for a given proxy.ingress.secrets entry.
  • In case you are going to manage TLS secrets separately, please know that you must create a TLS secret with name INGRESS_HOSTNAME-tls (where INGRESS_HOSTNAME is a placeholder to be replaced with the hostname you set using the proxy.ingress.hostname parameter).
  • To use self-signed certificates created by Helm, set proxy.ingress.tls to true, and proxy.ingress.certManager to false.
  • If your cluster has a cert-manager add-on to automate the management and issuance of TLS certificates, set proxy.ingress.certManager boolean to true to enable the corresponding annotations for cert-manager.

Setting Pod's affinity

This chart allows you to set your custom affinity using the hub.affinity and proxy.affinity parameters. Find more information about Pod's affinity in the kubernetes documentation.

As an alternative, you can use of the preset configurations for pod affinity, pod anti-affinity, and node affinity available at the bitnami/common chart. To do so, set the hub.podAffinityPreset, hub.podAntiAffinityPreset, or hub.nodeAffinityPreset parameters for the hub and proxy.podAffinityPreset, proxy.podAntiAffinityPreset, or ploxy.nodeAffinityPreset parameters for the proxy.

Deploying extra resources

There are cases where you may want to deploy extra objects, such a ConfigMap containing your app's configuration or some extra deployment with a micro service used by your app. For covering this case, the chart allows adding the full specification of other objects using the extraDeploy parameter.

Troubleshooting

Find more information about how to deal with common errors related to Bitnamis Helm charts in this troubleshooting guide.

Upgrading

$ helm upgrade my-release bitnami/jupyterhub