MetalLB
MetalLB is an open source, rock solid LoadBalancer. It handles the ServiceType: Loadbalancer.
TL;DR
$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/metallb
Introduction
Bitnami charts for Helm are carefully engineered, actively maintained and are the quickest and easiest way to deploy containers on a Kubernetes cluster that are ready to handle production workloads.
This chart bootstraps a MetalLB Controller Controller Deployment and a MetalLB Speaker Daemonset on a Kubernetes cluster using the Helm package manager.
Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters. This Helm chart has been tested on top of Bitnami Kubernetes Production Runtime (BKPR). Deploy BKPR to get automated TLS certificates, logging and monitoring for your applications.
Prerequisites
- Kubernetes 1.12+
- Helm 3.1.0
- Virtual IPs for Layer 2 or Route Reflector for BGP setup.
Installing the Chart
To install the chart with the release name my-release:
$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install my-release bitnami/metallb
These commands deploy metallb on the Kubernetes cluster in the default configuration. The Parameters section lists the parameters that can be configured during installation.
Tip
: List all releases using
helm list
Uninstalling the Chart
To uninstall/delete the my-release helm release:
$ helm uninstall my-release
The command removes all the Kubernetes components associated with the chart and deletes the release.
Parameters
Global parameters
| Name | Description | Value |
|---|---|---|
global.imageRegistry |
Global Docker image registry | "" |
global.imagePullSecrets |
Global Docker registry secret names as an array | [] |
Common parameters
| Name | Description | Value |
|---|---|---|
nameOverride |
String to partially override metallb.fullname include (will maintain the release name) | "" |
fullnameOverride |
String to fully override metallb.fullname template | "" |
commonLabels |
Add labels to all the deployed resources | {} |
commonAnnotations |
Add annotations to all the deployed resources | {} |
extraDeploy |
Array of extra objects to deploy with the release | [] |
MetalLB parameters
| Name | Description | Value |
|---|---|---|
existingConfigMap |
Specify the name of an externally-defined ConfigMap to use as the configuration. This is mutually exclusive with the configInline option. |
"" |
configInline |
Specifies MetalLB's configuration directly, in yaml format. | {} |
rbac.create |
Specifies whether to install and use RBAC rules | true |
psp.create |
Whether to create a PodSecurityPolicy. WARNING: PodSecurityPolicy is deprecated in Kubernetes v1.21 or later, unavailable in v1.25 or later | false |
networkPolicy.enabled |
Enable NetworkPolicy | false |
networkPolicy.ingressNSMatchLabels |
Allow connections from other namespaces | {} |
networkPolicy.ingressNSPodMatchLabels |
For other namespaces match by pod labels and namespace labels | {} |
prometheusRule.enabled |
Prometheus Operator alertmanager alerts are created | false |
Controller parameters
| Name | Description | Value |
|---|---|---|
controller.image.registry |
MetalLB Controller image registry | docker.io |
controller.image.repository |
MetalLB Controller image repository | bitnami/metallb-controller |
controller.image.tag |
MetalLB Controller image tag (immutable tags are recommended) | 0.11.0-debian-10-r27 |
controller.image.pullPolicy |
MetalLB Controller image pull policy | IfNotPresent |
controller.image.pullSecrets |
Specify docker-registry secret names as an array | [] |
controller.hostAliases |
Deployment pod host aliases | [] |
controller.rbac.create |
create specifies whether to install and use RBAC rules. | true |
controller.psp.create |
Whether to create a PodSecurityPolicy. WARNING: PodSecurityPolicy is deprecated in Kubernetes v1.21 or later, unavailable in v1.25 or later | true |
controller.priorityClassName |
Set pod priorityClassName | "" |
controller.resources.limits |
The resources limits for the container | {} |
controller.resources.requests |
The requested resources for the container | {} |
controller.nodeSelector |
Node labels for controller pod assignment | {} |
controller.tolerations |
Tolerations for controller pod assignment | [] |
controller.affinity |
Affinity for controller pod assignment | {} |
controller.podAnnotations |
Controller Pod annotations | {} |
controller.podLabels |
Controller Pod labels | {} |
controller.podAffinityPreset |
Controller Pod affinitypreset. Allowed values: soft, hard | "" |
controller.podAntiAffinityPreset |
Controller Pod anti affinitypreset. Allowed values: soft, hard | soft |
controller.nodeAffinityPreset.type |
Controller Pod Node affinity preset. Allowed values: soft, hard | "" |
controller.nodeAffinityPreset.key |
Controller Pod Node affinity label key to match | "" |
controller.nodeAffinityPreset.values |
Controller Pod Node affinity label values to match | [] |
controller.serviceAccount.create |
Specifies whether a ServiceAccount should be created | true |
controller.serviceAccount.name |
The name of the ServiceAccount to use. If not set and create is true, a name is generated using the fullname template | "" |
controller.securityContext.enabled |
Enable pods' security context | true |
controller.securityContext.runAsNonRoot |
MetalLB Controller must runs as nonRoot. | true |
controller.securityContext.runAsUser |
User ID for the pods. | 1001 |
controller.securityContext.fsGroup |
Group ID for the pods. | 1001 |
controller.securityContext.allowPrivilegeEscalation |
This defines if privilegeEscalation is allowed on that container | false |
controller.securityContext.readOnlyRootFilesystem |
This defines if the container can read the root fs on the host | true |
controller.securityContext.capabilities.drop |
Drop capabilities for the securityContext | [] |
controller.revisionHistoryLimit |
Configure the revisionHistoryLimit of the Controller deployment | 3 |
controller.terminationGracePeriodSeconds |
Configure the grace time period for sig term | 0 |
controller.containerPort.metrics |
Configures the ports the MetalLB Controller listens on for metrics | 7472 |
controller.livenessProbe.enabled |
Enable livenessProbe | true |
controller.livenessProbe.initialDelaySeconds |
Initial delay seconds for livenessProbe | 10 |
controller.livenessProbe.periodSeconds |
Period seconds for livenessProbe | 10 |
controller.livenessProbe.timeoutSeconds |
Timeout seconds for livenessProbe | 1 |
controller.livenessProbe.failureThreshold |
Failure threshold for livenessProbe | 3 |
controller.livenessProbe.successThreshold |
Success threshold for livenessProbe | 1 |
controller.readinessProbe.enabled |
Enable readinessProbe | true |
controller.readinessProbe.initialDelaySeconds |
Initial delay seconds for readinessProbe | 10 |
controller.readinessProbe.periodSeconds |
Period seconds for readinessProbe | 10 |
controller.readinessProbe.timeoutSeconds |
Timeout seconds for readinessProbe | 1 |
controller.readinessProbe.failureThreshold |
Failure threshold for readinessProbe | 3 |
controller.readinessProbe.successThreshold |
Success threshold for readinessProbe | 1 |
controller.prometheus.serviceMonitor.enabled |
Specify if a servicemonitor will be deployed for prometheus-operator | false |
controller.prometheus.serviceMonitor.jobLabel |
Specify the jobLabel to use for the prometheus-operator | app.kubernetes.io/name |
controller.prometheus.serviceMonitor.interval |
Scrape interval. If not set, the Prometheus default scrape interval is used | "" |
controller.prometheus.serviceMonitor.metricRelabelings |
Specify additional relabeling of metrics | [] |
controller.prometheus.serviceMonitor.relabelings |
Specify general relabeling | [] |
Speaker parameters
| Name | Description | Value |
|---|---|---|
speaker.image.registry |
MetalLB Speaker image registry | docker.io |
speaker.image.repository |
MetalLB Speaker image repository | bitnami/metallb-speaker |
speaker.image.tag |
MetalLB Speaker image tag (immutable tags are recommended) | 0.11.0-debian-10-r27 |
speaker.image.pullPolicy |
MetalLB Speaker image pull policy | IfNotPresent |
speaker.image.pullSecrets |
Specify docker-registry secret names as an array | [] |
speaker.rbac.create |
create specifies whether to install and use RBAC rules. | true |
speaker.hostAliases |
Deployment pod host aliases | [] |
speaker.psp.create |
Whether to create a PodSecurityPolicy. WARNING: PodSecurityPolicy is deprecated in Kubernetes v1.21 or later, unavailable in v1.25 or later | true |
speaker.priorityClassName |
Set pod priorityClassName. | "" |
speaker.resources.limits |
The resources limits for the container | {} |
speaker.resources.requests |
The requested resources for the container | {} |
speaker.nodeSelector |
Node labels for speaker pod assignment | {} |
speaker.tolerations |
Tolerations for speaker pod assignment | [] |
speaker.affinity |
Affinity for speaker pod assignment | {} |
speaker.podAnnotations |
Speaker Pod annotations | {} |
speaker.podLabels |
Speaker Pod labels | {} |
speaker.serviceAccount.create |
Specifies whether a ServiceAccount should be created | true |
speaker.serviceAccount.name |
The name of the ServiceAccount to use. If not set and create is true, a name is generated using the fullname template | "" |
speaker.daemonset.terminationGracePeriodSeconds |
Configure the grace time period for sig term | 2 |
speaker.daemonset.hostPorts.metrics |
HTTP Metrics Endpoint | 7472 |
speaker.secretName |
References a Secret name for the member secret outside of the helm chart | "" |
speaker.secretKey |
References a Secret key the member secret outside of the helm chart | "" |
speaker.secretValue |
Custom value for speaker.secretKey |
"" |
speaker.initContainers |
Extra initContainers to add to the daemonset | [] |
speaker.securityContext.enabled |
Enable pods' security context | true |
speaker.securityContext.runAsUser |
User ID for the pods. | 0 |
speaker.securityContext.allowPrivilegeEscalation |
Enables privilege Escalation context for the pod. | false |
speaker.securityContext.readOnlyRootFilesystem |
Allows the pod to mount the RootFS as ReadOnly | true |
speaker.securityContext.capabilities.drop |
Drop capabilities for the securityContext | [] |
speaker.securityContext.capabilities.add |
Add capabilities for the securityContext | [] |
speaker.extraEnvVars |
Extra environment variable to pass to the running container. | [] |
speaker.livenessProbe.enabled |
Enable livenessProbe | true |
speaker.livenessProbe.initialDelaySeconds |
Initial delay seconds for livenessProbe | 10 |
speaker.livenessProbe.periodSeconds |
Period seconds for livenessProbe | 10 |
speaker.livenessProbe.timeoutSeconds |
Timeout seconds for livenessProbe | 1 |
speaker.livenessProbe.failureThreshold |
Failure threshold for livenessProbe | 3 |
speaker.livenessProbe.successThreshold |
Success threshold for livenessProbe | 1 |
speaker.readinessProbe.enabled |
Enable readinessProbe | true |
speaker.readinessProbe.initialDelaySeconds |
Initial delay seconds for readinessProbe | 10 |
speaker.readinessProbe.periodSeconds |
Period seconds for readinessProbe | 10 |
speaker.readinessProbe.timeoutSeconds |
Timeout seconds for readinessProbe | 1 |
speaker.readinessProbe.failureThreshold |
Failure threshold for readinessProbe | 3 |
speaker.readinessProbe.successThreshold |
Success threshold for readinessProbe | 1 |
speaker.prometheus.serviceMonitor.enabled |
Enable support for Prometheus Operator | false |
speaker.prometheus.serviceMonitor.jobLabel |
Job label for scrape target | app.kubernetes.io/name |
speaker.prometheus.serviceMonitor.interval |
Scrape interval. If not set, the Prometheus default scrape interval is used | "" |
speaker.prometheus.serviceMonitor.metricRelabelings |
Specify additional relabeling of metrics | [] |
speaker.prometheus.serviceMonitor.relabelings |
Specify general relabeling | [] |
Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,
$ helm install my-release \
--set readinessProbe.successThreshold=5 \
bitnami/metallb
The above command sets the readinessProbe.successThreshold to 5.
Configuration and installation details
Rolling VS Immutable tags
It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.
Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.
To configure MetalLB please look into the configuration section MetalLB Configuration.
Example Layer 2 configuration
configInline:
# The address-pools section lists the IP addresses that MetalLB is
# allowed to allocate, along with settings for how to advertise
# those addresses over BGP once assigned. You can have as many
# address pools as you want.
address-pools:
- # A name for the address pool. Services can request allocation
# from a specific address pool using this name, by listing this
# name under the 'metallb.universe.tf/address-pool' annotation.
name: generic-cluster-pool
# Protocol can be used to select how the announcement is done.
# Supported values are bgp and layer2.
protocol: layer2
# A list of IP address ranges over which MetalLB has
# authority. You can list multiple ranges in a single pool, they
# will all share the same settings. Each range can be either a
# CIDR prefix, or an explicit start-end range of IPs.
addresses:
- 10.27.50.30-10.27.50.35
Troubleshooting
Find more information about how to deal with common errors related to Bitnami’s Helm charts in this troubleshooting guide.
Upgrading
To 2.0.0
What changes were introduced in this major version?
- The
.Values.prometheussection was moved into the components.Values.controller.prometheusand.Values.speaker.prometheus - The
prometheus.prometheusRulewhich is used to toggle the deployment of the metallb alerts is moved under the root of the.Values.prometheusRule - A globel
.Values.psp.createand.Values.rbac.createwas introduced together with the option of toggeling for each component. (global option overwrites component options)Values.controller.rbac.createandValues.controller.psp.createValues.speaker.rbac.createandValues.speaker.psp.create
Considerations when upgrading to this version
- Check if you used the
prometheussection in you deployment. - If you do so, place the configuration you made into the sections
controller.prometheusandspeaker.prometheus. prometheusRuleshould stay under the root of your values.
To 1.0.0
On November 13, 2020, Helm v2 support was formally finished, this major version is the result of the required changes applied to the Helm Chart to be able to incorporate the different features added in Helm v3 and to be consistent with the Helm project itself regarding the Helm v2 EOL.
What changes were introduced in this major version?
- Previous versions of this Helm Chart use
apiVersion: v1(installable by both Helm 2 and 3), this Helm Chart was updated toapiVersion: v2(installable by Helm 3 only). Here you can find more information about theapiVersionfield. - The different fields present in the Chart.yaml file has been ordered alphabetically in a homogeneous way for all the Bitnami Helm Charts
Considerations when upgrading to this version
- If you want to upgrade to this version from a previous one installed with Helm v3, you shouldn't face any issues
- If you want to upgrade to this version using Helm v2, this scenario is not supported as this version doesn't support Helm v2 anymore
- If you installed the previous version with Helm v2 and wants to upgrade to this version with Helm v3, please refer to the official Helm documentation about migrating from Helm v2 to v3
Useful links
- https://docs.bitnami.com/tutorials/resolve-helm2-helm3-post-migration-issues/
- https://helm.sh/docs/topics/v2_v3_migration/
- https://helm.sh/blog/migrate-from-helm-v2-to-helm-v3/
Community supported solution
Please, note this Helm chart is a community-supported solution. This means that the Bitnami team is not actively working on new features/improvements nor providing support through GitHub Issues for this Helm chart. Any new issue will stay open for 20 days to allow the community to contribute, after 15 days without activity the issue will be marked as stale being closed after 5 days.
The Bitnami team will review any PR that is created, feel free to create a PR if you find any issue or want to implement a new feature.
New versions are not going to be affected. Once a new version is released in the upstream project, the Bitnami container image will be updated to use the latest version.