mirror of
https://github.com/bitnami/charts.git
synced 2026-03-06 06:58:50 +08:00
[bitnami/kube-prometheus] Release 11.1.4 (#32759)
* [bitnami/kube-prometheus] Release 11.1.4 updating components versions Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * Update CHANGELOG.md Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * Update Prometheus rules Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> * [bitnami/kube-prometheus] Restore kube-prometheus-crds Signed-off-by: David Gomez <david.gomez@broadcom.com> --------- Signed-off-by: Bitnami Bot <bitnami.bot@broadcom.com> Signed-off-by: David Gomez <david.gomez@broadcom.com> Co-authored-by: David Gomez <david.gomez@broadcom.com>
This commit is contained in:
@@ -1,8 +1,13 @@
|
||||
# Changelog
|
||||
|
||||
## 11.1.3 (2025-03-17)
|
||||
## 11.1.4 (2025-04-02)
|
||||
|
||||
* [bitnami/kube-prometheus] Release 11.1.3 ([#32406](https://github.com/bitnami/charts/pull/32406))
|
||||
* [bitnami/kube-prometheus] Release 11.1.4 ([#32759](https://github.com/bitnami/charts/pull/32759))
|
||||
|
||||
## <small>11.1.3 (2025-03-17)</small>
|
||||
|
||||
* [bitnami/*] Add tanzuCategory annotation (#32409) ([a8fba5c](https://github.com/bitnami/charts/commit/a8fba5cb01f6f4464ca7f69c50b0fbe97d837a95)), closes [#32409](https://github.com/bitnami/charts/issues/32409)
|
||||
* [bitnami/kube-prometheus] Release 11.1.3 (#32406) ([bab0ee2](https://github.com/bitnami/charts/commit/bab0ee20e13d00d7d9a2c153da82abf7ac77bc15)), closes [#32406](https://github.com/bitnami/charts/issues/32406)
|
||||
|
||||
## <small>11.1.2 (2025-03-07)</small>
|
||||
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
dependencies:
|
||||
- name: node-exporter
|
||||
repository: oci://registry-1.docker.io/bitnamicharts
|
||||
version: 4.5.8
|
||||
version: 4.5.9
|
||||
- name: kube-state-metrics
|
||||
repository: oci://registry-1.docker.io/bitnamicharts
|
||||
version: 5.0.4
|
||||
@@ -11,5 +11,5 @@ dependencies:
|
||||
- name: kube-prometheus-crds
|
||||
repository: file://./charts/kube-prometheus-crds
|
||||
version: 0.1.0
|
||||
digest: sha256:2712a333e7677fef4c1ef8a9b9d264daacdaf6fe2c4cca41f86a1842bc14c27d
|
||||
generated: "2025-03-11T16:18:01.155035222Z"
|
||||
digest: sha256:c65a32e0c6b09d31ad61e0fd5e9672c89c93f6bab955a348d5109c782da8a2c2
|
||||
generated: "2025-04-02T01:27:24.670543816Z"
|
||||
|
||||
@@ -7,13 +7,13 @@ annotations:
|
||||
licenses: Apache-2.0
|
||||
images: |
|
||||
- name: alertmanager
|
||||
image: docker.io/bitnami/alertmanager:0.28.1-debian-12-r0
|
||||
image: docker.io/bitnami/alertmanager:0.28.1-debian-12-r2
|
||||
- name: blackbox-exporter
|
||||
image: docker.io/bitnami/blackbox-exporter:0.26.0-debian-12-r1
|
||||
image: docker.io/bitnami/blackbox-exporter:0.26.0-debian-12-r3
|
||||
- name: prometheus
|
||||
image: docker.io/bitnami/prometheus:2.55.1-debian-12-r14
|
||||
image: docker.io/bitnami/prometheus:2.55.1-debian-12-r16
|
||||
- name: prometheus-operator
|
||||
image: docker.io/bitnami/prometheus-operator:0.81.0-debian-12-r0
|
||||
image: docker.io/bitnami/prometheus-operator:0.81.0-debian-12-r2
|
||||
- name: thanos
|
||||
image: docker.io/bitnami/thanos:0.37.2-debian-12-r8
|
||||
apiVersion: v2
|
||||
@@ -50,4 +50,4 @@ maintainers:
|
||||
name: kube-prometheus
|
||||
sources:
|
||||
- https://github.com/bitnami/charts/tree/main/bitnami/kube-prometheus
|
||||
version: 11.1.3
|
||||
version: 11.1.4
|
||||
|
||||
@@ -50,9 +50,9 @@ spec:
|
||||
summary: An Alertmanager instance failed to send notifications.
|
||||
expr: |-
|
||||
(
|
||||
rate(alertmanager_notifications_failed_total{job="foo-alertmanager",namespace="default"}[5m])
|
||||
rate(alertmanager_notifications_failed_total{job="foo-alertmanager",namespace="default"}[15m])
|
||||
/
|
||||
ignoring (reason) group_left rate(alertmanager_notifications_total{job="foo-alertmanager",namespace="default"}[5m])
|
||||
ignoring (reason) group_left rate(alertmanager_notifications_total{job="foo-alertmanager",namespace="default"}[15m])
|
||||
)
|
||||
> 0.01
|
||||
for: 5m
|
||||
@@ -65,9 +65,9 @@ spec:
|
||||
summary: All Alertmanager instances in a cluster failed to send notifications to a critical integration.
|
||||
expr: |-
|
||||
min by (namespace,service, integration) (
|
||||
rate(alertmanager_notifications_failed_total{job="foo-alertmanager",namespace="default", integration=~`.*`}[5m])
|
||||
rate(alertmanager_notifications_failed_total{job="foo-alertmanager",namespace="default", integration=~`.*`}[15m])
|
||||
/
|
||||
ignoring (reason) group_left rate(alertmanager_notifications_total{job="foo-alertmanager",namespace="default", integration=~`.*`}[5m])
|
||||
ignoring (reason) group_left rate(alertmanager_notifications_total{job="foo-alertmanager",namespace="default", integration=~`.*`}[15m])
|
||||
)
|
||||
> 0.01
|
||||
for: 5m
|
||||
@@ -80,9 +80,9 @@ spec:
|
||||
summary: All Alertmanager instances in a cluster failed to send notifications to a non-critical integration.
|
||||
expr: |-
|
||||
min by (namespace,service, integration) (
|
||||
rate(alertmanager_notifications_failed_total{job="foo-alertmanager",namespace="default", integration!~`.*`}[5m])
|
||||
rate(alertmanager_notifications_failed_total{job="foo-alertmanager",namespace="default", integration!~`.*`}[15m])
|
||||
/
|
||||
ignoring (reason) group_left rate(alertmanager_notifications_total{job="foo-alertmanager",namespace="default", integration!~`.*`}[5m])
|
||||
ignoring (reason) group_left rate(alertmanager_notifications_total{job="foo-alertmanager",namespace="default", integration!~`.*`}[15m])
|
||||
)
|
||||
> 0.01
|
||||
for: 5m
|
||||
|
||||
@@ -113,7 +113,7 @@ spec:
|
||||
summary: Processes experience elevated CPU throttling.
|
||||
expr: |-
|
||||
sum(increase(container_cpu_cfs_throttled_periods_total{container!="", job="kubelet", metrics_path="/metrics/cadvisor", }[5m])) without (id, metrics_path, name, image, endpoint, job, node)
|
||||
/
|
||||
/ on (cluster, namespace, pod, container, instance) group_left
|
||||
sum(increase(container_cpu_cfs_periods_total{job="kubelet", metrics_path="/metrics/cadvisor", }[5m])) without (id, metrics_path, name, image, endpoint, job, node)
|
||||
> ( 25 / 100 )
|
||||
for: 15m
|
||||
|
||||
@@ -29,6 +29,18 @@ spec:
|
||||
for: 15m
|
||||
labels:
|
||||
severity: warning
|
||||
- alert: KubeNodePressure
|
||||
annotations:
|
||||
description: '{{`{{`}} $labels.node {{`}}`}} on cluster {{`{{`}} $labels.cluster {{`}}`}} has active Condition {{`{{`}} $labels.condition {{`}}`}}. This is caused by resource usage exceeding eviction thresholds.'
|
||||
runbook_url: https://runbooks.prometheus-operator.dev/runbooks/kubernetes/kubenodepressure
|
||||
summary: Node has as active Condition.
|
||||
expr: |-
|
||||
kube_node_status_condition{job="kube-state-metrics",condition=~"(MemoryPressure|DiskPressure|PIDPressure)",status="true"} == 1
|
||||
and on (cluster, node)
|
||||
kube_node_spec_unschedulable{job="kube-state-metrics"} == 0
|
||||
for: 10m
|
||||
labels:
|
||||
severity: info
|
||||
- alert: KubeNodeUnreachable
|
||||
annotations:
|
||||
description: '{{`{{`}} $labels.node {{`}}`}} is unreachable and some workloads may be rescheduled on cluster {{`{{`}} $labels.cluster {{`}}`}}.'
|
||||
@@ -72,6 +84,21 @@ spec:
|
||||
for: 15m
|
||||
labels:
|
||||
severity: warning
|
||||
- alert: KubeNodeEviction
|
||||
annotations:
|
||||
description: Node {{`{{`}} $labels.node {{`}}`}} on {{`{{`}} $labels.cluster {{`}}`}} is evicting Pods due to {{`{{`}} $labels.eviction_signal {{`}}`}}. Eviction occurs when eviction thresholds are crossed, typically caused by Pods exceeding RAM/ephemeral-storage limits.
|
||||
runbook_url: https://runbooks.prometheus-operator.dev/runbooks/kubernetes/kubenodeeviction
|
||||
summary: Node is evicting pods.
|
||||
expr: |-
|
||||
sum(rate(kubelet_evictions{job="kubelet", metrics_path="/metrics"}[15m])) by (cluster, eviction_signal, instance)
|
||||
* on (cluster, instance) group_left(node)
|
||||
max by (cluster, instance, node) (
|
||||
kubelet_node_name{job="kubelet", metrics_path="/metrics"}
|
||||
)
|
||||
> 0
|
||||
for: 0s
|
||||
labels:
|
||||
severity: info
|
||||
- alert: KubeletPlegDurationHigh
|
||||
annotations:
|
||||
description: The Kubelet Pod Lifecycle Event Generator has a 99th percentile duration of {{`{{`}} $value {{`}}`}} seconds on node {{`{{`}} $labels.node {{`}}`}} on cluster {{`{{`}} $labels.cluster {{`}}`}}.
|
||||
|
||||
@@ -78,7 +78,7 @@ operator:
|
||||
image:
|
||||
registry: docker.io
|
||||
repository: bitnami/prometheus-operator
|
||||
tag: 0.81.0-debian-12-r0
|
||||
tag: 0.81.0-debian-12-r2
|
||||
digest: ""
|
||||
## Specify a imagePullPolicy
|
||||
## ref: https://kubernetes.io/docs/concepts/containers/images/#pre-pulled-images
|
||||
@@ -638,7 +638,7 @@ prometheus:
|
||||
image:
|
||||
registry: docker.io
|
||||
repository: bitnami/prometheus
|
||||
tag: 2.55.1-debian-12-r14
|
||||
tag: 2.55.1-debian-12-r16
|
||||
digest: ""
|
||||
pullPolicy: IfNotPresent
|
||||
## Optionally specify an array of imagePullSecrets.
|
||||
@@ -1865,7 +1865,7 @@ alertmanager:
|
||||
image:
|
||||
registry: docker.io
|
||||
repository: bitnami/alertmanager
|
||||
tag: 0.28.1-debian-12-r0
|
||||
tag: 0.28.1-debian-12-r2
|
||||
digest: ""
|
||||
pullPolicy: IfNotPresent
|
||||
## Optionally specify an array of imagePullSecrets.
|
||||
@@ -2554,7 +2554,7 @@ blackboxExporter:
|
||||
image:
|
||||
registry: docker.io
|
||||
repository: bitnami/blackbox-exporter
|
||||
tag: 0.26.0-debian-12-r1
|
||||
tag: 0.26.0-debian-12-r3
|
||||
digest: ""
|
||||
pullPolicy: IfNotPresent
|
||||
## Optionally specify an array of imagePullSecrets.
|
||||
|
||||
Reference in New Issue
Block a user