From b6353b9452579b580109576edc853741a598c923 Mon Sep 17 00:00:00 2001 From: simu Date: Thu, 6 Jun 2024 08:03:29 +0000 Subject: [PATCH 1/2] Update from template Template version: main (073a838) --- .cruft.json | 4 +++- renovate.json | 14 +++++++++++++- 2 files changed, 16 insertions(+), 2 deletions(-) diff --git a/.cruft.json b/.cruft.json index 0617c66d..74c9b8e1 100644 --- a/.cruft.json +++ b/.cruft.json @@ -1,6 +1,6 @@ { "template": "https://github.com/projectsyn/commodore-component-template.git", - "commit": "e0fcaa457a78109244747e0879a87e0adaabbd10", + "commit": "073a838986e0d4a0af14095c5983d06faa9cd8fc", "checkout": "main", "context": { "cookiecutter": { @@ -13,6 +13,8 @@ "add_golden": "y", "add_matrix": "y", "add_go_unit": "n", + "automerge_patch": "y", + "automerge_patch_v0": "n", "copyright_holder": "VSHN AG ", "copyright_year": "2021", "github_owner": "appuio", diff --git a/renovate.json b/renovate.json index e174b12e..d07ecc64 100644 --- a/renovate.json +++ b/renovate.json @@ -18,5 +18,17 @@ "labels": [ "dependency" ], - "separateMinorPatch": true + "separateMinorPatch": true, + "packageRules": [ + { + "matchUpdateTypes": ["patch"], + "matchCurrentVersion": "!/^v?0\\./", + "automerge": true, + "platformAutomerge": false, + "labels": [ + "dependency", + "automerge" + ] + } + ] } From 32774eb0acefe135e89d31df17dcaa5cf4458865 Mon Sep 17 00:00:00 2001 From: Simon Gerber Date: Thu, 6 Jun 2024 10:48:33 +0200 Subject: [PATCH 2/2] Update golden test outputs --- .../openshift4-monitoring/prometheus_rules.yaml | 16 ++++++---------- .../openshift4-monitoring/prometheus_rules.yaml | 16 ++++++---------- .../openshift4-monitoring/prometheus_rules.yaml | 16 ++++++---------- .../openshift4-monitoring/prometheus_rules.yaml | 16 ++++++---------- .../openshift4-monitoring/prometheus_rules.yaml | 16 ++++++---------- .../openshift4-monitoring/prometheus_rules.yaml | 16 ++++++---------- .../openshift4-monitoring/prometheus_rules.yaml | 16 ++++++---------- .../openshift4-monitoring/prometheus_rules.yaml | 16 ++++++---------- 8 files changed, 48 insertions(+), 80 deletions(-) diff --git a/tests/golden/capacity-alerts-with-node-labels/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml b/tests/golden/capacity-alerts-with-node-labels/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml index 95929000..32d092fb 100644 --- a/tests/golden/capacity-alerts-with-node-labels/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml +++ b/tests/golden/capacity-alerts-with-node-labels/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml @@ -392,17 +392,13 @@ spec: syn_component: openshift4-monitoring - alert: SYN_HighOverallControlPlaneCPU annotations: - description: Given three control plane nodes, the overall CPU utilization - may only be about 2/3 of all available capacity. This is because if - a single control plane node fails, the remaining two must handle the - load of the cluster in order to be HA. If the cluster is using more - than 2/3 of all capacity, if one control plane node fails, the remaining - two are likely to fail when they take the load. To fix this, increase - the CPU and memory on your control plane nodes. + description: |- + On a multi-node cluster with three control plane nodes, the overall CPU utilization may only be about 2/3 of all available capacity. This is because if a single control plane node fails, the remaining two must handle the load of the cluster in order to be HA. If the cluster is using more than 2/3 of all capacity, if one control plane node fails, the remaining two are likely to fail when they take the load. To fix this, increase the CPU and memory on your control plane nodes. + On a single node OpenShift (SNO) cluster, this alert will also fire if the 2/3 of the CPU cores of the node are in use by any workload. This level of CPU utlization of an SNO cluster is probably not a problem under most circumstances, but high levels of utilization may result in degraded performance. To manage this alert or silence it in case of false positives see the following link: https://docs.openshift.com/container-platform/latest/monitoring/managing-alerts.html runbook_url: https://github.com/openshift/runbooks/blob/master/alerts/cluster-kube-apiserver-operator/ExtremelyHighIndividualControlPlaneCPU.md - summary: CPU utilization across all three control plane nodes is higher - than two control plane nodes can sustain; a single control plane node - outage may cause a cascading failure; increase available CPU. + summary: CPU utilization across all control plane nodes is more than 60% + of the total available CPU. Control plane node outage may cause a cascading + failure; increase available CPU. syn_component: openshift4-monitoring expr: | sum( diff --git a/tests/golden/capacity-alerts/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml b/tests/golden/capacity-alerts/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml index 95929000..32d092fb 100644 --- a/tests/golden/capacity-alerts/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml +++ b/tests/golden/capacity-alerts/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml @@ -392,17 +392,13 @@ spec: syn_component: openshift4-monitoring - alert: SYN_HighOverallControlPlaneCPU annotations: - description: Given three control plane nodes, the overall CPU utilization - may only be about 2/3 of all available capacity. This is because if - a single control plane node fails, the remaining two must handle the - load of the cluster in order to be HA. If the cluster is using more - than 2/3 of all capacity, if one control plane node fails, the remaining - two are likely to fail when they take the load. To fix this, increase - the CPU and memory on your control plane nodes. + description: |- + On a multi-node cluster with three control plane nodes, the overall CPU utilization may only be about 2/3 of all available capacity. This is because if a single control plane node fails, the remaining two must handle the load of the cluster in order to be HA. If the cluster is using more than 2/3 of all capacity, if one control plane node fails, the remaining two are likely to fail when they take the load. To fix this, increase the CPU and memory on your control plane nodes. + On a single node OpenShift (SNO) cluster, this alert will also fire if the 2/3 of the CPU cores of the node are in use by any workload. This level of CPU utlization of an SNO cluster is probably not a problem under most circumstances, but high levels of utilization may result in degraded performance. To manage this alert or silence it in case of false positives see the following link: https://docs.openshift.com/container-platform/latest/monitoring/managing-alerts.html runbook_url: https://github.com/openshift/runbooks/blob/master/alerts/cluster-kube-apiserver-operator/ExtremelyHighIndividualControlPlaneCPU.md - summary: CPU utilization across all three control plane nodes is higher - than two control plane nodes can sustain; a single control plane node - outage may cause a cascading failure; increase available CPU. + summary: CPU utilization across all control plane nodes is more than 60% + of the total available CPU. Control plane node outage may cause a cascading + failure; increase available CPU. syn_component: openshift4-monitoring expr: | sum( diff --git a/tests/golden/ovn-kubernetes/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml b/tests/golden/ovn-kubernetes/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml index b13b9a69..b4d53ec4 100644 --- a/tests/golden/ovn-kubernetes/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml +++ b/tests/golden/ovn-kubernetes/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml @@ -526,17 +526,13 @@ spec: syn_component: openshift4-monitoring - alert: SYN_HighOverallControlPlaneCPU annotations: - description: Given three control plane nodes, the overall CPU utilization - may only be about 2/3 of all available capacity. This is because if - a single control plane node fails, the remaining two must handle the - load of the cluster in order to be HA. If the cluster is using more - than 2/3 of all capacity, if one control plane node fails, the remaining - two are likely to fail when they take the load. To fix this, increase - the CPU and memory on your control plane nodes. + description: |- + On a multi-node cluster with three control plane nodes, the overall CPU utilization may only be about 2/3 of all available capacity. This is because if a single control plane node fails, the remaining two must handle the load of the cluster in order to be HA. If the cluster is using more than 2/3 of all capacity, if one control plane node fails, the remaining two are likely to fail when they take the load. To fix this, increase the CPU and memory on your control plane nodes. + On a single node OpenShift (SNO) cluster, this alert will also fire if the 2/3 of the CPU cores of the node are in use by any workload. This level of CPU utlization of an SNO cluster is probably not a problem under most circumstances, but high levels of utilization may result in degraded performance. To manage this alert or silence it in case of false positives see the following link: https://docs.openshift.com/container-platform/latest/monitoring/managing-alerts.html runbook_url: https://github.com/openshift/runbooks/blob/master/alerts/cluster-kube-apiserver-operator/ExtremelyHighIndividualControlPlaneCPU.md - summary: CPU utilization across all three control plane nodes is higher - than two control plane nodes can sustain; a single control plane node - outage may cause a cascading failure; increase available CPU. + summary: CPU utilization across all control plane nodes is more than 60% + of the total available CPU. Control plane node outage may cause a cascading + failure; increase available CPU. syn_component: openshift4-monitoring expr: | sum( diff --git a/tests/golden/release-4.14/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml b/tests/golden/release-4.14/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml index 95929000..32d092fb 100644 --- a/tests/golden/release-4.14/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml +++ b/tests/golden/release-4.14/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml @@ -392,17 +392,13 @@ spec: syn_component: openshift4-monitoring - alert: SYN_HighOverallControlPlaneCPU annotations: - description: Given three control plane nodes, the overall CPU utilization - may only be about 2/3 of all available capacity. This is because if - a single control plane node fails, the remaining two must handle the - load of the cluster in order to be HA. If the cluster is using more - than 2/3 of all capacity, if one control plane node fails, the remaining - two are likely to fail when they take the load. To fix this, increase - the CPU and memory on your control plane nodes. + description: |- + On a multi-node cluster with three control plane nodes, the overall CPU utilization may only be about 2/3 of all available capacity. This is because if a single control plane node fails, the remaining two must handle the load of the cluster in order to be HA. If the cluster is using more than 2/3 of all capacity, if one control plane node fails, the remaining two are likely to fail when they take the load. To fix this, increase the CPU and memory on your control plane nodes. + On a single node OpenShift (SNO) cluster, this alert will also fire if the 2/3 of the CPU cores of the node are in use by any workload. This level of CPU utlization of an SNO cluster is probably not a problem under most circumstances, but high levels of utilization may result in degraded performance. To manage this alert or silence it in case of false positives see the following link: https://docs.openshift.com/container-platform/latest/monitoring/managing-alerts.html runbook_url: https://github.com/openshift/runbooks/blob/master/alerts/cluster-kube-apiserver-operator/ExtremelyHighIndividualControlPlaneCPU.md - summary: CPU utilization across all three control plane nodes is higher - than two control plane nodes can sustain; a single control plane node - outage may cause a cascading failure; increase available CPU. + summary: CPU utilization across all control plane nodes is more than 60% + of the total available CPU. Control plane node outage may cause a cascading + failure; increase available CPU. syn_component: openshift4-monitoring expr: | sum( diff --git a/tests/golden/remote-write/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml b/tests/golden/remote-write/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml index 95929000..32d092fb 100644 --- a/tests/golden/remote-write/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml +++ b/tests/golden/remote-write/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml @@ -392,17 +392,13 @@ spec: syn_component: openshift4-monitoring - alert: SYN_HighOverallControlPlaneCPU annotations: - description: Given three control plane nodes, the overall CPU utilization - may only be about 2/3 of all available capacity. This is because if - a single control plane node fails, the remaining two must handle the - load of the cluster in order to be HA. If the cluster is using more - than 2/3 of all capacity, if one control plane node fails, the remaining - two are likely to fail when they take the load. To fix this, increase - the CPU and memory on your control plane nodes. + description: |- + On a multi-node cluster with three control plane nodes, the overall CPU utilization may only be about 2/3 of all available capacity. This is because if a single control plane node fails, the remaining two must handle the load of the cluster in order to be HA. If the cluster is using more than 2/3 of all capacity, if one control plane node fails, the remaining two are likely to fail when they take the load. To fix this, increase the CPU and memory on your control plane nodes. + On a single node OpenShift (SNO) cluster, this alert will also fire if the 2/3 of the CPU cores of the node are in use by any workload. This level of CPU utlization of an SNO cluster is probably not a problem under most circumstances, but high levels of utilization may result in degraded performance. To manage this alert or silence it in case of false positives see the following link: https://docs.openshift.com/container-platform/latest/monitoring/managing-alerts.html runbook_url: https://github.com/openshift/runbooks/blob/master/alerts/cluster-kube-apiserver-operator/ExtremelyHighIndividualControlPlaneCPU.md - summary: CPU utilization across all three control plane nodes is higher - than two control plane nodes can sustain; a single control plane node - outage may cause a cascading failure; increase available CPU. + summary: CPU utilization across all control plane nodes is more than 60% + of the total available CPU. Control plane node outage may cause a cascading + failure; increase available CPU. syn_component: openshift4-monitoring expr: | sum( diff --git a/tests/golden/team-routing/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml b/tests/golden/team-routing/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml index f82725a8..82c76e1f 100644 --- a/tests/golden/team-routing/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml +++ b/tests/golden/team-routing/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml @@ -413,17 +413,13 @@ spec: syn_team: clumsy-donkeys - alert: SYN_HighOverallControlPlaneCPU annotations: - description: Given three control plane nodes, the overall CPU utilization - may only be about 2/3 of all available capacity. This is because if - a single control plane node fails, the remaining two must handle the - load of the cluster in order to be HA. If the cluster is using more - than 2/3 of all capacity, if one control plane node fails, the remaining - two are likely to fail when they take the load. To fix this, increase - the CPU and memory on your control plane nodes. + description: |- + On a multi-node cluster with three control plane nodes, the overall CPU utilization may only be about 2/3 of all available capacity. This is because if a single control plane node fails, the remaining two must handle the load of the cluster in order to be HA. If the cluster is using more than 2/3 of all capacity, if one control plane node fails, the remaining two are likely to fail when they take the load. To fix this, increase the CPU and memory on your control plane nodes. + On a single node OpenShift (SNO) cluster, this alert will also fire if the 2/3 of the CPU cores of the node are in use by any workload. This level of CPU utlization of an SNO cluster is probably not a problem under most circumstances, but high levels of utilization may result in degraded performance. To manage this alert or silence it in case of false positives see the following link: https://docs.openshift.com/container-platform/latest/monitoring/managing-alerts.html runbook_url: https://github.com/openshift/runbooks/blob/master/alerts/cluster-kube-apiserver-operator/ExtremelyHighIndividualControlPlaneCPU.md - summary: CPU utilization across all three control plane nodes is higher - than two control plane nodes can sustain; a single control plane node - outage may cause a cascading failure; increase available CPU. + summary: CPU utilization across all control plane nodes is more than 60% + of the total available CPU. Control plane node outage may cause a cascading + failure; increase available CPU. syn_component: openshift4-monitoring expr: | sum( diff --git a/tests/golden/user-workload-monitoring/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml b/tests/golden/user-workload-monitoring/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml index 2b7ff60f..e974c1dc 100644 --- a/tests/golden/user-workload-monitoring/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml +++ b/tests/golden/user-workload-monitoring/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml @@ -392,17 +392,13 @@ spec: syn_component: openshift4-monitoring - alert: SYN_HighOverallControlPlaneCPU annotations: - description: Given three control plane nodes, the overall CPU utilization - may only be about 2/3 of all available capacity. This is because if - a single control plane node fails, the remaining two must handle the - load of the cluster in order to be HA. If the cluster is using more - than 2/3 of all capacity, if one control plane node fails, the remaining - two are likely to fail when they take the load. To fix this, increase - the CPU and memory on your control plane nodes. + description: |- + On a multi-node cluster with three control plane nodes, the overall CPU utilization may only be about 2/3 of all available capacity. This is because if a single control plane node fails, the remaining two must handle the load of the cluster in order to be HA. If the cluster is using more than 2/3 of all capacity, if one control plane node fails, the remaining two are likely to fail when they take the load. To fix this, increase the CPU and memory on your control plane nodes. + On a single node OpenShift (SNO) cluster, this alert will also fire if the 2/3 of the CPU cores of the node are in use by any workload. This level of CPU utlization of an SNO cluster is probably not a problem under most circumstances, but high levels of utilization may result in degraded performance. To manage this alert or silence it in case of false positives see the following link: https://docs.openshift.com/container-platform/latest/monitoring/managing-alerts.html runbook_url: https://github.com/openshift/runbooks/blob/master/alerts/cluster-kube-apiserver-operator/ExtremelyHighIndividualControlPlaneCPU.md - summary: CPU utilization across all three control plane nodes is higher - than two control plane nodes can sustain; a single control plane node - outage may cause a cascading failure; increase available CPU. + summary: CPU utilization across all control plane nodes is more than 60% + of the total available CPU. Control plane node outage may cause a cascading + failure; increase available CPU. syn_component: openshift4-monitoring expr: | sum( diff --git a/tests/golden/vsphere/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml b/tests/golden/vsphere/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml index 8a35c738..65b6699f 100644 --- a/tests/golden/vsphere/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml +++ b/tests/golden/vsphere/openshift4-monitoring/openshift4-monitoring/prometheus_rules.yaml @@ -392,17 +392,13 @@ spec: syn_component: openshift4-monitoring - alert: SYN_HighOverallControlPlaneCPU annotations: - description: Given three control plane nodes, the overall CPU utilization - may only be about 2/3 of all available capacity. This is because if - a single control plane node fails, the remaining two must handle the - load of the cluster in order to be HA. If the cluster is using more - than 2/3 of all capacity, if one control plane node fails, the remaining - two are likely to fail when they take the load. To fix this, increase - the CPU and memory on your control plane nodes. + description: |- + On a multi-node cluster with three control plane nodes, the overall CPU utilization may only be about 2/3 of all available capacity. This is because if a single control plane node fails, the remaining two must handle the load of the cluster in order to be HA. If the cluster is using more than 2/3 of all capacity, if one control plane node fails, the remaining two are likely to fail when they take the load. To fix this, increase the CPU and memory on your control plane nodes. + On a single node OpenShift (SNO) cluster, this alert will also fire if the 2/3 of the CPU cores of the node are in use by any workload. This level of CPU utlization of an SNO cluster is probably not a problem under most circumstances, but high levels of utilization may result in degraded performance. To manage this alert or silence it in case of false positives see the following link: https://docs.openshift.com/container-platform/latest/monitoring/managing-alerts.html runbook_url: https://github.com/openshift/runbooks/blob/master/alerts/cluster-kube-apiserver-operator/ExtremelyHighIndividualControlPlaneCPU.md - summary: CPU utilization across all three control plane nodes is higher - than two control plane nodes can sustain; a single control plane node - outage may cause a cascading failure; increase available CPU. + summary: CPU utilization across all control plane nodes is more than 60% + of the total available CPU. Control plane node outage may cause a cascading + failure; increase available CPU. syn_component: openshift4-monitoring expr: | sum(