diff --git a/clusters/cl01tl/manifests/immich/Deployment-immich.yaml b/clusters/cl01tl/manifests/immich/Deployment-immich.yaml index 68b88de1e..349f96443 100644 --- a/clusters/cl01tl/manifests/immich/Deployment-immich.yaml +++ b/clusters/cl01tl/manifests/immich/Deployment-immich.yaml @@ -70,7 +70,7 @@ spec: secretKeyRef: key: password name: immich-postgresql-18-cluster-app - image: ghcr.io/immich-app/immich-server:v2.5.6 + image: ghcr.io/immich-app/immich-server:v2.6.1 imagePullPolicy: IfNotPresent livenessProbe: failureThreshold: 3 diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Alertmanager-kube-prometheus-stack-alertmanager.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Alertmanager-kube-prometheus-stack-alertmanager.yaml index 814172697..98b3424a0 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Alertmanager-kube-prometheus-stack-alertmanager.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Alertmanager-kube-prometheus-stack-alertmanager.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-alertmanager app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-admission.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-admission.yaml index 9c05aba0b..59613f97c 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-admission.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-admission.yaml @@ -9,9 +9,9 @@ metadata: app: kube-prometheus-stack-admission app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-operator.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-operator.yaml index e1a2467b5..7686d6f63 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-operator.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-operator.yaml @@ -5,9 +5,9 @@ metadata: labels: app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app: kube-prometheus-stack-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus-crd-edit.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus-crd-edit.yaml index 5c091babe..1e194ace9 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus-crd-edit.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus-crd-edit.yaml @@ -7,9 +7,9 @@ metadata: rbac.authorization.k8s.io/aggregate-to-admin: "true" app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app: kube-prometheus-stack-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus-crd-view.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus-crd-view.yaml index 64bee3813..cff53966a 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus-crd-view.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus-crd-view.yaml @@ -8,9 +8,9 @@ metadata: rbac.authorization.k8s.io/aggregate-to-view: "true" app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app: kube-prometheus-stack-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus.yaml index 15784b02b..f68a1f17b 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRole-kube-prometheus-stack-prometheus.yaml @@ -6,9 +6,9 @@ metadata: app: kube-prometheus-stack-prometheus app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" rules: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-admission.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-admission.yaml index c89d139c8..fd8aeb8a1 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-admission.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-admission.yaml @@ -9,9 +9,9 @@ metadata: app: kube-prometheus-stack-admission app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-operator.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-operator.yaml index 2ea2e9984..6cc49b277 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-operator.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-operator.yaml @@ -5,9 +5,9 @@ metadata: labels: app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app: kube-prometheus-stack-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-prometheus.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-prometheus.yaml index 22c7aa7ac..b66b3d54d 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-prometheus.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ClusterRoleBinding-kube-prometheus-stack-prometheus.yaml @@ -6,9 +6,9 @@ metadata: app: kube-prometheus-stack-prometheus app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" roleRef: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Deployment-kube-prometheus-stack-operator.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Deployment-kube-prometheus-stack-operator.yaml index a4d9eb8d1..d501e2d9e 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Deployment-kube-prometheus-stack-operator.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Deployment-kube-prometheus-stack-operator.yaml @@ -6,9 +6,9 @@ metadata: labels: app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app: kube-prometheus-stack-operator @@ -26,9 +26,9 @@ spec: labels: app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app: kube-prometheus-stack-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/HTTPRoute-kube-prometheus-stack-alertmanager.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/HTTPRoute-kube-prometheus-stack-alertmanager.yaml index 945c2dfdc..36e178121 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/HTTPRoute-kube-prometheus-stack-alertmanager.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/HTTPRoute-kube-prometheus-stack-alertmanager.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-alertmanager app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/HTTPRoute-kube-prometheus-stack-prometheus.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/HTTPRoute-kube-prometheus-stack-prometheus.yaml index 78d512ed0..a1709f8ce 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/HTTPRoute-kube-prometheus-stack-prometheus.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/HTTPRoute-kube-prometheus-stack-prometheus.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-prometheus app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Job-kube-prometheus-stack-admission-create.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Job-kube-prometheus-stack-admission-create.yaml index 8ca094d45..b38bafb6d 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Job-kube-prometheus-stack-admission-create.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Job-kube-prometheus-stack-admission-create.yaml @@ -12,9 +12,9 @@ metadata: app: kube-prometheus-stack-admission-create app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator @@ -28,9 +28,9 @@ spec: app: kube-prometheus-stack-admission-create app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Job-kube-prometheus-stack-admission-patch.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Job-kube-prometheus-stack-admission-patch.yaml index 745557fe1..4c117136b 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Job-kube-prometheus-stack-admission-patch.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Job-kube-prometheus-stack-admission-patch.yaml @@ -12,9 +12,9 @@ metadata: app: kube-prometheus-stack-admission-patch app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator @@ -28,9 +28,9 @@ spec: app: kube-prometheus-stack-admission-patch app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/MutatingWebhookConfiguration-kube-prometheus-stack-admission.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/MutatingWebhookConfiguration-kube-prometheus-stack-admission.yaml index 8fe9fb47f..61807829d 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/MutatingWebhookConfiguration-kube-prometheus-stack-admission.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/MutatingWebhookConfiguration-kube-prometheus-stack-admission.yaml @@ -8,9 +8,9 @@ metadata: app: kube-prometheus-stack-admission app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Prometheus-kube-prometheus-stack-prometheus.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Prometheus-kube-prometheus-stack-prometheus.yaml index 4f7e05b4e..33aa4a71b 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Prometheus-kube-prometheus-stack-prometheus.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Prometheus-kube-prometheus-stack-prometheus.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-prometheus app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-alertmanager.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-alertmanager.rules.yaml index c50bcce04..636a145d1 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-alertmanager.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-alertmanager.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-config-reloaders.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-config-reloaders.yaml index 9a2a68f2e..5098b0639 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-config-reloaders.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-config-reloaders.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-etcd.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-etcd.yaml index a259f8c4f..fa0fe1da8 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-etcd.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-etcd.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-general.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-general.rules.yaml index d7b0c5960..cb34c5844 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-general.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-general.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-cpu-usage-seconds-tot.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-cpu-usage-seconds-tot.yaml index 6b9055af3..2724444fd 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-cpu-usage-seconds-tot.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-cpu-usage-seconds-tot.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-cache.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-cache.yaml index 9daf12098..cf3eca4a0 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-cache.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-cache.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-rss.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-rss.yaml index 8d800e4e2..9e8922614 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-rss.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-rss.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-swap.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-swap.yaml index be564cbbd..140bc5962 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-swap.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-swap.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-working-set-by.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-working-set-by.yaml index 8da02a12a..c57a232ec 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-working-set-by.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-memory-working-set-by.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-resource.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-resource.yaml index ecbb3ed3d..8724537fd 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-resource.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.container-resource.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.pod-owner.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.pod-owner.yaml index c12a2d835..afaf3a618 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.pod-owner.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-k8s.rules.pod-owner.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-availability.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-availability.rules.yaml index 2c219175d..c66c9adf9 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-availability.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-availability.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-burnrate.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-burnrate.rules.yaml index 712fe62a8..e5af94437 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-burnrate.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-burnrate.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-histogram.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-histogram.rules.yaml index a1d35ebbb..d6449b914 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-histogram.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-histogram.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-slos.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-slos.yaml index 58a3daf11..d38d9e8d9 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-slos.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-apiserver-slos.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-prometheus-general.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-prometheus-general.rules.yaml index e37eb4893..e33600a87 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-prometheus-general.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-prometheus-general.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-prometheus-node-recording.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-prometheus-node-recording.rules.yaml index 91dff7423..67950fb26 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-prometheus-node-recording.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-prometheus-node-recording.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-state-metrics.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-state-metrics.yaml index 3ee40cc20..453bec416 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-state-metrics.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kube-state-metrics.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubelet.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubelet.rules.yaml index 3d7556004..24e47133a 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubelet.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubelet.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-apps.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-apps.yaml index e16391be5..b2b0dd217 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-apps.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-apps.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-resources.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-resources.yaml index 29c18e2e6..d74858b11 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-resources.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-resources.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-storage.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-storage.yaml index 12ee17a3c..74ea5a7b5 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-storage.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-storage.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system-apiserver.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system-apiserver.yaml index 16b2be72f..37b0288bf 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system-apiserver.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system-apiserver.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system-kubelet.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system-kubelet.yaml index d544e3caa..29811311b 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system-kubelet.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system-kubelet.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system.yaml index 5582cd787..cb13b9143 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-kubernetes-system.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-exporter.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-exporter.rules.yaml index 1670854dc..5ea3e50f1 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-exporter.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-exporter.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-exporter.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-exporter.yaml index dd452d7dc..12ed37bb0 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-exporter.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-exporter.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-network.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-network.yaml index d8d5d0206..a63079bd2 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-network.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node-network.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node.rules.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node.rules.yaml index 5850b9fdc..a50460c28 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node.rules.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-node.rules.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-prometheus-operator.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-prometheus-operator.yaml index 07e44daec..8073d0e8e 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-prometheus-operator.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-prometheus-operator.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-prometheus.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-prometheus.yaml index 76ba6ec4a..64c45f358 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-prometheus.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/PrometheusRule-kube-prometheus-stack-prometheus.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Role-kube-prometheus-stack-admission.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Role-kube-prometheus-stack-admission.yaml index 7012ad731..212e9b518 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Role-kube-prometheus-stack-admission.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Role-kube-prometheus-stack-admission.yaml @@ -10,9 +10,9 @@ metadata: app: kube-prometheus-stack-admission app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/RoleBinding-kube-prometheus-stack-admission.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/RoleBinding-kube-prometheus-stack-admission.yaml index b85ec5559..a157f5418 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/RoleBinding-kube-prometheus-stack-admission.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/RoleBinding-kube-prometheus-stack-admission.yaml @@ -10,9 +10,9 @@ metadata: app: kube-prometheus-stack-admission app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Secret-alertmanager-kube-prometheus-stack-alertmanager.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Secret-alertmanager-kube-prometheus-stack-alertmanager.yaml index c157327fc..5cca92d57 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Secret-alertmanager-kube-prometheus-stack-alertmanager.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Secret-alertmanager-kube-prometheus-stack-alertmanager.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-alertmanager app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" data: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-alertmanager.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-alertmanager.yaml index fc7476fcd..20d341b62 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-alertmanager.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-alertmanager.yaml @@ -8,9 +8,9 @@ metadata: self-monitor: "true" app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-coredns.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-coredns.yaml index cf26884e5..81214e45f 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-coredns.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-coredns.yaml @@ -7,9 +7,9 @@ metadata: jobLabel: coredns app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" namespace: kube-system diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-kube-etcd.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-kube-etcd.yaml index 54cff2c85..51b00e8d9 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-kube-etcd.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-kube-etcd.yaml @@ -7,9 +7,9 @@ metadata: jobLabel: kube-etcd app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" namespace: kube-system diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-operator.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-operator.yaml index 6065007eb..59a90265d 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-operator.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-operator.yaml @@ -6,9 +6,9 @@ metadata: labels: app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app: kube-prometheus-stack-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-prometheus.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-prometheus.yaml index 7a3b24a44..00be29ae9 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-prometheus.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/Service-kube-prometheus-stack-prometheus.yaml @@ -8,9 +8,9 @@ metadata: self-monitor: "true" app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-admission.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-admission.yaml index 481d81737..19f4137b6 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-admission.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-admission.yaml @@ -10,9 +10,9 @@ metadata: app: kube-prometheus-stack-admission app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-alertmanager.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-alertmanager.yaml index 052024135..3b4cad1fd 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-alertmanager.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-alertmanager.yaml @@ -9,9 +9,9 @@ metadata: app.kubernetes.io/component: alertmanager app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" automountServiceAccountToken: true diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-operator.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-operator.yaml index b393be580..98874c5f0 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-operator.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-operator.yaml @@ -6,9 +6,9 @@ metadata: labels: app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app: kube-prometheus-stack-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-prometheus.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-prometheus.yaml index 343358f7a..f1ea4aae6 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-prometheus.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceAccount-kube-prometheus-stack-prometheus.yaml @@ -9,9 +9,9 @@ metadata: app.kubernetes.io/component: prometheus app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" automountServiceAccountToken: true diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-alertmanager.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-alertmanager.yaml index 4fef6a45e..132cf78bf 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-alertmanager.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-alertmanager.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-alertmanager app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-apiserver.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-apiserver.yaml index b0e4e7b95..a2be588d6 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-apiserver.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-apiserver.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-apiserver app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-coredns.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-coredns.yaml index b3a5b539f..da0efb464 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-coredns.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-coredns.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-coredns app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-kube-etcd.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-kube-etcd.yaml index d68e067e3..330abe21a 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-kube-etcd.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-kube-etcd.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-kube-etcd app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-kubelet.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-kubelet.yaml index 139575655..d5a644143 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-kubelet.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-kubelet.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-kubelet app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-operator.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-operator.yaml index 6b8438608..8257d035a 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-operator.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-operator.yaml @@ -6,9 +6,9 @@ metadata: labels: app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app: kube-prometheus-stack-operator diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-prometheus.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-prometheus.yaml index 439a2359c..ddd24dc09 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-prometheus.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ServiceMonitor-kube-prometheus-stack-prometheus.yaml @@ -7,9 +7,9 @@ metadata: app: kube-prometheus-stack-prometheus app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" spec: diff --git a/clusters/cl01tl/manifests/kube-prometheus-stack/ValidatingWebhookConfiguration-kube-prometheus-stack-admission.yaml b/clusters/cl01tl/manifests/kube-prometheus-stack/ValidatingWebhookConfiguration-kube-prometheus-stack-admission.yaml index 5aa427c7e..942ab6067 100644 --- a/clusters/cl01tl/manifests/kube-prometheus-stack/ValidatingWebhookConfiguration-kube-prometheus-stack-admission.yaml +++ b/clusters/cl01tl/manifests/kube-prometheus-stack/ValidatingWebhookConfiguration-kube-prometheus-stack-admission.yaml @@ -8,9 +8,9 @@ metadata: app: kube-prometheus-stack-admission app.kubernetes.io/managed-by: Helm app.kubernetes.io/instance: kube-prometheus-stack - app.kubernetes.io/version: "82.11.0" + app.kubernetes.io/version: "82.12.0" app.kubernetes.io/part-of: kube-prometheus-stack - chart: kube-prometheus-stack-82.11.0 + chart: kube-prometheus-stack-82.12.0 release: "kube-prometheus-stack" heritage: "Helm" app.kubernetes.io/name: kube-prometheus-stack-prometheus-operator diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-alertmanagerconfigs.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-alertmanagerconfigs.monitoring.coreos.com.yaml index a6f92ffcb..c7b1949bd 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-alertmanagerconfigs.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-alertmanagerconfigs.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: alertmanagerconfigs.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -997,6 +997,14 @@ spec: This is used for SMTP AUTH when the server requires authentication. minLength: 1 type: string + forceImplicitTLS: + description: |- + forceImplicitTLS defines whether to force use of implicit TLS (direct TLS connection) for better security. + true: force use of implicit TLS (direct TLS connection on any port) + false: force disable implicit TLS (use explicit TLS/STARTTLS if required) + nil (default): auto-detect based on port (465=implicit, other=explicit) for backward compatibility + It requires Alertmanager >= v0.31.0. + type: boolean from: description: |- from defines the sender address for email notifications. @@ -6650,6 +6658,13 @@ spec: linkNames enables automatic linking of channel names and usernames in the message. When true, @channel and @username will be converted to clickable links. type: boolean + messageText: + description: |- + messageText defines text content of the Slack message. + If set, this is sent as the top-level 'text' field in the Slack payload. + It requires Alertmanager >= v0.31.0. + minLength: 1 + type: string mrkdwnIn: description: |- mrkdwnIn defines which fields should be parsed as Slack markdown. diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-alertmanagers.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-alertmanagers.monitoring.coreos.com.yaml index 84a9711f2..b71f2c63a 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-alertmanagers.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-alertmanagers.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: alertmanagers.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -1940,6 +1940,14 @@ spec: authUsername: description: authUsername represents SMTP Auth using CRAM-MD5, LOGIN and PLAIN. If empty, Alertmanager doesn't authenticate to the SMTP server. type: string + forceImplicitTLS: + description: |- + forceImplicitTLS defines whether to force use of implicit TLS (direct TLS connection) for better security. + true: force use of implicit TLS (direct TLS connection on any port) + false: force disable implicit TLS (use explicit TLS/STARTTLS if required) + nil (default): auto-detect based on port (465=implicit, other=explicit) for backward compatibility + It requires Alertmanager >= v0.31.0. + type: boolean from: description: from defines the default SMTP From header field. type: string @@ -2684,14 +2692,20 @@ spec: type: string containers: description: |- - containers allows injecting additional containers. This is meant to - allow adding an authentication proxy to an Alertmanager pod. - Containers described here modify an operator generated container if they - share the same name and modifications are done via a strategic merge - patch. The current container names are: `alertmanager` and - `config-reloader`. Overriding containers is entirely outside the scope - of what the maintainers will support and by doing so, you accept that - this behaviour may break at any time without notice. + containers allows injecting additional containers or modifying operator + generated containers. This can be used to allow adding an authentication + proxy to the Pods or to change the behavior of an operator generated + container. Containers described here modify an operator generated + container if they share the same name and modifications are done via a + strategic merge patch. + + The names of containers managed by the operator are: + * `alertmanager` + * `config-reloader` + * `thanos-sidecar` + + Overriding containers which are managed by the operator require careful + testing, especially when upgrading to a new version of the operator. items: description: A single application container that you want to run within a pod. properties: @@ -3520,7 +3534,9 @@ spec: type: integer type: object resizePolicy: - description: Resources resize policy for the container. + description: |- + Resources resize policy for the container. + This field cannot be set on ephemeral containers. items: description: ContainerResizePolicy represents resource resize policy for the container. properties: @@ -4314,15 +4330,21 @@ spec: type: array initContainers: description: |- - initContainers allows adding initContainers to the pod definition. Those can be used to e.g. - fetch secrets for injection into the Alertmanager configuration from external sources. Any - errors during the execution of an initContainer will lead to a restart of the Pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - InitContainers described here modify an operator - generated init containers if they share the same name and modifications are - done via a strategic merge patch. The current init container name is: - `init-config-reloader`. Overriding init containers is entirely outside the - scope of what the maintainers will support and by doing so, you accept that - this behaviour may break at any time without notice. + initContainers allows injecting initContainers to the Pod definition. Those + can be used to e.g. fetch secrets for injection into the Prometheus + configuration from external sources. Any errors during the execution of + an initContainer will lead to a restart of the Pod. More info: + https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ + InitContainers described here modify an operator generated init + containers if they share the same name and modifications are done via a + strategic merge patch. + + The names of init container name managed by the operator are: + * `init-config-reloader`. + + Overriding init containers which are managed by the operator require + careful testing, especially when upgrading to a new version of the + operator. items: description: A single application container that you want to run within a pod. properties: @@ -5151,7 +5173,9 @@ spec: type: integer type: object resizePolicy: - description: Resources resize policy for the container. + description: |- + Resources resize policy for the container. + This field cannot be set on ephemeral containers. items: description: ContainerResizePolicy represents resource resize policy for the container. properties: @@ -6018,6 +6042,10 @@ spec: and the actual ExternalURL is still true, but the server serves requests under a different route prefix. For example for use with `kubectl proxy`. type: string + schedulerName: + description: schedulerName defines the scheduler to use for Pod scheduling. If not specified, the default scheduler is used. + minLength: 1 + type: string secrets: description: |- secrets is a list of Secrets in the same namespace as the Alertmanager @@ -6443,7 +6471,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -6692,7 +6720,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -6811,7 +6839,7 @@ spec: that it does not recognizes, then it should ignore that update and let other controllers handle it. type: string - description: "allocatedResourceStatuses stores status of resource being resized for the given PVC.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nClaimResourceStatus can be in any of following states:\n\t- ControllerResizeInProgress:\n\t\tState set when resize controller starts resizing the volume in control-plane.\n\t- ControllerResizeFailed:\n\t\tState set when resize has failed in resize controller with a terminal error.\n\t- NodeResizePending:\n\t\tState set when resize controller has finished resizing the volume but further resizing of\n\t\tvolume is needed on the node.\n\t- NodeResizeInProgress:\n\t\tState set when kubelet starts resizing the volume.\n\t- NodeResizeFailed:\n\t\tState set when resizing has failed in kubelet with a terminal error. Transient errors don't set\n\t\tNodeResizeFailed.\nFor example: if expanding a PVC for more capacity - this field can be one of the following states:\n\t- pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeFailed\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizePending\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeFailed\"\nWhen this field is not set, it means that no resize operation is in progress for the given PVC.\n\nA controller that receives PVC update with previously unknown resourceName or ClaimResourceStatus\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC.\n\nThis is an alpha field and requires enabling RecoverVolumeExpansionFailure feature." + description: "allocatedResourceStatuses stores status of resource being resized for the given PVC.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nClaimResourceStatus can be in any of following states:\n\t- ControllerResizeInProgress:\n\t\tState set when resize controller starts resizing the volume in control-plane.\n\t- ControllerResizeFailed:\n\t\tState set when resize has failed in resize controller with a terminal error.\n\t- NodeResizePending:\n\t\tState set when resize controller has finished resizing the volume but further resizing of\n\t\tvolume is needed on the node.\n\t- NodeResizeInProgress:\n\t\tState set when kubelet starts resizing the volume.\n\t- NodeResizeFailed:\n\t\tState set when resizing has failed in kubelet with a terminal error. Transient errors don't set\n\t\tNodeResizeFailed.\nFor example: if expanding a PVC for more capacity - this field can be one of the following states:\n\t- pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeFailed\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizePending\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeFailed\"\nWhen this field is not set, it means that no resize operation is in progress for the given PVC.\n\nA controller that receives PVC update with previously unknown resourceName or ClaimResourceStatus\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC." type: object x-kubernetes-map-type: granular allocatedResources: @@ -6821,7 +6849,7 @@ spec: - type: string pattern: ^(\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))(([KMGTPE]i)|[numkMGTPE]|([eE](\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))))?$ x-kubernetes-int-or-string: true - description: "allocatedResources tracks the resources allocated to a PVC including its capacity.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nCapacity reported here may be larger than the actual capacity when a volume expansion operation\nis requested.\nFor storage quota, the larger value from allocatedResources and PVC.spec.resources is used.\nIf allocatedResources is not set, PVC.spec.resources alone is used for quota calculation.\nIf a volume expansion capacity request is lowered, allocatedResources is only\nlowered if there are no expansion operations in progress and if the actual volume capacity\nis equal or lower than the requested capacity.\n\nA controller that receives PVC update with previously unknown resourceName\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC.\n\nThis is an alpha field and requires enabling RecoverVolumeExpansionFailure feature." + description: "allocatedResources tracks the resources allocated to a PVC including its capacity.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nCapacity reported here may be larger than the actual capacity when a volume expansion operation\nis requested.\nFor storage quota, the larger value from allocatedResources and PVC.spec.resources is used.\nIf allocatedResources is not set, PVC.spec.resources alone is used for quota calculation.\nIf a volume expansion capacity request is lowered, allocatedResources is only\nlowered if there are no expansion operations in progress and if the actual volume capacity\nis equal or lower than the requested capacity.\n\nA controller that receives PVC update with previously unknown resourceName\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC." type: object capacity: additionalProperties: @@ -6936,9 +6964,10 @@ spec: operator: description: |- Operator represents a key's relationship to the value. - Valid operators are Exists and Equal. Defaults to Equal. + Valid operators are Exists, Equal, Lt, and Gt. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category. + Lt and Gt perform numeric comparisons (requires feature gate TaintTolerationComparisonOperators). type: string tolerationSeconds: description: |- @@ -7796,7 +7825,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -8631,6 +8660,24 @@ spec: signerName: description: Kubelet's generated CSRs will be addressed to this signer. type: string + userAnnotations: + additionalProperties: + type: string + description: |- + userAnnotations allow pod authors to pass additional information to + the signer implementation. Kubernetes does not restrict or validate this + metadata in any way. + + These values are copied verbatim into the `spec.unverifiedUserAnnotations` field of + the PodCertificateRequest objects that Kubelet creates. + + Entries are subject to the same validation as object metadata annotations, + with the addition that all keys must be domain-prefixed. No restrictions + are placed on values, except an overall size limitation on the entire field. + + Signers should document the keys and values they support. Signers should + deny requests that contain keys they do not recognize. + type: object required: - keyType - signerName diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-podmonitors.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-podmonitors.monitoring.coreos.com.yaml index 5d0889436..b22f6fed0 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-podmonitors.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-podmonitors.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: podmonitors.monitoring.coreos.com spec: group: monitoring.coreos.com diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-probes.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-probes.monitoring.coreos.com.yaml index a759b702b..55ea78a47 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-probes.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-probes.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: probes.monitoring.coreos.com spec: group: monitoring.coreos.com diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheusagents.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheusagents.monitoring.coreos.com.yaml index d864d2650..4cbeb0257 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheusagents.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheusagents.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: prometheusagents.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -1404,9 +1404,8 @@ spec: * `config-reloader` * `thanos-sidecar` - Overriding containers is entirely outside the scope of what the - maintainers will support and by doing so, you accept that this behaviour - may break at any time without notice. + Overriding containers which are managed by the operator require careful + testing, especially when upgrading to a new version of the operator. items: description: A single application container that you want to run within a pod. properties: @@ -2235,7 +2234,9 @@ spec: type: integer type: object resizePolicy: - description: Resources resize policy for the container. + description: |- + Resources resize policy for the container. + This field cannot be set on ephemeral containers. items: description: ContainerResizePolicy represents resource resize policy for the container. properties: @@ -3248,7 +3249,7 @@ spec: initContainers: description: |- initContainers allows injecting initContainers to the Pod definition. Those - can be used to e.g. fetch secrets for injection into the Prometheus + can be used to e.g. fetch secrets for injection into the Prometheus configuration from external sources. Any errors during the execution of an initContainer will lead to a restart of the Pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ @@ -3259,9 +3260,9 @@ spec: The names of init container name managed by the operator are: * `init-config-reloader`. - Overriding init containers is entirely outside the scope of what the - maintainers will support and by doing so, you accept that this behaviour - may break at any time without notice. + Overriding init containers which are managed by the operator require + careful testing, especially when upgrading to a new version of the + operator. items: description: A single application container that you want to run within a pod. properties: @@ -4090,7 +4091,9 @@ spec: type: integer type: object resizePolicy: - description: Resources resize policy for the container. + description: |- + Resources resize policy for the container. + This field cannot be set on ephemeral containers. items: description: ContainerResizePolicy represents resource resize policy for the container. properties: @@ -6178,8 +6181,11 @@ spec: type: string type: object url: - description: url defines the URL of the endpoint to send samples to. - minLength: 1 + description: |- + url defines the URL of the endpoint to send samples to. + + It must use the HTTP or HTTPS scheme. + pattern: ^(http|https)://.+$ type: string writeRelabelConfigs: description: writeRelabelConfigs defines the list of remote write relabel configurations. @@ -6389,6 +6395,10 @@ spec: If you want to enforce a maximum limit for all scrape objects, refer to enforcedSampleLimit. format: int64 type: integer + schedulerName: + description: schedulerName defines the scheduler to use for Pod scheduling. If not specified, the default scheduler is used. + minLength: 1 + type: string scrapeClasses: description: |- scrapeClasses defines the list of scrape classes to expose to scraping objects such as @@ -7566,7 +7576,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -7815,7 +7825,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -7934,7 +7944,7 @@ spec: that it does not recognizes, then it should ignore that update and let other controllers handle it. type: string - description: "allocatedResourceStatuses stores status of resource being resized for the given PVC.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nClaimResourceStatus can be in any of following states:\n\t- ControllerResizeInProgress:\n\t\tState set when resize controller starts resizing the volume in control-plane.\n\t- ControllerResizeFailed:\n\t\tState set when resize has failed in resize controller with a terminal error.\n\t- NodeResizePending:\n\t\tState set when resize controller has finished resizing the volume but further resizing of\n\t\tvolume is needed on the node.\n\t- NodeResizeInProgress:\n\t\tState set when kubelet starts resizing the volume.\n\t- NodeResizeFailed:\n\t\tState set when resizing has failed in kubelet with a terminal error. Transient errors don't set\n\t\tNodeResizeFailed.\nFor example: if expanding a PVC for more capacity - this field can be one of the following states:\n\t- pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeFailed\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizePending\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeFailed\"\nWhen this field is not set, it means that no resize operation is in progress for the given PVC.\n\nA controller that receives PVC update with previously unknown resourceName or ClaimResourceStatus\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC.\n\nThis is an alpha field and requires enabling RecoverVolumeExpansionFailure feature." + description: "allocatedResourceStatuses stores status of resource being resized for the given PVC.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nClaimResourceStatus can be in any of following states:\n\t- ControllerResizeInProgress:\n\t\tState set when resize controller starts resizing the volume in control-plane.\n\t- ControllerResizeFailed:\n\t\tState set when resize has failed in resize controller with a terminal error.\n\t- NodeResizePending:\n\t\tState set when resize controller has finished resizing the volume but further resizing of\n\t\tvolume is needed on the node.\n\t- NodeResizeInProgress:\n\t\tState set when kubelet starts resizing the volume.\n\t- NodeResizeFailed:\n\t\tState set when resizing has failed in kubelet with a terminal error. Transient errors don't set\n\t\tNodeResizeFailed.\nFor example: if expanding a PVC for more capacity - this field can be one of the following states:\n\t- pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeFailed\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizePending\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeFailed\"\nWhen this field is not set, it means that no resize operation is in progress for the given PVC.\n\nA controller that receives PVC update with previously unknown resourceName or ClaimResourceStatus\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC." type: object x-kubernetes-map-type: granular allocatedResources: @@ -7944,7 +7954,7 @@ spec: - type: string pattern: ^(\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))(([KMGTPE]i)|[numkMGTPE]|([eE](\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))))?$ x-kubernetes-int-or-string: true - description: "allocatedResources tracks the resources allocated to a PVC including its capacity.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nCapacity reported here may be larger than the actual capacity when a volume expansion operation\nis requested.\nFor storage quota, the larger value from allocatedResources and PVC.spec.resources is used.\nIf allocatedResources is not set, PVC.spec.resources alone is used for quota calculation.\nIf a volume expansion capacity request is lowered, allocatedResources is only\nlowered if there are no expansion operations in progress and if the actual volume capacity\nis equal or lower than the requested capacity.\n\nA controller that receives PVC update with previously unknown resourceName\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC.\n\nThis is an alpha field and requires enabling RecoverVolumeExpansionFailure feature." + description: "allocatedResources tracks the resources allocated to a PVC including its capacity.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nCapacity reported here may be larger than the actual capacity when a volume expansion operation\nis requested.\nFor storage quota, the larger value from allocatedResources and PVC.spec.resources is used.\nIf allocatedResources is not set, PVC.spec.resources alone is used for quota calculation.\nIf a volume expansion capacity request is lowered, allocatedResources is only\nlowered if there are no expansion operations in progress and if the actual volume capacity\nis equal or lower than the requested capacity.\n\nA controller that receives PVC update with previously unknown resourceName\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC." type: object capacity: additionalProperties: @@ -8062,9 +8072,10 @@ spec: operator: description: |- Operator represents a key's relationship to the value. - Valid operators are Exists and Equal. Defaults to Equal. + Valid operators are Exists, Equal, Lt, and Gt. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category. + Lt and Gt perform numeric comparisons (requires feature gate TaintTolerationComparisonOperators). type: string tolerationSeconds: description: |- @@ -9160,7 +9171,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -9995,6 +10006,24 @@ spec: signerName: description: Kubelet's generated CSRs will be addressed to this signer. type: string + userAnnotations: + additionalProperties: + type: string + description: |- + userAnnotations allow pod authors to pass additional information to + the signer implementation. Kubernetes does not restrict or validate this + metadata in any way. + + These values are copied verbatim into the `spec.unverifiedUserAnnotations` field of + the PodCertificateRequest objects that Kubelet creates. + + Entries are subject to the same validation as object metadata annotations, + with the addition that all keys must be domain-prefixed. No restrictions + are placed on values, except an overall size limitation on the entire field. + + Signers should document the keys and values they support. Signers should + deny requests that contain keys they do not recognize. + type: object required: - keyType - signerName diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheuses.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheuses.monitoring.coreos.com.yaml index 52354b0ff..b2c5500bd 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheuses.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheuses.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: prometheuses.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -2108,9 +2108,8 @@ spec: * `config-reloader` * `thanos-sidecar` - Overriding containers is entirely outside the scope of what the - maintainers will support and by doing so, you accept that this behaviour - may break at any time without notice. + Overriding containers which are managed by the operator require careful + testing, especially when upgrading to a new version of the operator. items: description: A single application container that you want to run within a pod. properties: @@ -2939,7 +2938,9 @@ spec: type: integer type: object resizePolicy: - description: Resources resize policy for the container. + description: |- + Resources resize policy for the container. + This field cannot be set on ephemeral containers. items: description: ContainerResizePolicy represents resource resize policy for the container. properties: @@ -3994,7 +3995,7 @@ spec: initContainers: description: |- initContainers allows injecting initContainers to the Pod definition. Those - can be used to e.g. fetch secrets for injection into the Prometheus + can be used to e.g. fetch secrets for injection into the Prometheus configuration from external sources. Any errors during the execution of an initContainer will lead to a restart of the Pod. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ @@ -4005,9 +4006,9 @@ spec: The names of init container name managed by the operator are: * `init-config-reloader`. - Overriding init containers is entirely outside the scope of what the - maintainers will support and by doing so, you accept that this behaviour - may break at any time without notice. + Overriding init containers which are managed by the operator require + careful testing, especially when upgrading to a new version of the + operator. items: description: A single application container that you want to run within a pod. properties: @@ -4836,7 +4837,9 @@ spec: type: integer type: object resizePolicy: - description: Resources resize policy for the container. + description: |- + Resources resize policy for the container. + This field cannot be set on ephemeral containers. items: description: ContainerResizePolicy represents resource resize policy for the container. properties: @@ -7656,8 +7659,11 @@ spec: type: string type: object url: - description: url defines the URL of the endpoint to send samples to. - minLength: 1 + description: |- + url defines the URL of the endpoint to send samples to. + + It must use the HTTP or HTTPS scheme. + pattern: ^(http|https)://.+$ type: string writeRelabelConfigs: description: writeRelabelConfigs defines the list of remote write relabel configurations. @@ -8006,6 +8012,10 @@ spec: If you want to enforce a maximum limit for all scrape objects, refer to enforcedSampleLimit. format: int64 type: integer + schedulerName: + description: schedulerName defines the scheduler to use for Pod scheduling. If not specified, the default scheduler is used. + minLength: 1 + type: string scrapeClasses: description: |- scrapeClasses defines the list of scrape classes to expose to scraping objects such as @@ -9219,7 +9229,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -9468,7 +9478,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -9587,7 +9597,7 @@ spec: that it does not recognizes, then it should ignore that update and let other controllers handle it. type: string - description: "allocatedResourceStatuses stores status of resource being resized for the given PVC.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nClaimResourceStatus can be in any of following states:\n\t- ControllerResizeInProgress:\n\t\tState set when resize controller starts resizing the volume in control-plane.\n\t- ControllerResizeFailed:\n\t\tState set when resize has failed in resize controller with a terminal error.\n\t- NodeResizePending:\n\t\tState set when resize controller has finished resizing the volume but further resizing of\n\t\tvolume is needed on the node.\n\t- NodeResizeInProgress:\n\t\tState set when kubelet starts resizing the volume.\n\t- NodeResizeFailed:\n\t\tState set when resizing has failed in kubelet with a terminal error. Transient errors don't set\n\t\tNodeResizeFailed.\nFor example: if expanding a PVC for more capacity - this field can be one of the following states:\n\t- pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeFailed\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizePending\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeFailed\"\nWhen this field is not set, it means that no resize operation is in progress for the given PVC.\n\nA controller that receives PVC update with previously unknown resourceName or ClaimResourceStatus\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC.\n\nThis is an alpha field and requires enabling RecoverVolumeExpansionFailure feature." + description: "allocatedResourceStatuses stores status of resource being resized for the given PVC.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nClaimResourceStatus can be in any of following states:\n\t- ControllerResizeInProgress:\n\t\tState set when resize controller starts resizing the volume in control-plane.\n\t- ControllerResizeFailed:\n\t\tState set when resize has failed in resize controller with a terminal error.\n\t- NodeResizePending:\n\t\tState set when resize controller has finished resizing the volume but further resizing of\n\t\tvolume is needed on the node.\n\t- NodeResizeInProgress:\n\t\tState set when kubelet starts resizing the volume.\n\t- NodeResizeFailed:\n\t\tState set when resizing has failed in kubelet with a terminal error. Transient errors don't set\n\t\tNodeResizeFailed.\nFor example: if expanding a PVC for more capacity - this field can be one of the following states:\n\t- pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeFailed\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizePending\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeFailed\"\nWhen this field is not set, it means that no resize operation is in progress for the given PVC.\n\nA controller that receives PVC update with previously unknown resourceName or ClaimResourceStatus\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC." type: object x-kubernetes-map-type: granular allocatedResources: @@ -9597,7 +9607,7 @@ spec: - type: string pattern: ^(\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))(([KMGTPE]i)|[numkMGTPE]|([eE](\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))))?$ x-kubernetes-int-or-string: true - description: "allocatedResources tracks the resources allocated to a PVC including its capacity.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nCapacity reported here may be larger than the actual capacity when a volume expansion operation\nis requested.\nFor storage quota, the larger value from allocatedResources and PVC.spec.resources is used.\nIf allocatedResources is not set, PVC.spec.resources alone is used for quota calculation.\nIf a volume expansion capacity request is lowered, allocatedResources is only\nlowered if there are no expansion operations in progress and if the actual volume capacity\nis equal or lower than the requested capacity.\n\nA controller that receives PVC update with previously unknown resourceName\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC.\n\nThis is an alpha field and requires enabling RecoverVolumeExpansionFailure feature." + description: "allocatedResources tracks the resources allocated to a PVC including its capacity.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nCapacity reported here may be larger than the actual capacity when a volume expansion operation\nis requested.\nFor storage quota, the larger value from allocatedResources and PVC.spec.resources is used.\nIf allocatedResources is not set, PVC.spec.resources alone is used for quota calculation.\nIf a volume expansion capacity request is lowered, allocatedResources is only\nlowered if there are no expansion operations in progress and if the actual volume capacity\nis equal or lower than the requested capacity.\n\nA controller that receives PVC update with previously unknown resourceName\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC." type: object capacity: additionalProperties: @@ -9758,7 +9768,7 @@ spec: description: |- grpcServerTlsConfig defines the TLS parameters for the gRPC server providing the StoreAPI. - Note: Currently only the `caFile`, `certFile`, and `keyFile` fields are supported. + Note: Currently only the `minVersion`, `caFile`, `certFile`, and `keyFile` fields are supported. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -10205,9 +10215,10 @@ spec: operator: description: |- Operator represents a key's relationship to the value. - Valid operators are Exists and Equal. Defaults to Equal. + Valid operators are Exists, Equal, Lt, and Gt. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category. + Lt and Gt perform numeric comparisons (requires feature gate TaintTolerationComparisonOperators). type: string tolerationSeconds: description: |- @@ -11303,7 +11314,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -12138,6 +12149,24 @@ spec: signerName: description: Kubelet's generated CSRs will be addressed to this signer. type: string + userAnnotations: + additionalProperties: + type: string + description: |- + userAnnotations allow pod authors to pass additional information to + the signer implementation. Kubernetes does not restrict or validate this + metadata in any way. + + These values are copied verbatim into the `spec.unverifiedUserAnnotations` field of + the PodCertificateRequest objects that Kubelet creates. + + Entries are subject to the same validation as object metadata annotations, + with the addition that all keys must be domain-prefixed. No restrictions + are placed on values, except an overall size limitation on the entire field. + + Signers should document the keys and values they support. Signers should + deny requests that contain keys they do not recognize. + type: object required: - keyType - signerName diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheusrules.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheusrules.monitoring.coreos.com.yaml index 13f0d36c5..4c8263293 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheusrules.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-prometheusrules.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: prometheusrules.monitoring.coreos.com spec: group: monitoring.coreos.com diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-scrapeconfigs.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-scrapeconfigs.monitoring.coreos.com.yaml index 224b34896..8a6d6c34f 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-scrapeconfigs.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-scrapeconfigs.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: scrapeconfigs.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -1544,7 +1544,7 @@ spec: properties: authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the DigitalOcean API. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -1947,7 +1947,7 @@ spec: pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the DigitalOcean API. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -2156,7 +2156,7 @@ spec: properties: authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the Docker daemon. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -2252,6 +2252,7 @@ spec: properties: name: description: name of the Filter. + minLength: 1 type: string values: description: values defines values to filter on. @@ -2273,8 +2274,9 @@ spec: description: followRedirects defines whether HTTP requests follow HTTP 3xx redirects. type: boolean host: - description: host defines the address of the docker daemon + description: host defines the address of the docker daemon. minLength: 1 + pattern: ^[a-zA-Z][a-zA-Z0-9+.-]*://.+$ type: string hostNetworkingHost: description: hostNetworkingHost defines the host to use if the container is in host networking mode. @@ -2649,7 +2651,7 @@ spec: pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the Docker daemon. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -2811,7 +2813,7 @@ spec: properties: authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the Docker Swarm API. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -2913,6 +2915,7 @@ spec: properties: name: description: name of the Filter. + minLength: 1 type: string values: description: values defines values to filter on. @@ -3311,7 +3314,7 @@ spec: - Nodes type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the Docker Swarm daemon. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -3516,6 +3519,7 @@ spec: properties: name: description: name of the Filter. + minLength: 1 type: string values: description: values defines values to filter on. @@ -3634,7 +3638,7 @@ spec: x-kubernetes-map-type: atomic tlsConfig: description: |- - tlsConfig defines the TLS configuration to connect to the Consul API. + tlsConfig defines the TLS configuration to connect to the EC2 API. It requires Prometheus >= v2.41.0 properties: ca: @@ -3807,7 +3811,7 @@ spec: properties: authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the Eureka server. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -4257,11 +4261,10 @@ spec: type: string server: description: server defines the URL to connect to the Eureka server. - minLength: 1 - pattern: ^http(s)?://.+$ + pattern: ^https?://.+$ type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the Eureka server. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -4522,7 +4525,7 @@ spec: properties: authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the Hetzner API. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -4991,7 +4994,7 @@ spec: - Robot type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the Hetzner API. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -5768,8 +5771,7 @@ spec: type: object url: description: url defines the URL from which the targets are fetched. - minLength: 1 - pattern: ^http(s)?://.+$ + pattern: ^https?://.+$ type: string required: - url @@ -5784,7 +5786,7 @@ spec: properties: authorization: description: |- - authorization defines the header configuration to authenticate against the IONOS. + authorization defines the header configuration to authenticate against the IONOS API. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -6191,7 +6193,7 @@ spec: pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the IONOS API. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -7068,7 +7070,7 @@ spec: properties: authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the Kuma control plane. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -7531,7 +7533,7 @@ spec: pattern: ^https?://.+$ type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the Kuma control plane. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -7733,7 +7735,7 @@ spec: x-kubernetes-map-type: atomic authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the Lightsail API. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -8225,7 +8227,7 @@ spec: type: object x-kubernetes-map-type: atomic tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the Lightsail API. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -8385,7 +8387,7 @@ spec: properties: authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the Linode API. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -8798,7 +8800,7 @@ spec: minLength: 1 type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the Linode API. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -9098,7 +9100,7 @@ spec: type: boolean authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the Nomad API. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -9560,7 +9562,7 @@ spec: description: |- server defines the Nomad server address to connect to for service discovery. This should be the full URL including protocol (e.g., "https://nomad.example.com:4646"). - minLength: 1 + pattern: ^https?://.+$ type: string tagSeparator: description: |- @@ -9568,7 +9570,7 @@ spec: This determines how Nomad service tags are concatenated into Prometheus labels. type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the Nomad API. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -10092,7 +10094,7 @@ spec: description: |- identityEndpoint defines the HTTP endpoint that is required to work with the Identity API of the appropriate version. - pattern: ^http(s)?:\/\/.+$ + pattern: ^https?://.+$ type: string password: description: |- @@ -10472,7 +10474,7 @@ spec: properties: authorization: description: |- - authorization defines the header configuration to authenticate against the DigitalOcean API. + authorization defines the header configuration to authenticate against the PuppetDB API. Cannot be set at the same time as `oauth2`. properties: credentials: @@ -10943,7 +10945,7 @@ spec: pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the PuppetDB server. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -11094,8 +11096,7 @@ spec: type: object url: description: url defines the URL of the PuppetDB root query endpoint. - minLength: 1 - pattern: ^http(s)?://.+$ + pattern: ^https?://.+$ type: string required: - query @@ -11209,7 +11210,7 @@ spec: type: string apiURL: description: apiURL defines the API URL to use when doing the server listing requests. - pattern: ^http(s)?://.+$ + pattern: ^https?://.+$ type: string enableHTTP2: description: enableHTTP2 defines whether to enable HTTP2. @@ -11324,7 +11325,7 @@ spec: type: array x-kubernetes-list-type: set tlsConfig: - description: tlsConfig defines the TLS configuration to connect to the Consul API. + description: tlsConfig defines the TLS configuration to connect to the Scaleway API. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -11561,9 +11562,8 @@ spec: targets: description: targets defines the list of targets for this static configuration. items: - description: |- - Target represents a target for Prometheus to scrape - kubebuilder:validation:MinLength:=1 + description: Target represents a target for Prometheus to scrape + minLength: 1 type: string minItems: 1 type: array diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-servicemonitors.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-servicemonitors.monitoring.coreos.com.yaml index 81c91b8ec..7aa83b819 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-servicemonitors.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-servicemonitors.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: servicemonitors.monitoring.coreos.com spec: group: monitoring.coreos.com diff --git a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-thanosrulers.monitoring.coreos.com.yaml b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-thanosrulers.monitoring.coreos.com.yaml index d95f65566..5bd8248da 100644 --- a/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-thanosrulers.monitoring.coreos.com.yaml +++ b/clusters/cl01tl/manifests/prometheus-operator-crds/CustomResourceDefinition-thanosrulers.monitoring.coreos.com.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.19.0 - operator.prometheus.io/version: 0.89.0 + operator.prometheus.io/version: 0.90.0 name: thanosrulers.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -1075,13 +1075,19 @@ spec: type: array containers: description: |- - containers allows injecting additional containers or modifying operator generated - containers. This can be used to allow adding an authentication proxy to a ThanosRuler pod or - to change the behavior of an operator generated container. Containers described here modify - an operator generated container if they share the same name and modifications are done via a - strategic merge patch. The current container names are: `thanos-ruler` and `config-reloader`. - Overriding containers is entirely outside the scope of what the maintainers will support and by doing - so, you accept that this behaviour may break at any time without notice. + containers allows injecting additional containers or modifying operator + generated containers. This can be used to allow adding an authentication + proxy to the Pods or to change the behavior of an operator generated + container. Containers described here modify an operator generated + container if they share the same name and modifications are done via a + strategic merge patch. + + The names of containers managed by the operator are: + * `thanos-ruler` + * `config-reloader` + + Overriding containers which are managed by the operator require careful + testing, especially when upgrading to a new version of the operator. items: description: A single application container that you want to run within a pod. properties: @@ -1910,7 +1916,9 @@ spec: type: integer type: object resizePolicy: - description: Resources resize policy for the container. + description: |- + Resources resize policy for the container. + This field cannot be set on ephemeral containers. items: description: ContainerResizePolicy represents resource resize policy for the container. properties: @@ -2672,8 +2680,8 @@ spec: description: |- grpcServerTlsConfig defines the gRPC server from which Thanos Querier reads recorded rule data. - Note: Currently only the CAFile, CertFile, and KeyFile fields are supported. - Maps to the '--grpc-server-tls-*' CLI args. + + Note: Currently only the `minVersion`, `caFile`, `certFile`, and `keyFile` fields are supported. properties: ca: description: ca defines the Certificate authority used when verifying server certificates. @@ -2900,13 +2908,11 @@ spec: type: array initContainers: description: |- - initContainers allows adding initContainers to the pod definition. Those can be used to e.g. - fetch secrets for injection into the ThanosRuler configuration from external sources. Any - errors during the execution of an initContainer will lead to a restart of the Pod. - More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ - Using initContainers for any use case other then secret fetching is entirely outside the scope - of what the maintainers will support and by doing so, you accept that this behaviour may break - at any time without notice. + initContainers allows injecting initContainers to the Pod definition. + Those can be used to e.g. fetch secrets for injection into the + configuration from external sources. Any errors during the execution of + an initContainer will lead to a restart of the Pod. More info: + https://kubernetes.io/docs/concepts/workloads/pods/init-containers/ items: description: A single application container that you want to run within a pod. properties: @@ -3735,7 +3741,9 @@ spec: type: integer type: object resizePolicy: - description: Resources resize policy for the container. + description: |- + Resources resize policy for the container. + This field cannot be set on ephemeral containers. items: description: ContainerResizePolicy represents resource resize policy for the container. properties: @@ -5531,8 +5539,11 @@ spec: type: string type: object url: - description: url defines the URL of the endpoint to send samples to. - minLength: 1 + description: |- + url defines the URL of the endpoint to send samples to. + + It must use the HTTP or HTTPS scheme. + pattern: ^(http|https)://.+$ type: string writeRelabelConfigs: description: writeRelabelConfigs defines the list of remote write relabel configurations. @@ -5825,6 +5836,10 @@ spec: type: object type: object x-kubernetes-map-type: atomic + schedulerName: + description: schedulerName defines the scheduler to use for Pod scheduling. If not specified, the default scheduler is used. + minLength: 1 + type: string securityContext: description: |- securityContext defines the pod-level security attributes and common container settings. @@ -6232,7 +6247,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -6481,7 +6496,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -6600,7 +6615,7 @@ spec: that it does not recognizes, then it should ignore that update and let other controllers handle it. type: string - description: "allocatedResourceStatuses stores status of resource being resized for the given PVC.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nClaimResourceStatus can be in any of following states:\n\t- ControllerResizeInProgress:\n\t\tState set when resize controller starts resizing the volume in control-plane.\n\t- ControllerResizeFailed:\n\t\tState set when resize has failed in resize controller with a terminal error.\n\t- NodeResizePending:\n\t\tState set when resize controller has finished resizing the volume but further resizing of\n\t\tvolume is needed on the node.\n\t- NodeResizeInProgress:\n\t\tState set when kubelet starts resizing the volume.\n\t- NodeResizeFailed:\n\t\tState set when resizing has failed in kubelet with a terminal error. Transient errors don't set\n\t\tNodeResizeFailed.\nFor example: if expanding a PVC for more capacity - this field can be one of the following states:\n\t- pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeFailed\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizePending\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeFailed\"\nWhen this field is not set, it means that no resize operation is in progress for the given PVC.\n\nA controller that receives PVC update with previously unknown resourceName or ClaimResourceStatus\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC.\n\nThis is an alpha field and requires enabling RecoverVolumeExpansionFailure feature." + description: "allocatedResourceStatuses stores status of resource being resized for the given PVC.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nClaimResourceStatus can be in any of following states:\n\t- ControllerResizeInProgress:\n\t\tState set when resize controller starts resizing the volume in control-plane.\n\t- ControllerResizeFailed:\n\t\tState set when resize has failed in resize controller with a terminal error.\n\t- NodeResizePending:\n\t\tState set when resize controller has finished resizing the volume but further resizing of\n\t\tvolume is needed on the node.\n\t- NodeResizeInProgress:\n\t\tState set when kubelet starts resizing the volume.\n\t- NodeResizeFailed:\n\t\tState set when resizing has failed in kubelet with a terminal error. Transient errors don't set\n\t\tNodeResizeFailed.\nFor example: if expanding a PVC for more capacity - this field can be one of the following states:\n\t- pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"ControllerResizeFailed\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizePending\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeInProgress\"\n - pvc.status.allocatedResourceStatus['storage'] = \"NodeResizeFailed\"\nWhen this field is not set, it means that no resize operation is in progress for the given PVC.\n\nA controller that receives PVC update with previously unknown resourceName or ClaimResourceStatus\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC." type: object x-kubernetes-map-type: granular allocatedResources: @@ -6610,7 +6625,7 @@ spec: - type: string pattern: ^(\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))(([KMGTPE]i)|[numkMGTPE]|([eE](\+|-)?(([0-9]+(\.[0-9]*)?)|(\.[0-9]+))))?$ x-kubernetes-int-or-string: true - description: "allocatedResources tracks the resources allocated to a PVC including its capacity.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nCapacity reported here may be larger than the actual capacity when a volume expansion operation\nis requested.\nFor storage quota, the larger value from allocatedResources and PVC.spec.resources is used.\nIf allocatedResources is not set, PVC.spec.resources alone is used for quota calculation.\nIf a volume expansion capacity request is lowered, allocatedResources is only\nlowered if there are no expansion operations in progress and if the actual volume capacity\nis equal or lower than the requested capacity.\n\nA controller that receives PVC update with previously unknown resourceName\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC.\n\nThis is an alpha field and requires enabling RecoverVolumeExpansionFailure feature." + description: "allocatedResources tracks the resources allocated to a PVC including its capacity.\nKey names follow standard Kubernetes label syntax. Valid values are either:\n\t* Un-prefixed keys:\n\t\t- storage - the capacity of the volume.\n\t* Custom resources must use implementation-defined prefixed names such as \"example.com/my-custom-resource\"\nApart from above values - keys that are unprefixed or have kubernetes.io prefix are considered\nreserved and hence may not be used.\n\nCapacity reported here may be larger than the actual capacity when a volume expansion operation\nis requested.\nFor storage quota, the larger value from allocatedResources and PVC.spec.resources is used.\nIf allocatedResources is not set, PVC.spec.resources alone is used for quota calculation.\nIf a volume expansion capacity request is lowered, allocatedResources is only\nlowered if there are no expansion operations in progress and if the actual volume capacity\nis equal or lower than the requested capacity.\n\nA controller that receives PVC update with previously unknown resourceName\nshould ignore the update for the purpose it was designed. For example - a controller that\nonly is responsible for resizing capacity of the volume, should ignore PVC updates that change other valid\nresources associated with PVC." type: object capacity: additionalProperties: @@ -6719,9 +6734,10 @@ spec: operator: description: |- Operator represents a key's relationship to the value. - Valid operators are Exists and Equal. Defaults to Equal. + Valid operators are Exists, Equal, Lt, and Gt. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category. + Lt and Gt perform numeric comparisons (requires feature gate TaintTolerationComparisonOperators). type: string tolerationSeconds: description: |- @@ -7623,7 +7639,7 @@ spec: resources: description: |- resources represents the minimum resources the volume should have. - If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements + Users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources @@ -8458,6 +8474,24 @@ spec: signerName: description: Kubelet's generated CSRs will be addressed to this signer. type: string + userAnnotations: + additionalProperties: + type: string + description: |- + userAnnotations allow pod authors to pass additional information to + the signer implementation. Kubernetes does not restrict or validate this + metadata in any way. + + These values are copied verbatim into the `spec.unverifiedUserAnnotations` field of + the PodCertificateRequest objects that Kubelet creates. + + Entries are subject to the same validation as object metadata annotations, + with the addition that all keys must be domain-prefixed. No restrictions + are placed on values, except an overall size limitation on the entire field. + + Signers should document the keys and values they support. Signers should + deny requests that contain keys they do not recognize. + type: object required: - keyType - signerName diff --git a/clusters/cl01tl/manifests/qbittorrent/Deployment-qbittorrent-qui.yaml b/clusters/cl01tl/manifests/qbittorrent/Deployment-qbittorrent-qui.yaml index 6f8184d99..81d6bf353 100644 --- a/clusters/cl01tl/manifests/qbittorrent/Deployment-qbittorrent-qui.yaml +++ b/clusters/cl01tl/manifests/qbittorrent/Deployment-qbittorrent-qui.yaml @@ -59,7 +59,7 @@ spec: value: https://qui.alexlebens.net/api/auth/oidc/callback - name: QUI__OIDC_DISABLE_BUILT_IN_LOGIN value: "false" - image: ghcr.io/autobrr/qui:v1.14.1 + image: ghcr.io/autobrr/qui:v1.15.0 imagePullPolicy: IfNotPresent name: qui resources: