Skip to content

Commit 7d3c9d5

Browse files
Merge pull request #2741 from simonpasquier/bump-jsonnet-deps
NO-ISSUE: bump jsonnet dependencies
2 parents 5fe68e1 + 3f13b82 commit 7d3c9d5

21 files changed

+103
-87
lines changed

assets/alertmanager-user-workload/alertmanager.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,7 @@ spec:
2525
namespaces:
2626
- openshift-user-workload-monitoring
2727
topologyKey: kubernetes.io/hostname
28+
alertmanagerConfigSelector: null
2829
automountServiceAccountToken: true
2930
containers:
3031
- env:

assets/alertmanager/alertmanager.yaml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,7 @@ spec:
2525
namespaces:
2626
- openshift-monitoring
2727
topologyKey: kubernetes.io/hostname
28+
alertmanagerConfigSelector: null
2829
automountServiceAccountToken: true
2930
containers:
3031
- env:

assets/control-plane/prometheus-rule.yaml

Lines changed: 36 additions & 19 deletions
Original file line numberDiff line numberDiff line change
@@ -313,10 +313,16 @@ spec:
313313
description: Namespace {{ $labels.namespace }} is using {{ $value | humanizePercentage }} of its {{ $labels.resource }} quota.
314314
summary: Namespace quota is going to be full.
315315
expr: |
316-
kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="used"}
317-
/ ignoring(instance, job, type)
318-
(kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="hard"} > 0)
319-
> 0.9 < 1
316+
max without (instance, job, type) (
317+
kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="used"}
318+
)
319+
/ on (cluster, namespace, resource, resourcequota) group_left()
320+
(
321+
max without (instance, job, type) (
322+
kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="hard"}
323+
) > 0
324+
)
325+
> 0.9 < 1
320326
for: 15m
321327
labels:
322328
severity: info
@@ -325,10 +331,16 @@ spec:
325331
description: Namespace {{ $labels.namespace }} is using {{ $value | humanizePercentage }} of its {{ $labels.resource }} quota.
326332
summary: Namespace quota is fully used.
327333
expr: |
328-
kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="used"}
329-
/ ignoring(instance, job, type)
330-
(kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="hard"} > 0)
331-
== 1
334+
max without (instance, job, type) (
335+
kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="used"}
336+
)
337+
/ on (cluster, namespace, resource, resourcequota) group_left()
338+
(
339+
max without (instance, job, type) (
340+
kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="hard"}
341+
) > 0
342+
)
343+
== 1
332344
for: 15m
333345
labels:
334346
severity: info
@@ -337,10 +349,15 @@ spec:
337349
description: Namespace {{ $labels.namespace }} is using {{ $value | humanizePercentage }} of its {{ $labels.resource }} quota.
338350
summary: Namespace quota has exceeded the limits.
339351
expr: |
340-
kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="used"}
341-
/ ignoring(instance, job, type)
342-
(kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="hard"} > 0)
343-
> 1
352+
max without (instance, job, type) (
353+
kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="used"}
354+
)
355+
/ on (cluster, namespace, resource, resourcequota) group_left()
356+
(
357+
max without (instance, job, type) (
358+
kube_resourcequota{namespace=~"(openshift-.*|kube-.*|default)",job="kube-state-metrics", type="hard"}
359+
) > 0
360+
) > 1
344361
for: 15m
345362
labels:
346363
severity: warning
@@ -495,7 +512,7 @@ spec:
495512
runbook_url: https://github.com/openshift/runbooks/blob/master/alerts/cluster-monitoring-operator/KubeletDown.md
496513
summary: Target disappeared from Prometheus target discovery.
497514
expr: |
498-
absent(up{job="kubelet", metrics_path="/metrics"} == 1)
515+
absent(up{job="kubelet", metrics_path="/metrics"})
499516
for: 15m
500517
labels:
501518
namespace: kube-system
@@ -756,10 +773,10 @@ spec:
756773
- name: kube-scheduler.rules
757774
rules:
758775
- expr: |
759-
histogram_quantile(0.99, sum(rate(scheduler_e2e_scheduling_duration_seconds_bucket{job="scheduler"}[5m])) without(instance, pod))
776+
histogram_quantile(0.99, sum(rate(scheduler_scheduling_attempt_duration_seconds_bucket{job="scheduler"}[5m])) without(instance, pod))
760777
labels:
761778
quantile: "0.99"
762-
record: cluster_quantile:scheduler_e2e_scheduling_duration_seconds:histogram_quantile
779+
record: cluster_quantile:scheduler_scheduling_attempt_duration_seconds:histogram_quantile
763780
- expr: |
764781
histogram_quantile(0.99, sum(rate(scheduler_scheduling_algorithm_duration_seconds_bucket{job="scheduler"}[5m])) without(instance, pod))
765782
labels:
@@ -771,10 +788,10 @@ spec:
771788
quantile: "0.99"
772789
record: cluster_quantile:scheduler_binding_duration_seconds:histogram_quantile
773790
- expr: |
774-
histogram_quantile(0.9, sum(rate(scheduler_e2e_scheduling_duration_seconds_bucket{job="scheduler"}[5m])) without(instance, pod))
791+
histogram_quantile(0.9, sum(rate(scheduler_scheduling_attempt_duration_seconds_bucket{job="scheduler"}[5m])) without(instance, pod))
775792
labels:
776793
quantile: "0.9"
777-
record: cluster_quantile:scheduler_e2e_scheduling_duration_seconds:histogram_quantile
794+
record: cluster_quantile:scheduler_scheduling_attempt_duration_seconds:histogram_quantile
778795
- expr: |
779796
histogram_quantile(0.9, sum(rate(scheduler_scheduling_algorithm_duration_seconds_bucket{job="scheduler"}[5m])) without(instance, pod))
780797
labels:
@@ -786,10 +803,10 @@ spec:
786803
quantile: "0.9"
787804
record: cluster_quantile:scheduler_binding_duration_seconds:histogram_quantile
788805
- expr: |
789-
histogram_quantile(0.5, sum(rate(scheduler_e2e_scheduling_duration_seconds_bucket{job="scheduler"}[5m])) without(instance, pod))
806+
histogram_quantile(0.5, sum(rate(scheduler_scheduling_attempt_duration_seconds_bucket{job="scheduler"}[5m])) without(instance, pod))
790807
labels:
791808
quantile: "0.5"
792-
record: cluster_quantile:scheduler_e2e_scheduling_duration_seconds:histogram_quantile
809+
record: cluster_quantile:scheduler_scheduling_attempt_duration_seconds:histogram_quantile
793810
- expr: |
794811
histogram_quantile(0.5, sum(rate(scheduler_scheduling_algorithm_duration_seconds_bucket{job="scheduler"}[5m])) without(instance, pod))
795812
labels:

assets/prometheus-k8s/prometheus.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ spec:
4242
scheme: https
4343
tlsConfig:
4444
caFile: /etc/prometheus/configmaps/serving-certs-ca-bundle/service-ca.crt
45-
serverName: alertmanager-main
45+
serverName: alertmanager-main.openshift-monitoring.svc
4646
configMaps:
4747
- serving-certs-ca-bundle
4848
- kubelet-serving-ca-bundle

assets/prometheus-operator-user-workload/cluster-role.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -77,7 +77,7 @@ rules:
7777
- list
7878
- watch
7979
- apiGroups:
80-
- ""
80+
- events.k8s.io
8181
resources:
8282
- events
8383
verbs:

assets/prometheus-operator/cluster-role.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -77,7 +77,7 @@ rules:
7777
- list
7878
- watch
7979
- apiGroups:
80-
- ""
80+
- events.k8s.io
8181
resources:
8282
- events
8383
verbs:

jsonnet/components/alertmanager-user-workload.libsonnet

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -195,6 +195,8 @@ function(params)
195195
},
196196
},
197197
spec+: {
198+
// The value of alertmanagerConfigSelector is defined at runtime by the Cluster Monitoring Operator.
199+
alertmanagerConfigSelector: null,
198200
automountServiceAccountToken: true,
199201
securityContext: {
200202
fsGroup: 65534,

jsonnet/components/alertmanager.libsonnet

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -235,6 +235,8 @@ function(params)
235235
},
236236
},
237237
spec+: {
238+
// The value of alertmanagerConfigSelector is defined at runtime by the Cluster Monitoring Operator.
239+
alertmanagerConfigSelector: null,
238240
podMetadata+: {
239241
annotations+: {
240242
'openshift.io/required-scc': 'nonroot',

jsonnet/components/prometheus-user-workload.libsonnet

Lines changed: 0 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -308,24 +308,6 @@ function(params)
308308
},
309309
},
310310
},
311-
alerting+: {
312-
alertmanagers:
313-
std.map(
314-
function(a) a {
315-
scheme: 'https',
316-
// the user-workload alertmanager configuration points to the openshift-monitoring namespace
317-
// since there is no dedicated alertmanager in the user-workload monitoring stack.
318-
namespace: 'openshift-monitoring', //FIXME(paulfantom)
319-
tlsConfig: {
320-
caFile: '/etc/prometheus/configmaps/serving-certs-ca-bundle/service-ca.crt',
321-
serverName: 'alertmanager-main.openshift-monitoring.svc',
322-
},
323-
bearerTokenFile: '/var/run/secrets/kubernetes.io/serviceaccount/token',
324-
apiVersion: 'v2',
325-
},
326-
super.alertmanagers,
327-
),
328-
},
329311
podMetadata+: {
330312
annotations+: {
331313
'openshift.io/required-scc': 'nonroot-v2',

jsonnet/components/prometheus.libsonnet

Lines changed: 0 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -331,21 +331,6 @@ function(params)
331331
// Enable some experimental features.
332332
// More at https://prometheus.io/docs/prometheus/latest/feature_flags/
333333
enableFeatures+: ['delayed-compaction', 'use-uncached-io'],
334-
alerting+: {
335-
alertmanagers:
336-
std.map(
337-
function(a) a {
338-
scheme: 'https',
339-
tlsConfig: {
340-
caFile: '/etc/prometheus/configmaps/serving-certs-ca-bundle/service-ca.crt',
341-
serverName: 'alertmanager-main',
342-
},
343-
bearerTokenFile: '/var/run/secrets/kubernetes.io/serviceaccount/token',
344-
apiVersion: 'v2',
345-
},
346-
super.alertmanagers,
347-
),
348-
},
349334
resources: {
350335
requests: {
351336
memory: '1Gi',

0 commit comments

Comments
 (0)