diff --git a/jsonnet/kube-prometheus/versions.json b/jsonnet/kube-prometheus/versions.json index 56f84f0aced540ea0b1eb79301c6f68aa444d278..6e3a9e17efe15503072e63fbe2cdc17fa2957421 100644 --- a/jsonnet/kube-prometheus/versions.json +++ b/jsonnet/kube-prometheus/versions.json @@ -1,12 +1,12 @@ { "alertmanager": "0.27.0", "blackboxExporter": "0.24.0", - "grafana": "10.3.3", - "kubeStateMetrics": "2.10.1", + "grafana": "10.4.0", + "kubeStateMetrics": "2.11.0", "nodeExporter": "1.7.0", "prometheus": "2.50.1", "prometheusAdapter": "0.11.2", - "prometheusOperator": "0.71.2", + "prometheusOperator": "0.72.0", "kubeRbacProxy": "0.16.0", "configmapReload": "0.5.0", "pyrra": "0.6.4" diff --git a/jsonnetfile.lock.json b/jsonnetfile.lock.json index 2ad9c215384844053377c582b466939d2e2ef41c..8d0e447666018348eb1936204e7d1d790c1710b1 100644 --- a/jsonnetfile.lock.json +++ b/jsonnetfile.lock.json @@ -18,7 +18,7 @@ "subdir": "contrib/mixin" } }, - "version": "e68afe7eb7ce3e5f83deb2a8ad58faa3b35494b9", + "version": "7688ce3f9042f904d9f5da81b5a5b5d65232e772", "sum": "xuUBd2vqF7asyVDe5CE08uPT/RxAdy8O75EjFJoMXXU=" }, { @@ -88,7 +88,7 @@ "subdir": "grafana-builder" } }, - "version": "8e9c60e6464c5030dd3b506bdb92c3dbab0394f7", + "version": "f364a45c27c385d86e399cd29e0b63baab7ac10a", "sum": "B49EzIY2WZsFxNMJcgRxE/gcZ9ltnS8pkOOV6Q5qioc=" }, { @@ -108,8 +108,8 @@ "subdir": "" } }, - "version": "295816ad067c7408053702174634189af802f915", - "sum": "nIhYLrj+q4uP9bMwr36/ebaVlo7EjhmpCRj0qMuRRDE=" + "version": "fc2e57a8839902ed4ba6cab5a99d642500f7102b", + "sum": "43waffw1QzvpY4rKcWoo3L7Vpee+DCYexwLDd5cPG0M=" }, { "source": { @@ -128,7 +128,7 @@ "subdir": "jsonnet/kube-state-metrics" } }, - "version": "44c65d27238c2a2e58099ef56fdc40529561ca91", + "version": "9ba1c3702142918e09e8eb5ca530e15198624259", "sum": "msMZyUvcebzRILLzNlTIiSOwa1XgQKtP7jbZTkiqwM0=" }, { @@ -138,7 +138,7 @@ "subdir": "jsonnet/kube-state-metrics-mixin" } }, - "version": "44c65d27238c2a2e58099ef56fdc40529561ca91", + "version": "9ba1c3702142918e09e8eb5ca530e15198624259", "sum": "qclI7LwucTjBef3PkGBkKxF0mfZPbHnn4rlNWKGtR4c=" }, { @@ -148,7 +148,7 @@ "subdir": "jsonnet/mixin" } }, - "version": "f433619e9a0e9be14e8830661f5f75c9b0cdd269", + "version": "a9b337560d77d5e30321dd917bbc98f891ac0af7", "sum": "gi+knjdxs2T715iIQIntrimbHRgHnpM8IFBJDD1gYfs=", "name": "prometheus-operator-mixin" }, @@ -159,8 +159,8 @@ "subdir": "jsonnet/prometheus-operator" } }, - "version": "f433619e9a0e9be14e8830661f5f75c9b0cdd269", - "sum": "hZpmPn+vGQ6lNLQTp04ZUGrXXZZhLJta4/ks2Fqubv4=" + "version": "a9b337560d77d5e30321dd917bbc98f891ac0af7", + "sum": "yCSfxXNySUqEBRqEhjjDTSvvGD1/Q1pePUzH5EXwQhM=" }, { "source": { @@ -169,7 +169,7 @@ "subdir": "doc/alertmanager-mixin" } }, - "version": "1eb83c21ebf4515386933298ee07e87f1c67f35e", + "version": "f82574a3763c21ca9bf2ada7baf69bd09e01d977", "sum": "IpF46ZXsm+0wJJAPtAre8+yxTNZA57mBqGpBP/r7/kw=", "name": "alertmanager" }, @@ -180,8 +180,8 @@ "subdir": "docs/node-mixin" } }, - "version": "c371a7f582bf2b9a185fc51bd0d3e842937a8ab2", - "sum": "PqPMf2Lqp7tW7T752kxdbMqF7JJ6vC8XrBN8WL5MMCc=" + "version": "9d359d082f55114bba58fb0b4be4d6ffb5683cf0", + "sum": "vWhHvFqV7+fxrQddTeGVKi1e4EzB3VWtNyD8TjSmevY=" }, { "source": { @@ -190,7 +190,7 @@ "subdir": "documentation/prometheus-mixin" } }, - "version": "d5f0a240faade2c694dfc877e3ac151b896bf66c", + "version": "2c267389f69809b4092d823f8a64a849ba3ea253", "sum": "u/Fpz2MPkezy71/q+c7mF0vc3hE9fWt2W/YbvF0LP/8=", "name": "prometheus" }, @@ -212,7 +212,7 @@ "subdir": "mixin" } }, - "version": "4c7997d25acc57166bfd726a64cc3c5710fcfe29", + "version": "f731719f95152f902e1b8562ebb83b1c30c201c3", "sum": "HhSSbGGCNHCMy1ee5jElYDm0yS9Vesa7QB2/SHKdjsY=", "name": "thanos-mixin" }, diff --git a/manifests/grafana-config.yaml b/manifests/grafana-config.yaml index c907f5599fd614af7b273b1913532699c2c1f1b0..4bb975ac27753e14438cdb469247e7100be0e8a7 100644 --- a/manifests/grafana-config.yaml +++ b/manifests/grafana-config.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-config namespace: monitoring stringData: diff --git a/manifests/grafana-dashboardDatasources.yaml b/manifests/grafana-dashboardDatasources.yaml index 99cb5b66a1e539aa2002a18692401b5e8d485c15..68e9441fe4acc3d132f4aa45e863ad587d25716b 100644 --- a/manifests/grafana-dashboardDatasources.yaml +++ b/manifests/grafana-dashboardDatasources.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-datasources namespace: monitoring stringData: diff --git a/manifests/grafana-dashboardDefinitions.yaml b/manifests/grafana-dashboardDefinitions.yaml index 01e539f16dc4a1441462e696a7f29e506e626e79..9645bff093f86c6114ff6036d5ceb69e7d251746 100644 --- a/manifests/grafana-dashboardDefinitions.yaml +++ b/manifests/grafana-dashboardDefinitions.yaml @@ -600,7 +600,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-alertmanager-overview namespace: monitoring - apiVersion: v1 @@ -1469,7 +1469,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-apiserver namespace: monitoring - apiVersion: v1 @@ -3341,7 +3341,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-cluster-total namespace: monitoring - apiVersion: v1 @@ -4505,7 +4505,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-controller-manager namespace: monitoring - apiVersion: v1 @@ -5130,7 +5130,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-grafana-overview namespace: monitoring - apiVersion: v1 @@ -8060,7 +8060,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-k8s-resources-cluster namespace: monitoring - apiVersion: v1 @@ -9375,7 +9375,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-k8s-resources-multicluster namespace: monitoring - apiVersion: v1 @@ -12025,7 +12025,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-k8s-resources-namespace namespace: monitoring - apiVersion: v1 @@ -12994,7 +12994,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-k8s-resources-node namespace: monitoring - apiVersion: v1 @@ -15328,7 +15328,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-k8s-resources-pod namespace: monitoring - apiVersion: v1 @@ -17250,7 +17250,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-k8s-resources-workload namespace: monitoring - apiVersion: v1 @@ -19325,7 +19325,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-k8s-resources-workloads-namespace namespace: monitoring - apiVersion: v1 @@ -21569,7 +21569,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-kubelet namespace: monitoring - apiVersion: v1 @@ -23023,7 +23023,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-namespace-by-pod namespace: monitoring - apiVersion: v1 @@ -24749,7 +24749,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-namespace-by-workload namespace: monitoring - apiVersion: v1 @@ -25801,7 +25801,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-node-cluster-rsrc-use namespace: monitoring - apiVersion: v1 @@ -26879,7 +26879,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-node-rsrc-use namespace: monitoring - apiVersion: v1 @@ -26957,7 +26957,7 @@ items: "steppedLine": false, "targets": [ { - "expr": "(\n (1 - sum without (mode) (rate(node_cpu_seconds_total{job=\"node-exporter\", mode=~\"idle|iowait|steal\", instance=\"$instance\"}[$__rate_interval])))\n/ ignoring(cpu) group_left\n count without (cpu, mode) (node_cpu_seconds_total{job=\"node-exporter\", mode=\"idle\", instance=\"$instance\"})\n)\n", + "expr": "(\n (1 - sum without (mode) (rate(node_cpu_seconds_total{job=\"node-exporter\", mode=~\"idle|iowait|steal\", instance=\"$instance\", cluster=\"$cluster\"}[$__rate_interval])))\n/ ignoring(cpu) group_left\n count without (cpu, mode) (node_cpu_seconds_total{job=\"node-exporter\", mode=\"idle\", instance=\"$instance\", cluster=\"$cluster\"})\n)\n", "format": "time_series", "intervalFactor": 5, "legendFormat": "{{cpu}}", @@ -27050,28 +27050,28 @@ items: "steppedLine": false, "targets": [ { - "expr": "node_load1{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_load1{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "1m load average", "refId": "A" }, { - "expr": "node_load5{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_load5{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "5m load average", "refId": "B" }, { - "expr": "node_load15{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_load15{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "15m load average", "refId": "C" }, { - "expr": "count(node_cpu_seconds_total{job=\"node-exporter\", instance=\"$instance\", mode=\"idle\"})", + "expr": "count(node_cpu_seconds_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", mode=\"idle\"})", "format": "time_series", "intervalFactor": 2, "legendFormat": "logical cores", @@ -27177,35 +27177,35 @@ items: "steppedLine": false, "targets": [ { - "expr": "node_memory_total_bytes{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_memory_total_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "Physical Memory", "refId": "A" }, { - "expr": "(\n node_memory_internal_bytes{job=\"node-exporter\", instance=\"$instance\"} -\n node_memory_purgeable_bytes{job=\"node-exporter\", instance=\"$instance\"} +\n node_memory_wired_bytes{job=\"node-exporter\", instance=\"$instance\"} +\n node_memory_compressed_bytes{job=\"node-exporter\", instance=\"$instance\"}\n)\n", + "expr": "(\n node_memory_internal_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"} -\n node_memory_purgeable_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"} +\n node_memory_wired_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"} +\n node_memory_compressed_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}\n)\n", "format": "time_series", "intervalFactor": 2, "legendFormat": "Memory Used", "refId": "B" }, { - "expr": "(\n node_memory_internal_bytes{job=\"node-exporter\", instance=\"$instance\"} -\n node_memory_purgeable_bytes{job=\"node-exporter\", instance=\"$instance\"}\n)\n", + "expr": "(\n node_memory_internal_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"} -\n node_memory_purgeable_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}\n)\n", "format": "time_series", "intervalFactor": 2, "legendFormat": "App Memory", "refId": "C" }, { - "expr": "node_memory_wired_bytes{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_memory_wired_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "Wired Memory", "refId": "D" }, { - "expr": "node_memory_compressed_bytes{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_memory_compressed_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "Compressed", @@ -27284,7 +27284,7 @@ items: "span": 3, "targets": [ { - "expr": "(\n (\n avg(node_memory_internal_bytes{job=\"node-exporter\", instance=\"$instance\"}) -\n avg(node_memory_purgeable_bytes{job=\"node-exporter\", instance=\"$instance\"}) +\n avg(node_memory_wired_bytes{job=\"node-exporter\", instance=\"$instance\"}) +\n avg(node_memory_compressed_bytes{job=\"node-exporter\", instance=\"$instance\"})\n ) /\n avg(node_memory_total_bytes{job=\"node-exporter\", instance=\"$instance\"})\n)\n*\n100\n", + "expr": "(\n (\n avg(node_memory_internal_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}) -\n avg(node_memory_purgeable_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}) +\n avg(node_memory_wired_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}) +\n avg(node_memory_compressed_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"})\n ) /\n avg(node_memory_total_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"})\n)\n*\n100\n", "format": "time_series", "intervalFactor": 2, "legendFormat": "" @@ -27360,21 +27360,21 @@ items: "steppedLine": false, "targets": [ { - "expr": "rate(node_disk_read_bytes_total{job=\"node-exporter\", instance=\"$instance\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", + "expr": "rate(node_disk_read_bytes_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}} read", "refId": "A" }, { - "expr": "rate(node_disk_written_bytes_total{job=\"node-exporter\", instance=\"$instance\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", + "expr": "rate(node_disk_written_bytes_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}} written", "refId": "B" }, { - "expr": "rate(node_disk_io_time_seconds_total{job=\"node-exporter\", instance=\"$instance\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", + "expr": "rate(node_disk_io_time_seconds_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}} io time", @@ -27528,14 +27528,14 @@ items: "span": 6, "targets": [ { - "expr": "max by (mountpoint) (node_filesystem_size_bytes{job=\"node-exporter\", instance=\"$instance\", fstype!=\"\", mountpoint!=\"\"})\n", + "expr": "max by (mountpoint) (node_filesystem_size_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", fstype!=\"\", mountpoint!=\"\"})\n", "format": "table", "instant": true, "intervalFactor": 2, "legendFormat": "" }, { - "expr": "max by (mountpoint) (node_filesystem_avail_bytes{job=\"node-exporter\", instance=\"$instance\", fstype!=\"\", mountpoint!=\"\"})\n", + "expr": "max by (mountpoint) (node_filesystem_avail_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", fstype!=\"\", mountpoint!=\"\"})\n", "format": "table", "instant": true, "intervalFactor": 2, @@ -27700,7 +27700,7 @@ items: "steppedLine": false, "targets": [ { - "expr": "rate(node_network_receive_bytes_total{job=\"node-exporter\", instance=\"$instance\", device!=\"lo\"}[$__rate_interval]) * 8", + "expr": "rate(node_network_receive_bytes_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device!=\"lo\"}[$__rate_interval]) * 8", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}}", @@ -27794,7 +27794,7 @@ items: "steppedLine": false, "targets": [ { - "expr": "rate(node_network_transmit_bytes_total{job=\"node-exporter\", instance=\"$instance\", device!=\"lo\"}[$__rate_interval]) * 8", + "expr": "rate(node_network_transmit_bytes_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device!=\"lo\"}[$__rate_interval]) * 8", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}}", @@ -27878,6 +27878,32 @@ items: "allValue": null, "current": { + }, + "datasource": "$datasource", + "hide": 2, + "includeAll": false, + "label": "Cluster", + "multi": false, + "name": "cluster", + "options": [ + + ], + "query": "label_values(node_uname_info{job=\"node-exporter\", sysname=\"Darwin\"}, cluster)", + "refresh": 2, + "regex": "", + "sort": 0, + "tagValuesQuery": "", + "tags": [ + + ], + "tagsQuery": "", + "type": "query", + "useTags": false + }, + { + "allValue": null, + "current": { + }, "datasource": "$datasource", "hide": 0, @@ -27888,7 +27914,7 @@ items: "options": [ ], - "query": "label_values(node_uname_info{job=\"node-exporter\", sysname=\"Darwin\"}, instance)", + "query": "label_values(node_uname_info{job=\"node-exporter\", cluster=\"$cluster\", sysname=\"Darwin\"}, instance)", "refresh": 2, "regex": "", "sort": 0, @@ -27941,7 +27967,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-nodes-darwin namespace: monitoring - apiVersion: v1 @@ -28019,7 +28045,7 @@ items: "steppedLine": false, "targets": [ { - "expr": "(\n (1 - sum without (mode) (rate(node_cpu_seconds_total{job=\"node-exporter\", mode=~\"idle|iowait|steal\", instance=\"$instance\"}[$__rate_interval])))\n/ ignoring(cpu) group_left\n count without (cpu, mode) (node_cpu_seconds_total{job=\"node-exporter\", mode=\"idle\", instance=\"$instance\"})\n)\n", + "expr": "(\n (1 - sum without (mode) (rate(node_cpu_seconds_total{job=\"node-exporter\", mode=~\"idle|iowait|steal\", instance=\"$instance\", cluster=\"$cluster\"}[$__rate_interval])))\n/ ignoring(cpu) group_left\n count without (cpu, mode) (node_cpu_seconds_total{job=\"node-exporter\", mode=\"idle\", instance=\"$instance\", cluster=\"$cluster\"})\n)\n", "format": "time_series", "intervalFactor": 5, "legendFormat": "{{cpu}}", @@ -28112,28 +28138,28 @@ items: "steppedLine": false, "targets": [ { - "expr": "node_load1{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_load1{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "1m load average", "refId": "A" }, { - "expr": "node_load5{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_load5{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "5m load average", "refId": "B" }, { - "expr": "node_load15{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_load15{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "15m load average", "refId": "C" }, { - "expr": "count(node_cpu_seconds_total{job=\"node-exporter\", instance=\"$instance\", mode=\"idle\"})", + "expr": "count(node_cpu_seconds_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", mode=\"idle\"})", "format": "time_series", "intervalFactor": 2, "legendFormat": "logical cores", @@ -28239,28 +28265,28 @@ items: "steppedLine": false, "targets": [ { - "expr": "(\n node_memory_MemTotal_bytes{job=\"node-exporter\", instance=\"$instance\"}\n-\n node_memory_MemFree_bytes{job=\"node-exporter\", instance=\"$instance\"}\n-\n node_memory_Buffers_bytes{job=\"node-exporter\", instance=\"$instance\"}\n-\n node_memory_Cached_bytes{job=\"node-exporter\", instance=\"$instance\"}\n)\n", + "expr": "(\n node_memory_MemTotal_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}\n-\n node_memory_MemFree_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}\n-\n node_memory_Buffers_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}\n-\n node_memory_Cached_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}\n)\n", "format": "time_series", "intervalFactor": 2, "legendFormat": "memory used", "refId": "A" }, { - "expr": "node_memory_Buffers_bytes{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_memory_Buffers_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "memory buffers", "refId": "B" }, { - "expr": "node_memory_Cached_bytes{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_memory_Cached_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "memory cached", "refId": "C" }, { - "expr": "node_memory_MemFree_bytes{job=\"node-exporter\", instance=\"$instance\"}", + "expr": "node_memory_MemFree_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}", "format": "time_series", "intervalFactor": 2, "legendFormat": "memory free", @@ -28339,7 +28365,7 @@ items: "span": 3, "targets": [ { - "expr": "100 -\n(\n avg(node_memory_MemAvailable_bytes{job=\"node-exporter\", instance=\"$instance\"}) /\n avg(node_memory_MemTotal_bytes{job=\"node-exporter\", instance=\"$instance\"})\n* 100\n)\n", + "expr": "100 -\n(\n avg(node_memory_MemAvailable_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"}) /\n avg(node_memory_MemTotal_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\"})\n* 100\n)\n", "format": "time_series", "intervalFactor": 2, "legendFormat": "" @@ -28415,21 +28441,21 @@ items: "steppedLine": false, "targets": [ { - "expr": "rate(node_disk_read_bytes_total{job=\"node-exporter\", instance=\"$instance\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", + "expr": "rate(node_disk_read_bytes_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}} read", "refId": "A" }, { - "expr": "rate(node_disk_written_bytes_total{job=\"node-exporter\", instance=\"$instance\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", + "expr": "rate(node_disk_written_bytes_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}} written", "refId": "B" }, { - "expr": "rate(node_disk_io_time_seconds_total{job=\"node-exporter\", instance=\"$instance\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", + "expr": "rate(node_disk_io_time_seconds_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device=~\"(/dev/)?(mmcblk.p.+|nvme.+|rbd.+|sd.+|vd.+|xvd.+|dm-.+|md.+|dasd.+)\"}[$__rate_interval])", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}} io time", @@ -28583,14 +28609,14 @@ items: "span": 6, "targets": [ { - "expr": "max by (mountpoint) (node_filesystem_size_bytes{job=\"node-exporter\", instance=\"$instance\", fstype!=\"\", mountpoint!=\"\"})\n", + "expr": "max by (mountpoint) (node_filesystem_size_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", fstype!=\"\", mountpoint!=\"\"})\n", "format": "table", "instant": true, "intervalFactor": 2, "legendFormat": "" }, { - "expr": "max by (mountpoint) (node_filesystem_avail_bytes{job=\"node-exporter\", instance=\"$instance\", fstype!=\"\", mountpoint!=\"\"})\n", + "expr": "max by (mountpoint) (node_filesystem_avail_bytes{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", fstype!=\"\", mountpoint!=\"\"})\n", "format": "table", "instant": true, "intervalFactor": 2, @@ -28755,7 +28781,7 @@ items: "steppedLine": false, "targets": [ { - "expr": "rate(node_network_receive_bytes_total{job=\"node-exporter\", instance=\"$instance\", device!=\"lo\"}[$__rate_interval]) * 8", + "expr": "rate(node_network_receive_bytes_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device!=\"lo\"}[$__rate_interval]) * 8", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}}", @@ -28849,7 +28875,7 @@ items: "steppedLine": false, "targets": [ { - "expr": "rate(node_network_transmit_bytes_total{job=\"node-exporter\", instance=\"$instance\", device!=\"lo\"}[$__rate_interval]) * 8", + "expr": "rate(node_network_transmit_bytes_total{job=\"node-exporter\", instance=\"$instance\", cluster=\"$cluster\", device!=\"lo\"}[$__rate_interval]) * 8", "format": "time_series", "intervalFactor": 1, "legendFormat": "{{device}}", @@ -28933,6 +28959,32 @@ items: "allValue": null, "current": { + }, + "datasource": "$datasource", + "hide": 2, + "includeAll": false, + "label": "Cluster", + "multi": false, + "name": "cluster", + "options": [ + + ], + "query": "label_values(node_uname_info{job=\"node-exporter\", sysname!=\"Darwin\"}, cluster)", + "refresh": 2, + "regex": "", + "sort": 0, + "tagValuesQuery": "", + "tags": [ + + ], + "tagsQuery": "", + "type": "query", + "useTags": false + }, + { + "allValue": null, + "current": { + }, "datasource": "$datasource", "hide": 0, @@ -28943,7 +28995,7 @@ items: "options": [ ], - "query": "label_values(node_uname_info{job=\"node-exporter\", sysname!=\"Darwin\"}, instance)", + "query": "label_values(node_uname_info{job=\"node-exporter\", cluster=\"$cluster\", sysname!=\"Darwin\"}, instance)", "refresh": 2, "regex": "", "sort": 0, @@ -28996,7 +29048,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-nodes namespace: monitoring - apiVersion: v1 @@ -29557,7 +29609,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-persistentvolumesusage namespace: monitoring - apiVersion: v1 @@ -30775,7 +30827,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-pod-total namespace: monitoring - apiVersion: v1 @@ -32434,7 +32486,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-prometheus-remote-write namespace: monitoring - apiVersion: v1 @@ -33675,7 +33727,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-prometheus namespace: monitoring - apiVersion: v1 @@ -34919,7 +34971,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-proxy namespace: monitoring - apiVersion: v1 @@ -36006,7 +36058,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-scheduler namespace: monitoring - apiVersion: v1 @@ -37434,7 +37486,7 @@ items: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboard-workload-total namespace: monitoring kind: ConfigMapList diff --git a/manifests/grafana-dashboardSources.yaml b/manifests/grafana-dashboardSources.yaml index 3599b6fbb2718728a4ea1fa65b26eaf60ce4e1d8..b9a06ddb43f78b02c24101961f86c3e2d93dd496 100644 --- a/manifests/grafana-dashboardSources.yaml +++ b/manifests/grafana-dashboardSources.yaml @@ -22,6 +22,6 @@ metadata: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana-dashboards namespace: monitoring diff --git a/manifests/grafana-deployment.yaml b/manifests/grafana-deployment.yaml index 1bf281bed3221c4047bd27ad78b443001a973ebb..4dac9bad064cc02258525f9bca08f7bb0cd4d9c7 100644 --- a/manifests/grafana-deployment.yaml +++ b/manifests/grafana-deployment.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana namespace: monitoring spec: @@ -18,19 +18,19 @@ spec: template: metadata: annotations: - checksum/grafana-config: 098832873ac431e10436214f4da40273 - checksum/grafana-dashboardproviders: a7193910063bb76138fe42e9b56ed412 - checksum/grafana-datasources: a416639eb8ee8a83644d2f613a4c10d0 + checksum/grafana-config: 4ae9fc7e2a73ddf9a73f52561505ee7e + checksum/grafana-dashboardproviders: 214f4a62a17313df84bb417b1aade52d + checksum/grafana-datasources: 12a0d2d2166e2b6c5b656d98c5e2110c labels: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 spec: automountServiceAccountToken: false containers: - env: [] - image: grafana/grafana:10.3.3 + image: grafana/grafana:10.4.0 name: grafana ports: - containerPort: 3000 diff --git a/manifests/grafana-networkPolicy.yaml b/manifests/grafana-networkPolicy.yaml index 18a67a4afc5809774aab97e0f89160de1f771754..019a5a611b96109e17447b3b2685fe52177feb0e 100644 --- a/manifests/grafana-networkPolicy.yaml +++ b/manifests/grafana-networkPolicy.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana namespace: monitoring spec: diff --git a/manifests/grafana-prometheusRule.yaml b/manifests/grafana-prometheusRule.yaml index d749f77df5c7669597147323873d3c559402fe61..487c5c3fcd63d13a3b715173855531e9de0e4731 100644 --- a/manifests/grafana-prometheusRule.yaml +++ b/manifests/grafana-prometheusRule.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 prometheus: k8s role: alert-rules name: grafana-rules diff --git a/manifests/grafana-service.yaml b/manifests/grafana-service.yaml index 954b6d7ce9dc34fb4e30c2ad456bc4f742852b38..fedc3ca2498216b004ddafed9717ede48aad759c 100644 --- a/manifests/grafana-service.yaml +++ b/manifests/grafana-service.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana namespace: monitoring spec: diff --git a/manifests/grafana-serviceAccount.yaml b/manifests/grafana-serviceAccount.yaml index 85328fb89ee2dc5d3a72b088eeb25340bd1aeea6..3ab5a3558f8605b11739e8f8dceb36d0a0c7eeff 100644 --- a/manifests/grafana-serviceAccount.yaml +++ b/manifests/grafana-serviceAccount.yaml @@ -6,6 +6,6 @@ metadata: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana namespace: monitoring diff --git a/manifests/grafana-serviceMonitor.yaml b/manifests/grafana-serviceMonitor.yaml index be1f11e43377be733f136071b943c157f56983f6..e1d001b17c710f7656991164280df8aabd3b9eb8 100644 --- a/manifests/grafana-serviceMonitor.yaml +++ b/manifests/grafana-serviceMonitor.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: grafana app.kubernetes.io/name: grafana app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 10.3.3 + app.kubernetes.io/version: 10.4.0 name: grafana namespace: monitoring spec: diff --git a/manifests/kubeStateMetrics-clusterRole.yaml b/manifests/kubeStateMetrics-clusterRole.yaml index 9cc599de2d6b13b85666fdbec00c0f7fbbfae971..388f6e01b9064dd9857014477c76799a60ec4958 100644 --- a/manifests/kubeStateMetrics-clusterRole.yaml +++ b/manifests/kubeStateMetrics-clusterRole.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: exporter app.kubernetes.io/name: kube-state-metrics app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 2.10.1 + app.kubernetes.io/version: 2.11.0 name: kube-state-metrics rules: - apiGroups: diff --git a/manifests/kubeStateMetrics-clusterRoleBinding.yaml b/manifests/kubeStateMetrics-clusterRoleBinding.yaml index 5b67f2a790ac0be2764c0a8ef8493c198df2284b..55062bc456337d95c61b20bf393be1d49ce5c7a7 100644 --- a/manifests/kubeStateMetrics-clusterRoleBinding.yaml +++ b/manifests/kubeStateMetrics-clusterRoleBinding.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: exporter app.kubernetes.io/name: kube-state-metrics app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 2.10.1 + app.kubernetes.io/version: 2.11.0 name: kube-state-metrics roleRef: apiGroup: rbac.authorization.k8s.io diff --git a/manifests/kubeStateMetrics-deployment.yaml b/manifests/kubeStateMetrics-deployment.yaml index 5a4b4c9815b92ce539df07808bfe45a632ca901e..a70096788950e436271ca67893362b5cfacefd8c 100644 --- a/manifests/kubeStateMetrics-deployment.yaml +++ b/manifests/kubeStateMetrics-deployment.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: exporter app.kubernetes.io/name: kube-state-metrics app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 2.10.1 + app.kubernetes.io/version: 2.11.0 name: kube-state-metrics namespace: monitoring spec: @@ -23,7 +23,7 @@ spec: app.kubernetes.io/component: exporter app.kubernetes.io/name: kube-state-metrics app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 2.10.1 + app.kubernetes.io/version: 2.11.0 spec: automountServiceAccountToken: true containers: @@ -32,7 +32,7 @@ spec: - --port=8081 - --telemetry-host=127.0.0.1 - --telemetry-port=8082 - image: registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.10.1 + image: registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.11.0 name: kube-state-metrics resources: limits: diff --git a/manifests/kubeStateMetrics-networkPolicy.yaml b/manifests/kubeStateMetrics-networkPolicy.yaml index 17f500814b547a21070c23373b73009a15d557fb..061a40aed2c0093acdc1162c3a888a70c6520736 100644 --- a/manifests/kubeStateMetrics-networkPolicy.yaml +++ b/manifests/kubeStateMetrics-networkPolicy.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: exporter app.kubernetes.io/name: kube-state-metrics app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 2.10.1 + app.kubernetes.io/version: 2.11.0 name: kube-state-metrics namespace: monitoring spec: diff --git a/manifests/kubeStateMetrics-prometheusRule.yaml b/manifests/kubeStateMetrics-prometheusRule.yaml index 52d02f53da7150532435d66e49d59facdf99a3af..0350e0ee6fa31d9ca30cad7c9dcf56e78f170b1a 100644 --- a/manifests/kubeStateMetrics-prometheusRule.yaml +++ b/manifests/kubeStateMetrics-prometheusRule.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: exporter app.kubernetes.io/name: kube-state-metrics app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 2.10.1 + app.kubernetes.io/version: 2.11.0 prometheus: k8s role: alert-rules name: kube-state-metrics-rules diff --git a/manifests/kubeStateMetrics-service.yaml b/manifests/kubeStateMetrics-service.yaml index 32bba754f5507fc6f50a48f937193d1976b57cfd..3a79c2b95aefd27891a320699644307bea3e17dd 100644 --- a/manifests/kubeStateMetrics-service.yaml +++ b/manifests/kubeStateMetrics-service.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: exporter app.kubernetes.io/name: kube-state-metrics app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 2.10.1 + app.kubernetes.io/version: 2.11.0 name: kube-state-metrics namespace: monitoring spec: diff --git a/manifests/kubeStateMetrics-serviceAccount.yaml b/manifests/kubeStateMetrics-serviceAccount.yaml index 47ae2b3cf26a6f891d82af07dc746c9cc5fa82ba..49360ca4e0c194ddc39e5cadaf56c2c0c2e65777 100644 --- a/manifests/kubeStateMetrics-serviceAccount.yaml +++ b/manifests/kubeStateMetrics-serviceAccount.yaml @@ -6,6 +6,6 @@ metadata: app.kubernetes.io/component: exporter app.kubernetes.io/name: kube-state-metrics app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 2.10.1 + app.kubernetes.io/version: 2.11.0 name: kube-state-metrics namespace: monitoring diff --git a/manifests/kubeStateMetrics-serviceMonitor.yaml b/manifests/kubeStateMetrics-serviceMonitor.yaml index ebcb619bc25b1802fdf907177048756fd9c1648d..c458521eb5891f88efb560c2b4d5edc2d47e0c34 100644 --- a/manifests/kubeStateMetrics-serviceMonitor.yaml +++ b/manifests/kubeStateMetrics-serviceMonitor.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: exporter app.kubernetes.io/name: kube-state-metrics app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 2.10.1 + app.kubernetes.io/version: 2.11.0 name: kube-state-metrics namespace: monitoring spec: diff --git a/manifests/prometheusOperator-clusterRole.yaml b/manifests/prometheusOperator-clusterRole.yaml index a799fc32e2910d4ff7ac2302321c7681ec4f346f..ac9b2f41558c676358648f7d2d8f8b465eb8b4fb 100644 --- a/manifests/prometheusOperator-clusterRole.yaml +++ b/manifests/prometheusOperator-clusterRole.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 name: prometheus-operator rules: - apiGroups: diff --git a/manifests/prometheusOperator-clusterRoleBinding.yaml b/manifests/prometheusOperator-clusterRoleBinding.yaml index 2e7c21edd743bcf49ed5adcda58e64e8afea5153..75dfe78bfb6f6fb5179bd21d3612fd841436c616 100644 --- a/manifests/prometheusOperator-clusterRoleBinding.yaml +++ b/manifests/prometheusOperator-clusterRoleBinding.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 name: prometheus-operator roleRef: apiGroup: rbac.authorization.k8s.io diff --git a/manifests/prometheusOperator-deployment.yaml b/manifests/prometheusOperator-deployment.yaml index 42e089d8c9a21a41f6a80cbaefda7cf2fec8a038..44f891a084e437c751319c4c029e35c5aedc3ad6 100644 --- a/manifests/prometheusOperator-deployment.yaml +++ b/manifests/prometheusOperator-deployment.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 name: prometheus-operator namespace: monitoring spec: @@ -23,17 +23,17 @@ spec: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 spec: automountServiceAccountToken: true containers: - args: - --kubelet-service=kube-system/kubelet - - --prometheus-config-reloader=quay.io/prometheus-operator/prometheus-config-reloader:v0.71.2 + - --prometheus-config-reloader=quay.io/prometheus-operator/prometheus-config-reloader:v0.72.0 env: - name: GOGC value: "30" - image: quay.io/prometheus-operator/prometheus-operator:v0.71.2 + image: quay.io/prometheus-operator/prometheus-operator:v0.72.0 name: prometheus-operator ports: - containerPort: 8080 diff --git a/manifests/prometheusOperator-networkPolicy.yaml b/manifests/prometheusOperator-networkPolicy.yaml index 30ecd991e056e1b82d23c54ff320e52f660cd1dc..8251a209774d47dd42a1a74bb0762cfaeefa5251 100644 --- a/manifests/prometheusOperator-networkPolicy.yaml +++ b/manifests/prometheusOperator-networkPolicy.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 name: prometheus-operator namespace: monitoring spec: diff --git a/manifests/prometheusOperator-prometheusRule.yaml b/manifests/prometheusOperator-prometheusRule.yaml index e163f10ba8569d381774219ca7ed2e45183936fd..8e566a951f8dc8d9211c8f1ec07a1db729b2ae55 100644 --- a/manifests/prometheusOperator-prometheusRule.yaml +++ b/manifests/prometheusOperator-prometheusRule.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 prometheus: k8s role: alert-rules name: prometheus-operator-rules diff --git a/manifests/prometheusOperator-service.yaml b/manifests/prometheusOperator-service.yaml index 8aba259be48d72c26fd5ffe78c2e79d840e13f75..02545392c3c188c0960c54c3eb3ce1b8be700092 100644 --- a/manifests/prometheusOperator-service.yaml +++ b/manifests/prometheusOperator-service.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 name: prometheus-operator namespace: monitoring spec: diff --git a/manifests/prometheusOperator-serviceAccount.yaml b/manifests/prometheusOperator-serviceAccount.yaml index d600dfa227f926593435874590ed77150b8e25c4..7f67bfdf2d4bc24cf9e20723ec3acee39ed90e61 100644 --- a/manifests/prometheusOperator-serviceAccount.yaml +++ b/manifests/prometheusOperator-serviceAccount.yaml @@ -6,6 +6,6 @@ metadata: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 name: prometheus-operator namespace: monitoring diff --git a/manifests/prometheusOperator-serviceMonitor.yaml b/manifests/prometheusOperator-serviceMonitor.yaml index a33dde903565b3fdaed86de2bc7c98edc56983cd..aca2ee9d91471d8c65cad3ff7116004261fcb528 100644 --- a/manifests/prometheusOperator-serviceMonitor.yaml +++ b/manifests/prometheusOperator-serviceMonitor.yaml @@ -5,7 +5,7 @@ metadata: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 name: prometheus-operator namespace: monitoring spec: @@ -21,4 +21,4 @@ spec: app.kubernetes.io/component: controller app.kubernetes.io/name: prometheus-operator app.kubernetes.io/part-of: kube-prometheus - app.kubernetes.io/version: 0.71.2 + app.kubernetes.io/version: 0.72.0 diff --git a/manifests/setup/0alertmanagerConfigCustomResourceDefinition.yaml b/manifests/setup/0alertmanagerConfigCustomResourceDefinition.yaml index 67387c72a5a92a27f1e538644138c5314ee87142..4762db224bafd57987caa77123be5b17bfb09eb0 100644 --- a/manifests/setup/0alertmanagerConfigCustomResourceDefinition.yaml +++ b/manifests/setup/0alertmanagerConfigCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: alertmanagerconfigs.monitoring.coreos.com spec: group: monitoring.coreos.com diff --git a/manifests/setup/0alertmanagerCustomResourceDefinition.yaml b/manifests/setup/0alertmanagerCustomResourceDefinition.yaml index b2dc41d6e69fafb3c696f4253cc890001c028ffb..5d1685f4c9b4e66b47b764ad0de99c261b85eea7 100644 --- a/manifests/setup/0alertmanagerCustomResourceDefinition.yaml +++ b/manifests/setup/0alertmanagerCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: alertmanagers.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -1961,6 +1961,11 @@ spec: - name type: object type: array + enableFeatures: + description: "Enable access to Alertmanager feature flags. By default, no features are enabled. Enabling features which are disabled by default is entirely outside the scope of what the maintainers will support and by doing so, you accept that this behaviour may break at any time without notice. \n It requires Alertmanager >= 0.27.0." + items: + type: string + type: array externalUrl: description: The external URL the Alertmanager instances will be available under. This is necessary to generate correct URLs. This is necessary if Alertmanager is not served from root of a DNS name. type: string diff --git a/manifests/setup/0podmonitorCustomResourceDefinition.yaml b/manifests/setup/0podmonitorCustomResourceDefinition.yaml index 64eb9b979e185f832107842113aba7dd3e4eb754..8ae9e2af1a6a1ccfd2853e586c693d57d6580765 100644 --- a/manifests/setup/0podmonitorCustomResourceDefinition.yaml +++ b/manifests/setup/0podmonitorCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: podmonitors.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -41,6 +41,10 @@ spec: description: When set to true, Prometheus must have the `get` permission on the `Nodes` objects. type: boolean type: object + bodySizeLimit: + description: "When defined, bodySizeLimit specifies a job level limit on the size of uncompressed response body that will be accepted by Prometheus. \n It requires Prometheus >= v2.28.0." + pattern: (^0|([0-9]*[.])?[0-9]+((K|M|G|T|E|P)i?)?B)$ + type: string jobLabel: description: "The label to use to retrieve the job name from. `jobLabel` selects the label from the associated Kubernetes `Pod` object which will be used as the `job` label for all metrics. \n For example if `jobLabel` is set to `foo` and the Kubernetes `Pod` object is labeled with `foo: bar`, then Prometheus adds the `job=\"bar\"` label to all ingested metrics. \n If the value of this field is empty, the `job` label of the metrics defaults to the namespace and name of the PodMonitor object (e.g. `<namespace>/<name>`)." type: string diff --git a/manifests/setup/0probeCustomResourceDefinition.yaml b/manifests/setup/0probeCustomResourceDefinition.yaml index 05617c7d53eadadf303cbfcb1362371c8b43874d..30b0816bcf62c174e5fadcb84e35625cbd48ad05 100644 --- a/manifests/setup/0probeCustomResourceDefinition.yaml +++ b/manifests/setup/0probeCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: probes.monitoring.coreos.com spec: group: monitoring.coreos.com diff --git a/manifests/setup/0prometheusCustomResourceDefinition.yaml b/manifests/setup/0prometheusCustomResourceDefinition.yaml index 11504617c583593f62824a32cf2c77f079ff9f0f..bb729372348fc2cb81b0ba3462122a94b7e10caf 100644 --- a/manifests/setup/0prometheusCustomResourceDefinition.yaml +++ b/manifests/setup/0prometheusCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: prometheuses.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -3727,12 +3727,14 @@ spec: properties: batchSendDeadline: description: BatchSendDeadline is the maximum time a sample will wait in buffer. + pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string capacity: description: Capacity is the number of samples to buffer per shard before we start dropping them. type: integer maxBackoff: description: MaxBackoff is the maximum retry delay. + pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string maxRetries: description: MaxRetries is the maximum number of times to retry a batch on recoverable errors. @@ -3745,6 +3747,7 @@ spec: type: integer minBackoff: description: MinBackoff is the initial retry delay. Gets doubled for every retry. + pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string minShards: description: MinShards is the minimum number of shards, i.e. amount of concurrency. @@ -3752,6 +3755,10 @@ spec: retryOnRateLimit: description: Retry upon receiving a 429 status code from the remote-write storage. This is experimental feature and might change in the future. type: boolean + sampleAgeLimit: + description: SampleAgeLimit drops samples older than the limit. It requires Prometheus >= v2.50.0. + pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ + type: string type: object remoteTimeout: description: Timeout for requests to the remote write endpoint. diff --git a/manifests/setup/0prometheusagentCustomResourceDefinition.yaml b/manifests/setup/0prometheusagentCustomResourceDefinition.yaml index b3ec457a4d1129c020765821b8c4d86a5b68670e..8c889d93c8daf2f43ba2bb23bea7fc21003d7b2d 100644 --- a/manifests/setup/0prometheusagentCustomResourceDefinition.yaml +++ b/manifests/setup/0prometheusagentCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: prometheusagents.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -3082,12 +3082,14 @@ spec: properties: batchSendDeadline: description: BatchSendDeadline is the maximum time a sample will wait in buffer. + pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string capacity: description: Capacity is the number of samples to buffer per shard before we start dropping them. type: integer maxBackoff: description: MaxBackoff is the maximum retry delay. + pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string maxRetries: description: MaxRetries is the maximum number of times to retry a batch on recoverable errors. @@ -3100,6 +3102,7 @@ spec: type: integer minBackoff: description: MinBackoff is the initial retry delay. Gets doubled for every retry. + pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ type: string minShards: description: MinShards is the minimum number of shards, i.e. amount of concurrency. @@ -3107,6 +3110,10 @@ spec: retryOnRateLimit: description: Retry upon receiving a 429 status code from the remote-write storage. This is experimental feature and might change in the future. type: boolean + sampleAgeLimit: + description: SampleAgeLimit drops samples older than the limit. It requires Prometheus >= v2.50.0. + pattern: ^(0|(([0-9]+)y)?(([0-9]+)w)?(([0-9]+)d)?(([0-9]+)h)?(([0-9]+)m)?(([0-9]+)s)?(([0-9]+)ms)?)$ + type: string type: object remoteTimeout: description: Timeout for requests to the remote write endpoint. diff --git a/manifests/setup/0prometheusruleCustomResourceDefinition.yaml b/manifests/setup/0prometheusruleCustomResourceDefinition.yaml index 3b9a9f337ff7fe11ad8378681dd4af67438c003b..a618f50117bc46148ee8980e91dbc8968fdadf10 100644 --- a/manifests/setup/0prometheusruleCustomResourceDefinition.yaml +++ b/manifests/setup/0prometheusruleCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: prometheusrules.monitoring.coreos.com spec: group: monitoring.coreos.com diff --git a/manifests/setup/0scrapeconfigCustomResourceDefinition.yaml b/manifests/setup/0scrapeconfigCustomResourceDefinition.yaml index 6e2317e8a8a211968ea7008b0753a629b08e564e..b7000b99444b3eeab723a08d9bcb920f2b837a65 100644 --- a/manifests/setup/0scrapeconfigCustomResourceDefinition.yaml +++ b/manifests/setup/0scrapeconfigCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: scrapeconfigs.monitoring.coreos.com spec: group: monitoring.coreos.com diff --git a/manifests/setup/0servicemonitorCustomResourceDefinition.yaml b/manifests/setup/0servicemonitorCustomResourceDefinition.yaml index 2148fe09421f752ba99c3f645de9f430484c65c0..c2e3a48214ac7741622af839628932cbea3adfbd 100644 --- a/manifests/setup/0servicemonitorCustomResourceDefinition.yaml +++ b/manifests/setup/0servicemonitorCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: servicemonitors.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -41,6 +41,10 @@ spec: description: When set to true, Prometheus must have the `get` permission on the `Nodes` objects. type: boolean type: object + bodySizeLimit: + description: "When defined, bodySizeLimit specifies a job level limit on the size of uncompressed response body that will be accepted by Prometheus. \n It requires Prometheus >= v2.28.0." + pattern: (^0|([0-9]*[.])?[0-9]+((K|M|G|T|E|P)i?)?B)$ + type: string endpoints: description: List of endpoints part of this ServiceMonitor. items: diff --git a/manifests/setup/0thanosrulerCustomResourceDefinition.yaml b/manifests/setup/0thanosrulerCustomResourceDefinition.yaml index cc5a25901f7382056ff83d54e3d4f7d7ccdaefda..bf8740993d8f7d199d65e7425cbce6144153139c 100644 --- a/manifests/setup/0thanosrulerCustomResourceDefinition.yaml +++ b/manifests/setup/0thanosrulerCustomResourceDefinition.yaml @@ -3,7 +3,7 @@ kind: CustomResourceDefinition metadata: annotations: controller-gen.kubebuilder.io/version: v0.13.0 - operator.prometheus.io/version: 0.71.2 + operator.prometheus.io/version: 0.72.0 name: thanosrulers.monitoring.coreos.com spec: group: monitoring.coreos.com @@ -4329,6 +4329,160 @@ spec: - name type: object type: array + web: + description: Defines the configuration of the ThanosRuler web server. + properties: + httpConfig: + description: Defines HTTP parameters for web server. + properties: + headers: + description: List of headers that can be added to HTTP responses. + properties: + contentSecurityPolicy: + description: Set the Content-Security-Policy header to HTTP responses. Unset if blank. + type: string + strictTransportSecurity: + description: Set the Strict-Transport-Security header to HTTP responses. Unset if blank. Please make sure that you use this with care as this header might force browsers to load Prometheus and the other applications hosted on the same domain and subdomains over HTTPS. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Strict-Transport-Security + type: string + xContentTypeOptions: + description: Set the X-Content-Type-Options header to HTTP responses. Unset if blank. Accepted value is nosniff. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Content-Type-Options + enum: + - "" + - NoSniff + type: string + xFrameOptions: + description: Set the X-Frame-Options header to HTTP responses. Unset if blank. Accepted values are deny and sameorigin. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-Frame-Options + enum: + - "" + - Deny + - SameOrigin + type: string + xXSSProtection: + description: Set the X-XSS-Protection header to all responses. Unset if blank. https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/X-XSS-Protection + type: string + type: object + http2: + description: Enable HTTP/2 support. Note that HTTP/2 is only supported with TLS. When TLSConfig is not configured, HTTP/2 will be disabled. Whenever the value of the field changes, a rolling update will be triggered. + type: boolean + type: object + tlsConfig: + description: Defines the TLS parameters for HTTPS. + properties: + cert: + description: Contains the TLS certificate for the server. + properties: + configMap: + description: ConfigMap containing data to use for the targets. + properties: + key: + description: The key to select. + type: string + name: + description: 'Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?' + type: string + optional: + description: Specify whether the ConfigMap or its key must be defined + type: boolean + required: + - key + type: object + x-kubernetes-map-type: atomic + secret: + description: Secret containing data to use for the targets. + properties: + key: + description: The key of the secret to select from. Must be a valid secret key. + type: string + name: + description: 'Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?' + type: string + optional: + description: Specify whether the Secret or its key must be defined + type: boolean + required: + - key + type: object + x-kubernetes-map-type: atomic + type: object + cipherSuites: + description: 'List of supported cipher suites for TLS versions up to TLS 1.2. If empty, Go default cipher suites are used. Available cipher suites are documented in the go documentation: https://golang.org/pkg/crypto/tls/#pkg-constants' + items: + type: string + type: array + client_ca: + description: Contains the CA certificate for client certificate authentication to the server. + properties: + configMap: + description: ConfigMap containing data to use for the targets. + properties: + key: + description: The key to select. + type: string + name: + description: 'Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?' + type: string + optional: + description: Specify whether the ConfigMap or its key must be defined + type: boolean + required: + - key + type: object + x-kubernetes-map-type: atomic + secret: + description: Secret containing data to use for the targets. + properties: + key: + description: The key of the secret to select from. Must be a valid secret key. + type: string + name: + description: 'Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?' + type: string + optional: + description: Specify whether the Secret or its key must be defined + type: boolean + required: + - key + type: object + x-kubernetes-map-type: atomic + type: object + clientAuthType: + description: 'Server policy for client authentication. Maps to ClientAuth Policies. For more detail on clientAuth options: https://golang.org/pkg/crypto/tls/#ClientAuthType' + type: string + curvePreferences: + description: 'Elliptic curves that will be used in an ECDHE handshake, in preference order. Available curves are documented in the go documentation: https://golang.org/pkg/crypto/tls/#CurveID' + items: + type: string + type: array + keySecret: + description: Secret containing the TLS key for the server. + properties: + key: + description: The key of the secret to select from. Must be a valid secret key. + type: string + name: + description: 'Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?' + type: string + optional: + description: Specify whether the Secret or its key must be defined + type: boolean + required: + - key + type: object + x-kubernetes-map-type: atomic + maxVersion: + description: Maximum TLS version that is acceptable. Defaults to TLS13. + type: string + minVersion: + description: Minimum TLS version that is acceptable. Defaults to TLS12. + type: string + preferServerCipherSuites: + description: Controls whether the server selects the client's most preferred cipher suite, or the server's most preferred cipher suite. If true then the server's preference, as expressed in the order of elements in cipherSuites, is used. + type: boolean + required: + - cert + - keySecret + type: object + type: object type: object status: description: 'Most recent observed status of the ThanosRuler cluster. Read-only. More info: https://github.com/kubernetes/community/blob/master/contributors/devel/sig-architecture/api-conventions.md#spec-and-status'