Description of problem: Prometheus is gather metrics for the pods which are in completed/succeeded state. prometheus is producing loads of logs within a matter of minutes and elasticsearch is filling up exponentially. The errors are not directly harmful to the environment but it should not be reported by prometheus. 2021-02-02T10:22:42.982873977Z E0202 10:22:42.982863 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-16-ip-10-0-156-87.eu-central-1.compute.internal, skipping 2021-02-02T10:22:42.982873977Z E0202 10:22:42.982870 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-16-ip-10-0-162-147.eu-central-1.compute.internal, skipping 2021-02-02T10:22:42.982881426Z E0202 10:22:42.982875 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-16-ip-10-0-202-64.eu-central-1.compute.internal, skipping 2021-02-02T10:22:42.982881426Z E0202 10:22:42.982879 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-17-ip-10-0-156-87.eu-central-1.compute.internal, skipping 2021-02-02T10:22:42.982889626Z E0202 10:22:42.982884 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-17-ip-10-0-162-147.eu-central-1.compute.internal, skipping Version-Release number of selected component (if applicable): 4.6.8 How reproducible: NA Steps to Reproduce: 1. 2. 3. Actual results: 2021-02-02T10:22:42.982873977Z E0202 10:22:42.982863 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-16-ip-10-0-156-87.eu-central-1.compute.internal, skipping Expected results: No error messages should be there for completed/succeeded pods. Additional info:
Just to note - we have merged a fix for this in an upstream project in https://github.com/kubernetes-monitoring/kubernetes-mixin/pull/552. It should trickle down into OpenShift 4.8.
encounter the issue in 4.6 cluster [hongyli@hongyli-fed Downloads]$ oc -n openshift-monitoring logs prometheus-adapter-7c46fc8dfb-4xt9j I0302 05:14:29.793955 1 adapter.go:94] successfully using in-cluster auth I0302 05:14:31.593087 1 dynamic_cafile_content.go:166] Starting request-header::/etc/tls/private/requestheader-client-ca-file I0302 05:14:31.593139 1 dynamic_cafile_content.go:166] Starting client-ca-bundle::/etc/tls/private/client-ca-file I0302 05:14:31.593895 1 dynamic_serving_content.go:129] Starting serving-cert::/etc/tls/private/tls.crt::/etc/tls/private/tls.key I0302 05:14:31.595559 1 secure_serving.go:178] Serving securely on [::]:6443 I0302 05:14:31.595875 1 tlsconfig.go:219] Starting DynamicServingCertificateController E0302 05:19:44.698614 1 authentication.go:104] Unable to authenticate the request due to an error: [invalid bearer token, [invalid bearer token, square/go-jose: error in cryptographic primitive, token lookup failed]] E0302 05:20:14.695671 1 authentication.go:104] Unable to authenticate the request due to an error: [invalid bearer token, [invalid bearer token, square/go-jose: error in cryptographic primitive, token lookup failed]] E0302 05:20:44.716807 1 authentication.go:104] Unable to authenticate the request due to an error: [invalid bearer token, [invalid bearer token, square/go-jose: error in cryptographic primitive, token lookup failed]] E0302 05:21:14.694884 1 authentication.go:104] Unable to authenticate the request due to an error: [invalid bearer token, [invalid bearer token, square/go-jose: error in cryptographic primitive, token lookup failed]] E0302 05:21:44.695863 1 authentication.go:104] Unable to authenticate the request due to an error: [invalid bearer token, [invalid bearer token, square/go-jose: error in cryptographic primitive, token lookup failed]] E0302 05:22:14.692911 1 authentication.go:104] Unable to authenticate the request due to an error: [invalid bearer token, [invalid bearer token, square/go-jose: error in cryptographic primitive, token lookup failed]] E0302 06:57:35.600598 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/installer-2-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600626 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/installer-2-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600634 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/installer-2-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600641 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/installer-3-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600650 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/installer-3-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600657 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/installer-3-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600664 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/revision-pruner-2-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600671 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/revision-pruner-2-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600680 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/revision-pruner-2-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600692 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/revision-pruner-3-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600704 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/revision-pruner-3-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600715 1 provider.go:187] unable to fetch CPU metrics for pod openshift-etcd/revision-pruner-3-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600763 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/installer-2-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600771 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/installer-2-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600778 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/installer-2-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600785 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/installer-3-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600792 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/installer-4-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600799 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/installer-5-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600806 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/installer-5-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600813 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/installer-5-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600849 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-2-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600865 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-2-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600872 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-2-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600884 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-3-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600896 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-4-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600909 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-5-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600921 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-5-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600933 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-apiserver/revision-pruner-5-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600956 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/installer-4-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600968 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/installer-4-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600977 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/installer-4-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.600985 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/installer-5-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.600992 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/installer-5-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.600999 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/installer-5-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601006 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/installer-6-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601013 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/installer-6-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601020 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/installer-6-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601044 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/revision-pruner-4-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601065 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/revision-pruner-4-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601077 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/revision-pruner-4-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601118 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/revision-pruner-5-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601153 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/revision-pruner-5-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601179 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/revision-pruner-5-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601192 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/revision-pruner-6-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601206 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/revision-pruner-6-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601220 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-controller-manager/revision-pruner-6-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601237 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-2-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601249 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-3-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601260 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-4-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601272 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-5-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601285 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-6-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601298 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-6-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601333 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-6-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601356 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-7-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601369 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-7-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601408 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/installer-7-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601472 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-2-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601517 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-3-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601530 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-4-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601543 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-5-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601556 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-6-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601570 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-6-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601583 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-6-ip-10-0-222-158.us-west-2.compute.internal, skipping E0302 06:57:35.601594 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-7-ip-10-0-132-218.us-west-2.compute.internal, skipping E0302 06:57:35.601607 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-7-ip-10-0-142-172.us-west-2.compute.internal, skipping E0302 06:57:35.601619 1 provider.go:187] unable to fetch CPU metrics for pod openshift-kube-scheduler/revision-pruner-7-ip-10-0-222-158.us-west-2.compute.internal, skipping [hongyli@hongyli-fed Downloads]$ oc get co NAME VERSION AVAILABLE PROGRESSING DEGRADED SINCE authentication 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 100m cloud-credential 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 124m cluster-autoscaler 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 118m config-operator 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 119m console 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 104m csi-snapshot-controller 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 119m dns 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 118m etcd 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 117m image-registry 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 110m ingress 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 110m insights 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 119m kube-apiserver 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 117m kube-controller-manager 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 117m kube-scheduler 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 116m kube-storage-version-migrator 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 109m machine-api 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 114m machine-approver 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 118m machine-config 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 117m marketplace 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 117m monitoring 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 104m network 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 120m node-tuning 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 119m openshift-apiserver 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 114m openshift-controller-manager 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 117m openshift-samples 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 114m operator-lifecycle-manager 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 118m operator-lifecycle-manager-catalog 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 118m operator-lifecycle-manager-packageserver 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 114m service-ca 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 119m storage 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt True False False 118m
The the prometheus adapter pod has no issue oc -n openshift-monitoring logs prometheus-adapter-7c46fc8dfb-rvzzh I0302 05:14:29.794739 1 adapter.go:94] successfully using in-cluster auth I0302 05:14:31.215270 1 dynamic_cafile_content.go:166] Starting request-header::/etc/tls/private/requestheader-client-ca-file I0302 05:14:31.215337 1 dynamic_cafile_content.go:166] Starting client-ca-bundle::/etc/tls/private/client-ca-file I0302 05:14:31.215795 1 dynamic_serving_content.go:129] Starting serving-cert::/etc/tls/private/tls.crt::/etc/tls/private/tls.key I0302 05:14:31.217684 1 secure_serving.go:178] Serving securely on [::]:6443 I0302 05:14:31.221153 1 tlsconfig.go:219] Starting DynamicServingCertificateController
Saw issue in comments 12 on the cluster with payload 4.6.0-0.ci.test-2021-03-02-045033-ci-ln-piym2wt and PR openshift/node_exporter#83
The fix has been merged upstream and will be brought downstream with the next release of prometheus-adapter. It will most likely happen during the 4.9 timeframe.
Moving to MODIFIED state since https://github.com/openshift/cluster-monitoring-operator/pull/1325 and https://github.com/openshift/k8s-prometheus-adapter/pull/53 have been merged.
is the Target Release 4.9.0 or 4.10.0, I see the fix is in 4.9 and 4.10
Target release is 4.9.0. For some reason, the bug wasn't moved from MODIFIED to ON_QA automatically as I would have expected.
checked with 4.9.0-0.nightly-2021-09-08-233235,no error for the completed pods # oc -n openshift-kube-scheduler get pod NAME READY STATUS RESTARTS AGE installer-2-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 148m installer-3-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 148m installer-4-master-00.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 146m installer-4-master-01.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 145m installer-4-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 147m installer-5-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 143m installer-6-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 142m installer-7-master-00.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 140m installer-7-master-01.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 139m installer-7-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 141m openshift-kube-scheduler-master-00.pamoedo-dualtest23.qe.devcluster.openshift.com 3/3 Running 0 140m openshift-kube-scheduler-master-01.pamoedo-dualtest23.qe.devcluster.openshift.com 3/3 Running 0 139m openshift-kube-scheduler-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 3/3 Running 0 141m revision-pruner-4-master-00.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 145m revision-pruner-4-master-01.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 143m revision-pruner-4-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 146m revision-pruner-7-master-00.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 139m revision-pruner-7-master-01.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 138m revision-pruner-7-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 0/1 Completed 0 140m # oc adm top pod -n openshift-kube-scheduler NAME CPU(cores) MEMORY(bytes) openshift-kube-scheduler-master-00.pamoedo-dualtest23.qe.devcluster.openshift.com 5m 127Mi openshift-kube-scheduler-master-01.pamoedo-dualtest23.qe.devcluster.openshift.com 3m 141Mi openshift-kube-scheduler-master-02.pamoedo-dualtest23.qe.devcluster.openshift.com 4m 156Mi # oc -n openshift-monitoring logs prometheus-adapter-bc6cc65df-64vxk I0909 08:41:28.058691 1 adapter.go:108] successfully using in-cluster auth I0909 08:41:28.713173 1 dynamic_cafile_content.go:155] "Starting controller" name="request-header::/etc/tls/private/requestheader-client-ca-file" I0909 08:41:28.713195 1 dynamic_cafile_content.go:155] "Starting controller" name="client-ca-bundle::/etc/tls/private/client-ca-file" I0909 08:41:28.714249 1 dynamic_serving_content.go:129] "Starting controller" name="serving-cert::/etc/tls/private/tls.crt::/etc/tls/private/tls.key" I0909 08:41:28.715124 1 secure_serving.go:256] Serving securely on [::]:6443 I0909 08:41:28.715161 1 tlsconfig.go:240] "Starting DynamicServingCertificateController"
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: OpenShift Container Platform 4.9.0 bug fix and security update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:3759