Hide Forgot
Created attachment 1894658 [details] must-gather Description of problem: Created a single node cluster with monitoring stack disabled, now the kube-controller-manager operator shows degraded state due to dependency with thanos-querier Version-Release number of selected component (if applicable): 4.11.0-rc.0 Actual results: ```bash [cloud-user@rhel-crcqe /]$ oc get co NAME VERSION AVAILABLE PROGRESSING DEGRADED SINCE MESSAGE authentication 4.11.0-rc.0 True False False 65m config-operator 4.11.0-rc.0 True False False 46h console 4.11.0-rc.0 True False False 66m dns 4.11.0-rc.0 True False False 71m etcd 4.11.0-rc.0 True False False 46h image-registry 4.11.0-rc.0 True False False 66m ingress 4.11.0-rc.0 True False False 46h kube-apiserver 4.11.0-rc.0 True False False 46h kube-controller-manager 4.11.0-rc.0 True False True 46h GarbageCollectorDegraded: error fetching rules: Get "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/rules": dial tcp: lookup thanos-querier.openshift-monitoring.svc on 10.217.4.10:53: no such host kube-scheduler 4.11.0-rc.0 True False False 46h machine-api 4.11.0-rc.0 True False False 46h machine-approver 4.11.0-rc.0 True False False 46h machine-config 4.11.0-rc.0 True False False 46h marketplace 4.11.0-rc.0 True False False 46h network 4.11.0-rc.0 True False False 46h node-tuning 4.11.0-rc.0 True False False 66m openshift-apiserver 4.11.0-rc.0 True False False 65m openshift-controller-manager 4.11.0-rc.0 True False False 67m openshift-samples 4.11.0-rc.0 True False False 46h operator-lifecycle-manager 4.11.0-rc.0 True False False 46h operator-lifecycle-manager-catalog 4.11.0-rc.0 True False False 46h operator-lifecycle-manager-packageserver 4.11.0-rc.0 True False False 71m service-ca 4.11.0-rc.0 True False False 46h ```
Looking at the commits https://github.com/openshift/cluster-kube-controller-manager-operator/pull/623 is what causes it now. This is now blocking CRC because monitoring stack is disabled by default using the CVO overrides https://github.com/code-ready/snc/blob/master/cvo-overrides.yaml
Tested on 4.12.0-0.nightly-2022-07-06-023534 and still got the issue NAME VERSION AVAILABLE PROGRESSING DEGRADED SINCE MESSAGE authentication 4.12.0-0.nightly-2022-07-06-023534 True False False 4h52m config-operator 4.12.0-0.nightly-2022-07-06-023534 True False False 34h console 4.12.0-0.nightly-2022-07-06-023534 True False False 4h58m dns 4.12.0-0.nightly-2022-07-06-023534 True False False 5h2m etcd 4.12.0-0.nightly-2022-07-06-023534 True False False 34h image-registry 4.12.0-0.nightly-2022-07-06-023534 True False False 4h57m ingress 4.12.0-0.nightly-2022-07-06-023534 True False False 34h kube-apiserver 4.12.0-0.nightly-2022-07-06-023534 True False False 34h kube-controller-manager 4.12.0-0.nightly-2022-07-06-023534 True False True 34h GarbageCollectorDegraded: error fetching rules: Get "https://thanos-querier.openshift-monitoring.svc:9091/api/v1/rules": dial tcp: lookup thanos-querier.openshift-monitoring.svc on 10.217.4.10:53: no such host kube-scheduler 4.12.0-0.nightly-2022-07-06-023534 True False False 34h machine-api 4.12.0-0.nightly-2022-07-06-023534 True False False 34h machine-approver 4.12.0-0.nightly-2022-07-06-023534 True False False 34h machine-config 4.12.0-0.nightly-2022-07-06-023534 True False False 34h marketplace 4.12.0-0.nightly-2022-07-06-023534 True False False 34h network 4.12.0-0.nightly-2022-07-06-023534 True False False 34h node-tuning 4.12.0-0.nightly-2022-07-06-023534 True False False 4h56m openshift-apiserver 4.12.0-0.nightly-2022-07-06-023534 True False False 4h58m openshift-controller-manager 4.12.0-0.nightly-2022-07-06-023534 True False False 4h57m openshift-samples 4.12.0-0.nightly-2022-07-06-023534 True False False 34h operator-lifecycle-manager 4.12.0-0.nightly-2022-07-06-023534 True False False 34h operator-lifecycle-manager-catalog 4.12.0-0.nightly-2022-07-06-023534 True False False 34h operator-lifecycle-manager-packageserver 4.12.0-0.nightly-2022-07-06-023534 True False False 5h2m service-ca 4.12.0-0.nightly-2022-07-06-023534 True False False 34h
I tested the 4.12.0-0.nightly-2022-07-06-221008 and it is now not showing degraded state. ``` $ oc get co NAME VERSION AVAILABLE PROGRESSING DEGRADED SINCE MESSAGE authentication 4.12.0-0.nightly-2022-07-06-221008 True False False 6m37s config-operator 4.12.0-0.nightly-2022-07-06-221008 True False False 26h console 4.12.0-0.nightly-2022-07-06-221008 True False False 10m dns 4.12.0-0.nightly-2022-07-06-221008 True False False 15m etcd 4.12.0-0.nightly-2022-07-06-221008 True False False 26h image-registry 4.12.0-0.nightly-2022-07-06-221008 True False False 10m ingress 4.12.0-0.nightly-2022-07-06-221008 True False False 26h kube-apiserver 4.12.0-0.nightly-2022-07-06-221008 True False False 26h kube-controller-manager 4.12.0-0.nightly-2022-07-06-221008 True False False 26h kube-scheduler 4.12.0-0.nightly-2022-07-06-221008 True False False 26h machine-api 4.12.0-0.nightly-2022-07-06-221008 True False False 26h machine-approver 4.12.0-0.nightly-2022-07-06-221008 True False False 26h machine-config 4.12.0-0.nightly-2022-07-06-221008 True False False 26h marketplace 4.12.0-0.nightly-2022-07-06-221008 True False False 26h network 4.12.0-0.nightly-2022-07-06-221008 True False False 26h node-tuning 4.12.0-0.nightly-2022-07-06-221008 True False False 10m openshift-apiserver 4.12.0-0.nightly-2022-07-06-221008 True False False 10m openshift-controller-manager 4.12.0-0.nightly-2022-07-06-221008 True False False 11m openshift-samples 4.12.0-0.nightly-2022-07-06-221008 True False False 26h operator-lifecycle-manager 4.12.0-0.nightly-2022-07-06-221008 True False False 26h operator-lifecycle-manager-catalog 4.12.0-0.nightly-2022-07-06-221008 True False False 26h operator-lifecycle-manager-packageserver 4.12.0-0.nightly-2022-07-06-221008 True False False 16m service-ca 4.12.0-0.nightly-2022-07-06-221008 True False False 26h ```
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: OpenShift Container Platform 4.12.0 bug fix and security update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2022:7399