Bug 1960278
Summary: | alert KubePodCrashLooping: kube-controller-manager-recovery-controller | ||
---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | Petr Muller <pmuller> |
Component: | kube-controller-manager | Assignee: | Maciej Szulik <maszulik> |
Status: | CLOSED DUPLICATE | QA Contact: | zhou ying <yinzhou> |
Severity: | medium | Docs Contact: | |
Priority: | unspecified | ||
Version: | 4.8 | CC: | aos-bugs, mfojtik |
Target Milestone: | --- | ||
Target Release: | --- | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | If docs needed, set a value | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: |
test: openshift-tests.[sig-instrumentation][Late] Alerts shouldn't report any alerts in firing or pending state apart from Watchdog and AlertmanagerReceiversNotConfigured and have no gaps in Watchdog firing [Suite:openshift/conformance/parallel]
|
|
Last Closed: | 2021-05-17 08:53:58 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Petr Muller
2021-05-13 13:51:58 UTC
Not 100% sure if this related to https://bugzilla.redhat.com/show_bug.cgi?id=1958974 "kube-scheduler-recovery-controller is reported as crashlooping in 4.8 on about 8% of multiple types of runs" or not - the crashlooping thing seems to be different, so leaving this as a separate bug for now. Yeah, this looks very similar looking at the pod logs: 2021-05-08T21:28:45.715477293Z E0508 21:28:45.715424 1 csrcontroller.go:146] key failed with : Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager-operator/configmaps/csr-signer-ca": dial tcp [::1]:6443: connect: connection refused 2021-05-08T21:28:50.393085920Z I0508 21:28:50.393009 1 leaderelection.go:278] failed to renew lease openshift-kube-controller-manager/cert-recovery-controller-lock: timed out waiting for the condition 2021-05-08T21:28:50.393179604Z E0508 21:28:50.393155 1 leaderelection.go:301] Failed to release lock: resource name may not be empty 2021-05-08T21:28:50.393209737Z W0508 21:28:50.393186 1 leaderelection.go:75] leader election lost since the investigation is an ongoing, I'll close that as duplicate of the other one. *** This bug has been marked as a duplicate of bug 1948311 *** |