Bug 1811530
Summary: | Install failed due to mdns record changed | ||
---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | weiwei jiang <wjiang> |
Component: | Etcd Operator | Assignee: | Sam Batschelet <sbatsche> |
Status: | CLOSED DUPLICATE | QA Contact: | ge liu <geliu> |
Severity: | high | Docs Contact: | |
Priority: | high | ||
Version: | 4.4 | CC: | fbrychta, ikarpukh, jzmeskal, m.andre, pprinett, scuppett, smilner, wewang, wsun, yanyang, yprokule |
Target Milestone: | --- | Keywords: | Regression, TestBlocker |
Target Release: | 4.4.0 | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | If docs needed, set a value | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2020-03-11 13:29:46 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | |||
Bug Blocks: | 1809238, 1810490 |
Description
weiwei jiang
2020-03-09 07:05:01 UTC
This is also affecting upstream CI. I'm looking into it. This is likely caused by https://github.com/openshift/cluster-etcd-operator/pull/233 (backported to 4.4 in https://github.com/openshift/cluster-etcd-operator/pull/239), I'll port https://github.com/openshift/machine-config-operator/commit/2908ca449b46200cbed67ae5a465243a7919f144 to openstack, hopefully this is enough to fix our issue. When install OCP in GCP, met the same issue as follow: level=debug msg="Still waiting for the cluster to initialize: Working towards 4.4.0-0.nightly-2020-03-09-234759: 76% complete" level=error msg="Cluster operator authentication Degraded is True with IngressStateEndpoints_MissingSubsets::RouterCerts_NoRouterCertSecret: RouterCertsDegraded: secret/v4-0-config-system-router-certs -n openshift-authentication: could not be retrieved: secret \"v4-0-config-system-router-certs\" not found\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server" level=info msg="Cluster operator authentication Progressing is Unknown with NoData: " level=info msg="Cluster operator authentication Available is Unknown with NoData: " level=error msg="Cluster operator kube-apiserver Degraded is True with InstallerPodContainerWaiting_CreateContainerError::StaticPods_Error: InstallerPodContainerWaitingDegraded: Pod \"installer-2-wewang-vw88w-m-1.c.openshift-qe.internal\" on node \"wewang-vw88w-m-1.c.openshift-qe.internal\" container \"installer\" is waiting for 38m21.732926905s because \"the container name \\\"k8s_installer_installer-2-wewang-vw88w-m-1.c.openshift-qe.internal_openshift-kube-apiserver_243574fe-ebe9-4756-9e7f-6e8a446bf457_0\\\" is already in use by \\\"df46526127e942582cf15846967911f37d4a8db5abd712b0500d561131974176\\\". You have to remove that container to be able to reuse that name.: that name is already in use\"\nStaticPodsDegraded: nodes/wewang-vw88w-m-2.c.openshift-qe.internal pods/kube-apiserver-wewang-vw88w-m-2.c.openshift-qe.internal container=\"kube-apiserver-cert-regeneration-controller\" is not ready\nStaticPodsDegraded: nodes/wewang-vw88w-m-2.c.openshift-qe.internal pods/kube-apiserver-wewang-vw88w-m-2.c.openshift-qe.internal container=\"kube-apiserver-cert-regeneration-controller\" is waiting: \"CrashLoopBackOff\" - \"back-off 5m0s restarting failed container=kube-apiserver-cert-regeneration-controller pod=kube-apiserver-wewang-vw88w-m-2.c.openshift-qe.internal_openshift-kube-apiserver(b3014b7a2f1c6b8515fe65cbb22372bd)\"\nStaticPodsDegraded: pods \"kube-apiserver-wewang-vw88w-m-1.c.openshift-qe.internal\" not found\nStaticPodsDegraded: pods \"kube-apiserver-wewang-vw88w-m-0.c.openshift-qe.internal\" not found" Fyi, About comment 8, the cluster is about ocp in ipi on gcp. Should be fixed with https://github.com/openshift/cluster-kube-apiserver-operator/pull/791 Today set up the cluster successfully for 'IPI on GCP with http_proxy&ovn' against 4.4.0-0.nightly-2020-03-10-194324 I can confirm this issue was also encountered when deploying OPC on RHV with 4.4.0-0.nightly-2020-03-09-175442 *** Bug 1811855 has been marked as a duplicate of this bug. *** *** This bug has been marked as a duplicate of bug 1812071 *** |