Bug 1699456
Summary: | bootstrap-kube-apiserver failing to connect to etcd with invalid cert error | ||||||
---|---|---|---|---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | Mike Fiedler <mifiedle> | ||||
Component: | Etcd | Assignee: | Sam Batschelet <sbatsche> | ||||
Status: | CLOSED DUPLICATE | QA Contact: | ge liu <geliu> | ||||
Severity: | high | Docs Contact: | |||||
Priority: | unspecified | ||||||
Version: | 4.1.0 | CC: | aos-bugs, deads, eparis, gblomqui, jiajliu, jialiu, jokerman, mfojtik, mmccomas, wking | ||||
Target Milestone: | --- | Keywords: | BetaBlocker, Reopened | ||||
Target Release: | 4.1.0 | ||||||
Hardware: | Unspecified | ||||||
OS: | Unspecified | ||||||
Whiteboard: | |||||||
Fixed In Version: | Doc Type: | If docs needed, set a value | |||||
Doc Text: | Story Points: | --- | |||||
Clone Of: | Environment: | ||||||
Last Closed: | 2019-05-02 13:31:17 UTC | Type: | Bug | ||||
Regression: | --- | Mount Type: | --- | ||||
Documentation: | --- | CRM: | |||||
Verified Versions: | Category: | --- | |||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
Cloudforms Team: | --- | Target Upstream Version: | |||||
Embargoed: | |||||||
Attachments: |
|
Description
Mike Fiedler
2019-04-12 18:55:10 UTC
Still hit the etcd issue on 4.1.0-0.nightly-2019-04-29-235412 when upi installation on vmware. refer to https://bugzilla.redhat.com/show_bug.cgi?id=1698456#c14 bootkube.sh log Apr 30 09:12:40 bootstrap-0 bootkube.sh[11321]: E0430 09:12:40.713004 1 reflector.go:134] github.com/openshift/cluster-bootstrap/pkg/start/status.go:66: Failed to lis> Apr 30 09:31:50 bootstrap-0 bootkube.sh[11321]: Error: error while checking pod status: timed out waiting for the condition Apr 30 09:31:50 bootstrap-0 bootkube.sh[11321]: Tearing down temporary bootstrap control plane... Apr 30 09:31:50 bootstrap-0 bootkube.sh[11321]: Error: error while checking pod status: timed out waiting for the condition Apr 30 09:31:50 bootstrap-0 bootkube.sh[11321]: unable to find container etcd-signer: no container with name or ID etcd-signer found: no such container Apr 30 09:31:50 bootstrap-0 systemd[1]: bootkube.service: Main process exited, code=exited, status=125/n/a Apr 30 09:31:50 bootstrap-0 systemd[1]: bootkube.service: Failed with result 'exit-code'. Apr 30 09:31:55 bootstrap-0 systemd[1]: bootkube.service: Service RestartSec=5s expired, scheduling restart. Apr 30 09:31:55 bootstrap-0 systemd[1]: bootkube.service: Scheduled restart job, restart counter is at 2. Apr 30 09:31:55 bootstrap-0 systemd[1]: Stopped Bootstrap a Kubernetes cluster. Apr 30 09:31:55 bootstrap-0 systemd[1]: Started Bootstrap a Kubernetes cluster. ... Apr 30 09:33:00 bootstrap-0 bootkube.sh[16365]: E0430 09:33:00.493809 1 reflector.go:251] github.com/openshift/cluster-bootstrap/pkg/start/status.go:66: Failed to watch *v1.Pod: Get https://api.jliu-demo.qe.devcluster.openshift.com:6443/api/v1/pods?watch=true: dial tcp 139.178.89.199:6443: connect: connection refused Apr 30 09:33:01 bootstrap-0 bootkube.sh[16365]: E0430 09:33:01.527050 1 reflector.go:134] github.com/openshift/cluster-bootstrap/pkg/start/status.go:66: Failed to list *v1.Pod: Get https://api.jliu-demo.qe.devcluster.openshift.com:6443/api/v1/pods: dial tcp 139.178.89.198:6443: connect: connection refused Since the issue in bug 1698456 has been fixed, So re-open this bug to track the installation issue. > Still hit the etcd issue on 4.1.0-0.nightly-2019-04-29-235412 when upi installation on vmware.
The issue reported was a cert SAN issue but the above logs are regarding a failing signer container. Did you see errors like reported? If yes can you please share them? Thanks!
Jia Liu, the log in comment #4 looks like a new issue related to bootkube restarting. I can't tell by the logs provided what's causing bootkube to restart. Please open a new issue against the installer component to track the bootkube restart. The team will likely need more context from the logs to track down what's happening there. I'm going to go ahead and re-close this as a dup. Johnny, please feel free to open a new bug with the new (we believe unrelated) information in #4. *** This bug has been marked as a duplicate of bug 1698456 *** I'm ok to close this one. But since it's not a 100% re-produce issue, QE will file a new bug for it if hit it again. |