Bug 1934782
Summary: | Unable to apply 4.7.0: SDN unavailable on upgrade – Was: the cluster operator kube-storage-version-migrator has not yet successfully rolled out | |||
---|---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | mchebbi <mchebbi> | |
Component: | Networking | Assignee: | jamo luhrsen <jluhrsen> | |
Networking sub component: | openshift-sdn | QA Contact: | zhaozhanqi <zzhao> | |
Status: | CLOSED DUPLICATE | Docs Contact: | ||
Severity: | high | |||
Priority: | high | CC: | anbhat, aos-bugs, aprajapa, nnosenzo, sanchezl, sttts | |
Version: | 4.7 | |||
Target Milestone: | --- | |||
Target Release: | 4.8.0 | |||
Hardware: | Unspecified | |||
OS: | Unspecified | |||
Whiteboard: | ||||
Fixed In Version: | Doc Type: | If docs needed, set a value | ||
Doc Text: | Story Points: | --- | ||
Clone Of: | ||||
: | 1941740 (view as bug list) | Environment: | ||
Last Closed: | 2021-03-23 17:57:32 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | ||||
Bug Blocks: | 1941740 |
Description
mchebbi@redhat.com
2021-03-03 20:02:18 UTC
Kube-storage-migration-operator is the smallest of your problems. It's on the far end of the root cause chain. dial tcp 172.30.0.1:443: i/o timeout This suggests that networking is broken. (In reply to Stefan Schimanski from comment #1) > Kube-storage-migration-operator is the smallest of your problems. It's on > the far end of the root cause chain. > > dial tcp 172.30.0.1:443: i/o timeout > > This suggests that networking is broken. Thanks Stefan for your feedback. Could you tell me how to fix the issue. Thanks in advance for your help. Just adding a bit more info about Moez case. openshift-sdn: ~~~ NAME READY STATUS RESTARTS AGE ovs-455qw 1/1 Running 0 5d ovs-5dntd 0/1 Running 1544 6d ovs-8h26z 0/1 Running 1544 5d ovs-crxx4 0/1 Running 1546 6d ovs-lft66 0/1 Running 1545 6d ovs-lwbl9 0/1 Running 1545 7d ovs-nd5bw 0/1 Running 1544 6d ovs-nwz7k 0/1 Running 1545 6d ovs-s5rcz 1/1 Running 0 5d ovs-trvqd 1/1 Running 0 5d ovs-v2png 0/1 Running 1544 7d sdn-2ddbk 2/2 Running 0 7d sdn-72w2m 1/2 Running 1289 7d sdn-c2d6h 1/2 Running 1289 7d sdn-controller-kw9w2 1/1 Running 0 7d sdn-controller-qj6hd 1/1 Running 0 7d sdn-controller-rzt4f 1/1 Running 0 7d sdn-g9phv 1/2 Running 1288 7d sdn-gsptx 1/2 Running 1289 7d sdn-h2m76 2/2 Running 0 7d sdn-m5pht 1/2 Running 1288 7d sdn-r7bsz 2/2 Running 0 7d sdn-sg5ml 1/2 Running 1288 7d sdn-vbj7l 1/2 Running 1288 7d sdn-wsmvs 1/2 Running 1289 7d ~~~ SDN pods failure message: """ 2021-03-03T10:08:36.955546539-07:00 I0303 17:08:36.955476 14072 healthcheck.go:42] waiting for OVS to start: dial unix /var/run/openvswitch/db.sock: connect: no such file or directory 2021-03-03T10:08:36.955546539-07:00 F0303 17:08:36.955499 14072 cmd.go:111] Failed to start sdn: node SDN setup failed: timed out waiting for the condition """ OVS pods error: """ id: openvswitch: no such user """ This issue seems to be reported in ticket [0], which was linked to a systemd bug ([1]), I have asked the CU to restart one of the failing nodes and see if that solves the issue. [0] https://bugzilla.redhat.com/show_bug.cgi?id=1887040 [1] https://bugzilla.redhat.com/show_bug.cgi?id=1888017 *** This bug has been marked as a duplicate of bug 1907353 *** |