Description of problem: [root@wj45ios1015a-xtq4n-master-2 core]# journalctl -f -u kubelet|grep -i kubelet_node_status Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.305412 362998 kubelet_node_status.go:294] Setting node annotation to enable volume controller attach/detach Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.730980 362998 kubelet_node_status.go:342] Adding node label from cloud provider: beta.kubernetes.io/instance-type=m1.xlarge Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.731009 362998 kubelet_node_status.go:344] Adding node label from cloud provider: node.kubernetes.io/instance-type=m1.xlarge Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.731050 362998 kubelet_node_status.go:355] Adding node label from cloud provider: failure-domain.beta.kubernetes.io/zone=nova Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.731057 362998 kubelet_node_status.go:357] Adding node label from cloud provider: topology.kubernetes.io/zone=nova Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.731062 362998 kubelet_node_status.go:361] Adding node label from cloud provider: failure-domain.beta.kubernetes.io/region=regionOne Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.731066 362998 kubelet_node_status.go:363] Adding node label from cloud provider: topology.kubernetes.io/region=regionOne Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.741407 362998 kubelet_node_status.go:486] Recording NodeHasSufficientMemory event message for node wj45ios1015a-xtq4n-master-2 Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.741431 362998 kubelet_node_status.go:486] Recording NodeHasNoDiskPressure event message for node wj45ios1015a-xtq4n-master-2 Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.741445 362998 kubelet_node_status.go:486] Recording NodeHasSufficientPID event message for node wj45ios1015a-xtq4n-master-2 Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: I1015 05:27:51.741476 362998 kubelet_node_status.go:70] Attempting to register node wj45ios1015a-xtq4n-master-2 Oct 15 05:27:51 wj45ios1015a-xtq4n-master-2 hyperkube[362998]: E1015 05:27:51.742645 362998 kubelet_node_status.go:92] Unable to register node "wj45ios1015a-xtq4n-master-2" with API server: Post https://api-int.wj45ios1015a.1015-r-b.qe.rhcloud.com:6443/api/v1/nodes: dial tcp 192.168.0.5:6443: connect: connection refused Version-Release number of selected component (if applicable): 4.5.0-0.nightly-2020-10-14-095539 How reproducible: Always? Steps to Reproduce: 1. Setup one IPI on OSP cluster 2. Check if all master nodes are ready 3. Actual results: $ oc get nodes -o wide NAME STATUS ROLES AGE VERSION INTERNAL-IP EXTERNAL-IP OS-IMAGE KERNEL-VERSION CONTAINER-RUNTIME wj45ios1015a-xtq4n-master-0 Ready master 170m v1.18.3+2fbd7c7 192.168.2.45 <none> Red Hat Enterprise Linux CoreOS 45.82.202010091130-0 (Ootpa) 4.18.0-193.24.1.el8_2.dt1.x86_64 cri-o://1.18.3-19.rhaos4.5.git9264b 4f.el8 wj45ios1015a-xtq4n-master-1 Ready master 170m v1.18.3+2fbd7c7 192.168.0.248 <none> Red Hat Enterprise Linux CoreOS 45.82.202010091130-0 (Ootpa) 4.18.0-193.24.1.el8_2.dt1.x86_64 cri-o://1.18.3-19.rhaos4.5.git9264b 4f.el8 wj45ios1015a-xtq4n-master-2 NotReady master 170m v1.18.3+2fbd7c7 192.168.2.193 <none> Red Hat Enterprise Linux CoreOS 45.82.202010091130-0 (Ootpa) 4.18.0-193.24.1.el8_2.dt1.x86_64 cri-o://1.18.3-19.rhaos4.5.git9264b 4f.el8 Expected results: All master nodes should be ready without error. Additional info:
FYI. The expected progress should be the following: Kubelet -> preroute iptables change target port from 6443 to 9445(haproxy-monitor) -> haproxy(listen 9445 port) -> dispatch traffic to the backends based on healthy status(should be one of the masters). The 192.168.0.5(VIP, which is managed by keepalived) is assigned to one of the masters. in my scenario, it's master-2.
*** This bug has been marked as a duplicate of bug 1888301 ***