BZ#2002552 was backported to 4.8, but we still have instances of this failing on both IPv4 and IPv6. Example runs where this failed: https://prow.ci.openshift.org/view/gs/origin-ci-test/logs/periodic-ci-openshift-release-master-nightly-4.8-e2e-metal-ipi-ovn-ipv6/1453456250908971008 https://prow.ci.openshift.org/view/gs/origin-ci-test/logs/periodic-ci-openshift-release-master-nightly-4.8-e2e-metal-ipi/1453425238455881728
By looking at the sparse test failures, the oc debug command - triggered for each master found - failed intermittently due the unavailability of the kube-apiserver for that particular master. By cross-checking the test timestamps failures with the kube-apiserver events then it appeared that the server gets several restarts, and sometimes the test tries to execute command when the server it's not yet ready. Even though the test implemented a retry mechanism, it's not enough to cover the restart gaps. For example, in the 1453817495201779712 job [1], the debug command failed for master-1.ostest.test.metalkube.org at 21:36:52.652: > STEP: Testing master node master-1.ostest.test.metalkube.org > Oct 28 21:36:50.073: INFO: Verifying kubeconfig "localhost-recovery.kubeconfig" on master "master-1.ostest.test.metalkube.org" > Oct 28 21:36:50.073: INFO: Running 'oc --namespace=e2e-test-apiserver-rvg62 --kubeconfig=/tmp/secret/kubeconfig debug node/master-1.ostest.test.metalkube.org -- chroot /host /bin/bash -euxo pipefail -c oc --kubeconfig "/etc/kubernetes/static-pod-resources/kube-apiserver-certs/secrets/node-kubeconfigs/localhost-recovery.kubeconfig" get namespace > kube-system' > Oct 28 21:36:52.652: INFO: Error running /usr/bin/oc --namespace=e2e-test-apiserver-rvg62 --kubeconfig=/tmp/secret/kubeconfig debug node/master-1.ostest.test.metalkube.org -- chroot /host /bin/bash -euxo pipefail -c oc --kubeconfig "/etc/kubernetes/static-pod-resources/kube-apiserver-certs/secrets/node-kubeconfigs/localhost-recovery.kubeconfig" > get namespace kube-system: > StdOut> > Starting pod/master-1ostesttestmetalkubeorg-debug ... > To use host binaries, run `chroot /host` > + oc --kubeconfig /etc/kubernetes/static-pod-resources/kube-apiserver-certs/secrets/node-kubeconfigs/localhost-recovery.kubeconfig get namespace kube-system > The connection to the server localhost:6443 was refused - did you specify the right host or port? master-1.ostest.test.metalkube.org kube-apiserver logs [2] shows that last server restart was on 21:38:15.058617 (last kill event happened at 21:35:03, see [3]) Same behavior has been observed also in other job failures for the same tests. To prevent the tests being triggered too early, we're introducing a first waiting condition in the metal-ipi jobs [4] (and it should address also https://bugzilla.redhat.com/show_bug.cgi?id=2018208) [1] https://prow.ci.openshift.org/view/gs/origin-ci-test/logs/periodic-ci-openshift-release-master-nightly-4.8-e2e-metal-ipi-ovn-ipv6/1453817495201779712 [2] https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/logs/periodic-ci-openshift-release-master-nightly-4.8-e2e-metal-ipi-ovn-ipv6/1453817495201779712/artifacts/e2e-metal-ipi-ovn-ipv6/gather-extra/artifacts/pods/openshift-kube-apiserver_kube-apiserver-master-1.ostest.test.metalkube.org_kube-apiserver.log [3] https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/logs/periodic-ci-openshift-release-master-nightly-4.8-e2e-metal-ipi-ovn-ipv6/1453817495201779712/artifacts/e2e-metal-ipi-ovn-ipv6/gather-must-gather/artifacts/event-filter.html [4] https://github.com/openshift/release/pull/23131
From looking at https://testgrid.k8s.io/redhat-openshift-ocp-release-4.8-blocking#periodic-ci-openshift-release-master-nightly-4.8-e2e-metal-ipi-ovn-ipv6 I see this is not super-high pressing issue so I'm lowering the priority on this one.
This bug hasn't had any activity in the last 30 days. Maybe the problem got resolved, was a duplicate of something else, or became less pressing for some reason - or maybe it's still relevant but just hasn't been looked at yet. As such, we're marking this bug as "LifecycleStale" and decreasing the severity/priority. If you have further information on the current state of the bug, please update it, otherwise this bug can be closed in about 7 days. The information can be, for example, that the problem still occurs, that you still want the feature, that more information is needed, or that the bug is (for whatever reason) no longer relevant. Additionally, you can add LifecycleFrozen into Whiteboard if you think this bug should never be marked as stale. Please consult with bug assignee before you do that.
From looking at https://testgrid.k8s.io/redhat-openshift-ocp-release-4.8-blocking#periodic-ci-openshift-release-master-nightly-4.8-e2e-metal-ipi-ovn-ipv6 the test seems pretty green so I'll move this over to qa.
The LifecycleStale keyword was removed because the bug moved to QE. The bug assignee was notified.
Checked the kubeconfig related tests from https://testgrid.k8s.io/redhat-openshift-ocp-release-4.8-blocking#periodic-ci-openshift-release-master-nightly-4.8-e2e-metal-ipi-ovn-ipv6 and https://testgrid.k8s.io/redhat-openshift-ocp-release-4.8-blocking#periodic-ci-openshift-release-master-nightly-4.8-e2e-metal-ipi, they were passed during the past month, so move the bug VERIFIED.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (OpenShift Container Platform 4.8.26 bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2022:0021