The test case: [sig-cluster-lifecycle] TestAdminAck should succeed [Suite:openshift/conformance/parallel] is vulnerable to brief API-server hiccups like [1]: Dec 6 00:00:10.440: FAIL: Error accessing configmap openshift-config-managed/admin-gates: Get "https://api.ci-op-g2m38jp7-eafe9.origin-ci-int-aws.dev.rhcloud.com:6443/api/v1/namespaces/openshift-config-managed/configmaps/admin-gates": dial tcp: lookup api.ci-op-g2m38jp7-eafe9.origin-ci-int-aws.dev.rhcloud.com on 172.30.0.10:53: no such host and [2]: Dec 9 19:53:20.747: FAIL: Error accessing configmap openshift-config-managed/admin-gates: Get "https://api.ci-op-w5q90zpi-9278e.origin-ci-int-aws.dev.rhcloud.com:6443/api/v1/namespaces/openshift-config-managed/configmaps/admin-gates": dial tcp 100.21.251.165:6443: i/o timeout We should... do something to make those non-fatal. Logging the error and then bailing out to wait for the next poll round might work, but we want to ensure that we actually get a successful run and don't claim "success" if all our attempts were "I couldn't actually connect to the Kube API-server to check". [1]: https://bugzilla.redhat.com/show_bug.cgi?id=2026806#c8 [2]: https://bugzilla.redhat.com/show_bug.cgi?id=2027929#c1
Moving back to NEW, because I haven't had time to work on it. Leaving myself in as the assignee, because I don't want to dump fixing this on the Test Framework folks.
[1] and [2] haven't flaked this test-case since the fix landed, so yup, marking complete. [1]: https://testgrid.k8s.io/redhat-openshift-ocp-release-4.11-informing#periodic-ci-openshift-release-master-ci-4.11-e2e-aws-upgrade-single-node [2]: https://testgrid.k8s.io/redhat-openshift-ocp-release-4.11-informing#periodic-ci-openshift-release-master-ci-4.11-e2e-azure-upgrade-single-node