Hide Forgot
It's currently maxUnavailable: 1, but we want maxUnavailable: 10%, so we scale better on clusters with large node-counts. Checking a recent 4.8 CI release: $ curl -s https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/logs/release-openshift-origin-installer-e2e-gcp-4.8/1364411741282242560/artifacts/e2e-gcp/daemonsets.json | jq -r '.items[] | select(.spec.template.spec.nodeSelector["node-role.kubernetes.io/master"] != "" and .spec.updateStrategy.rollingUpdate.maxUnavailable != "10%") | .metadata.namespace + " " + .metadata.name + " " + (.spec.template.spec.nodeSelector | tostring) + " " + (.spec.updateStrategy | tostring)' ... openshift-sdn sdn {"kubernetes.io/os":"linux"} {"rollingUpdate":{"maxUnavailable":1},"type":"RollingUpdate"} Like bug 1933159, but different DaemonSet.
Verified this on 4.8.0-0.nightly-2021-03-08-184701
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: OpenShift Container Platform 4.8.2 bug fix and security update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:2438
This is not my component, so it's not my decision about whether or not to backport to 4.7. Clearing my NEEDINFO.
from what I can tell this did not get cherry-picked down to 4.7. I can handle that if it's needed, but I'll wait until it's asked for.