+++ This bug was initially created as a clone of Bug #1781283 +++
Noticed this in skew tests, but it can also happen when the CVO gets rescheduled onto a different machine during an upgrade.
We should be setting resource requests, looking at e2e runs it's ~ 10m CPU and 50Mi mem, roughly 1Mi disk.
Run upgrade from 4.3.0-0.nightly-2019-12-19-185024 to 4.3.0-0.nightly-2019-12-19-204222, checked extract job runs in burstable qos:
# oc get pod/version--np5tt-wrpmp -o json|jq .spec.containers.resources
# oc get pod/version--np5tt-wrpmp -o json|jq .status.qosClass
But checked the 4.2-4.3 upgrade ci test result, still failed with [Feature:Platform][Smoke] Managed cluster should ensure control plane pods do not run in best-effort QoS.
So i think we need backport the fix to 4.2 to let the ci test pass, so clone a bug to 4.2 and verify this bug.
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.
For information on the advisory, and where to find the updated
files, follow the link below.
If the solution does not work for you, open a new bug report.