+++ This bug was initially created as a clone of Bug #1852647 +++ Description of problem: Nodes are drained in parallel if 2 policies are applied at the same time Version-Release number of selected component (if applicable): 4.5 How reproducible: Steps to Reproduce: 1. Deploy sriov network operator on a cluster with at least 2 sriov capable worker nodes 2. Apply the following policies together ``` --- apiVersion: sriovnetwork.openshift.io/v1 kind: SriovNetworkNodePolicy metadata: name: policy-net-2 spec: resourceName: nic2 nodeSelector: kubernetes.io/hostname: worker-0 feature.node.kubernetes.io/network-sriov.capable: "true" priority: 99 mtu: 9000 numVfs: 4 nicSelector: pfNames: ['ens803f0#0-0'] isRdma: false --- apiVersion: sriovnetwork.openshift.io/v1 kind: SriovNetworkNodePolicy metadata: name: policy-net-2-vfio spec: resourceName: nic2vfio nodeSelector: feature.node.kubernetes.io/network-sriov.capable: "true" priority: 99 mtu: 9000 numVfs: 4 nicSelector: vendor: "8086" pfNames: ['ens803f0#0-0'] deviceType: vfio-pci isRdma: false ``` 3. Actual results: More than 1 worker node was drained and set to 'unschedulable' in parallel. Expected results: All worker nodes shall be drained one by one in sequence. Additional info: The workaround is to apply the second policy after the first one was fully synced on all the nodes. --- Additional comment from Peng Liu on 2020-07-01 02:58:48 UTC --- Fixed in PR https://github.com/openshift/sriov-network-operator/pull/249 and https://github.com/openshift/sriov-network-operator/pull/260
Verified this bug on 4.5.0-202007172106.p0 when creating policy with 2 nodes matched. the node will SchedulingDisabled one by one.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2020:2956