Bug 2250883

Summary: [Backport to 4.13] CSI pods and customer workloads both have 'priority=0' and race for resources
Product: [Red Hat Storage] Red Hat OpenShift Data Foundation Reporter: Eran Tamir <etamir>
Component: ocs-operatorAssignee: Nitin Goyal <nigoyal>
Status: CLOSED ERRATA QA Contact: Aviad Polak <apolak>
Severity: high Docs Contact:
Priority: unspecified    
Version: 4.13CC: apolak, bniver, kramdoss, mparida, mrajanna, muagarwa, nigoyal, odf-bz-bot, rdomingo, sostapov, tdesala
Target Milestone: ---   
Target Release: ODF 4.13.7   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: 4.13.7-1 Doc Type: Bug Fix
Doc Text:
Story Points: ---
Clone Of: 2232464 Environment:
Last Closed: 2024-01-29 08:22:12 UTC Type: ---
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 2232464, 2250884    
Bug Blocks: 2244409    

Description Eran Tamir 2023-11-21 16:44:13 UTC
+++ This bug was initially created as a clone of Bug #2232464 +++

Description of problem (please be detailed as possible and provide log
snippests):
- CSI pods have priority 0 instead of using openshift priorityclasses (https://docs.openshift.com/container-platform/4.13/nodes/pods/nodes-pods-priority.html).  Customer workloads also have priority 0 by default, which leads to a race for scheduling and resources.  

Does this issue impact your ability to continue to work with the product
(please explain in detail what is the user impact)?
- Customer have upcoming activities which this issue may occur again 


Is there any workaround available to the best of your knowledge?
1. The most practical way is by moving a pod that has at least 350 of MB of requests to a different node.

2. If the 1st does not help then contact us (Red Hat) with the node name which we can recycle the whole node with help from SRE.


Rate from 1 - 5 the complexity of the scenario you performed that caused this
bug (1 - very simple, 5 - very complex)? 3


Can this issue reproducible? Yes
This was triggered via a DR testing scenario.  So I think trying to quickly spin up full workloads on an empty cluster

Can this issue reproduce from the UI?
- Not sure


If this is a regression, please provide more details to justify this:


Steps to Reproduce:
1. Apply CSI pods and customer workloads at the same time
2. Repeat until race is hit where CSI pods cannot be scheduled.


Actual results:
CSI pods can't be scheduled when customer workloads get there first

Expected results:
CSI pods are scheduled before customer workloads

Additional info:
Per SRE - From OCP point of view, AWS EBS CSI driver on ROSA should have enough priority - linking 4.11 yamls:

https://github.com/openshift/aws-ebs-csi-driver-operator/blob/release-4.11/assets/node.yaml#L24

https://github.com/openshift/aws-ebs-csi-driver-operator/blob/release-4.11/assets/controller.yaml#L25

--- Additional comment from RHEL Program Management on 2023-08-17 03:14:10 UTC ---

This bug having no release flag set previously, is now set with release flag 'odf‑4.14.0' to '?', and so is being proposed to be fixed at the ODF 4.14.0 release. Note that the 3 Acks (pm_ack, devel_ack, qa_ack), if any previously set while release flag was missing, have now been reset since the Acks are to be set against a release flag.

--- Additional comment from Venky Shankar on 2023-08-17 04:24:50 UTC ---

This seems CSI related - moving component.

--- Additional comment from Madhu Rajanna on 2023-08-17 04:37:37 UTC ---

CSI Pods doesn't have the priority 0 rather it also has the priority class https://github.com/rook/rook/blob/master/deploy/examples/operator.yaml#L120-L124 in upstream this can also be set by ocs-operator
Moving it to ocs-operator as it creates the Rook configmap.

--- Additional comment from RHEL Program Management on 2023-08-17 08:17:41 UTC ---

This BZ is being approved for ODF 4.14.0 release, upon receipt of the 3 ACKs (PM,Devel,QA) for the release flag 'odf‑4.14.0

--- Additional comment from RHEL Program Management on 2023-08-17 08:17:41 UTC ---

Since this bug has been approved for ODF 4.14.0 release, through release flag 'odf-4.14.0+', the Target Release is being set to 'ODF 4.14.0

--- Additional comment from errata-xmlrpc on 2023-08-18 06:28:37 UTC ---

This bug has been added to advisory RHBA-2023:115514 by ceph-build service account (ceph-build.COM)

--- Additional comment from Aviad Polak on 2023-08-24 09:37:32 UTC ---

LGTM in build       full_version: 4.14.0-112

priority: 2000001000
priorityClassName: system-node-critical

--- Additional comment from errata-xmlrpc on 2023-11-08 17:54:16 UTC ---

Bug report changed to RELEASE_PENDING status by Errata System.
Advisory RHSA-2023:115514-11 has been changed to PUSH_READY status.
https://errata.devel.redhat.com/advisory/115514

--- Additional comment from errata-xmlrpc on 2023-11-08 18:54:15 UTC ---

Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Important: Red Hat OpenShift Data Foundation 4.14.0 security, enhancement & bug fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHSA-2023:6832

Comment 10 errata-xmlrpc 2024-01-29 08:22:12 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Red Hat OpenShift Data Foundation 4.13.7 Bug Fix Update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2024:0540