Platform: ---------- Vmware 3M 3W RHCOS cluster Versions: ---------- OCP - 4.7.12 OCS - ocs-operator.v4.7.1-403.ci Testcases Executed: ---------------------- 1. Perform single node drain and Restart rook-ceph operator during mon failover to happen a. Wait for >= 20 mins and restart the rook-ceph operator b. Uncordon drained node after 20 mins c. Check for mons running in healthy state and no mons should be observed in pending state post recovery of node 2. Delete the mon deployment a. Create the cluster, wait for it to be initially deployed b. Scale down a mon (e.g. mon-a) so it falls out of quorum c. Wait for the mon failover to be initiated (10 min) d. As soon as the new mon is created and before the bad mon deployment (rook-ceph-mon-a) is deleted, restart the operator e. All 3 mons will be running Restarted rook-ceph operator just during the failover I did not see 4 mons, I always saw 3 mons, Its working as expected. we dont see 2 pending Mons @santosh let me know if this BZ needs any more verifications or if it can be moved to verified state
(In reply to Shrivaibavi Raghaventhiran from comment #8) > @santosh let me know if this BZ needs any more verifications or if it can be > moved to verified state Looks good to me. Can be moved to verified state.
Based on comment 8 and comment 9 moving the BZ to verified state
needs info separately answered between comments 7 and 8
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Red Hat OpenShift Container Storage 4.7.1 bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2021:2449