Description of problem (please be detailed as possible and provide log snippests): In Multus enabled cluster, deletion of CephBlockPool is getting stuck and pool doesn't get deleted from Ceph side. This also blocks creation of new CephBlockPools as they never enter Ready Phase and new pool is not created in Ceph side. Version of all relevant components (if applicable): OCS: ocs-operator.v4.8.0-455.ci Does this issue impact your ability to continue to work with the product (please explain in detail what is the user impact)? Yes, I'm unable to create delete old CephBlockPool and create new one Is there any workaround available to the best of your knowledge? Delete pool manually from Ceph side using toolbox Rate from 1 - 5 the complexity of the scenario you performed that caused this bug (1 - very simple, 5 - very complex)? 2 Can this issue reproducible? Yes Can this issue reproduce from the UI? Yes If this is a regression, please provide more details to justify this: Steps to Reproduce: Manual steps: 1. Install OCS with Multus enabled 2. Create a new blockpool 3. Delete the blockpool created in step 2 (will not succeed) 4. Try to create another blockpool (will not succeed) Or Run tests from tests/manage/storageclass [1] Let say, first run tests/manage/storageclass/test_create_2_sc_with_1_pool_comp_rep2.py::TestMultipleScOnePoolRep2Comp::test_multiple_sc_one_pool_rep2_comp > test will pass but there will be error in teardown due to failure in cephblockpool deletion Now run tests/manage/storageclass/test_create_2sc_at_once_with_io.py::TestCreate2ScAtOnceWithIo::test_new_sc_rep2_rep3_at_once >This test will fail during creation of new cephblockpool [1] https://github.com/red-hat-storage/ocs-ci/tree/master/tests/manage/storageclass Actual results: Existing pool deletion get stuck and unable to create new one Expected results: Should be able to delete an existing pool and create new pools Additional info:
Downstream PR is ready but on-hold, waiting for various acks and a blocker flag.
Moving it out of 4.8 as discussed in the thread mentioned in the above comment. Sebastien, please fill the doc text.
.Deletion of `CephBlockPool` gets stuck and blocks the creation of new pools When Rook is deployed along with Multus, the Rook Operator does not have the network annotations and thus does not have access to the OSD network. This means that when running "rbd" type commands during pool cleanup, the command hangs since it cannot contact the OSDs. The workaround is to delete the CephBlockPool manually using toolbox.
*** Bug 1982672 has been marked as a duplicate of this bug. ***
Sidhant, please provide debug logs or access to the env, thanks.
Shared cluster details via gchat. Clearing NI.
Resync PR https://github.com/red-hat-storage/rook/pull/19
Fix should be available in the latest ODF builds
I don't need the env anymore, thanks for providing it.
https://github.com/red-hat-storage/rook/pull/305
Doc text needs to be changed, earlier it was a known issue and now a bug fix.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat OpenShift Data Foundation 4.9.0 enhancement, security, and bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:5086