Bug 2180329
Summary: | [RDR][tracker for BZ 2215392] RBD images left behind in managed cluster after deleting the application | |||
---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat OpenShift Data Foundation | Reporter: | Sidhant Agrawal <sagrawal> | |
Component: | ceph | Assignee: | N Balachandran <nibalach> | |
ceph sub component: | RBD-Mirror | QA Contact: | kmanohar | |
Status: | CLOSED ERRATA | Docs Contact: | ||
Severity: | urgent | |||
Priority: | unspecified | CC: | amagrawa, bniver, ebenahar, idryomov, kseeger, muagarwa, nibalach, odf-bz-bot, sostapov | |
Version: | 4.13 | Keywords: | Automation, Tracking | |
Target Milestone: | --- | |||
Target Release: | ODF 4.14.0 | |||
Hardware: | Unspecified | |||
OS: | Unspecified | |||
Whiteboard: | ||||
Fixed In Version: | 4.14.0-110 | Doc Type: | No Doc Update | |
Doc Text: | Story Points: | --- | ||
Clone Of: | ||||
: | 2215392 (view as bug list) | Environment: | ||
Last Closed: | 2023-11-08 18:50:04 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | 2215392 | |||
Bug Blocks: |
Description
Sidhant Agrawal
2023-03-21 08:07:06 UTC
Moving this one to 4.14 as it's not reproducible in recent builds and RDR is in TP state. Fix is merged upstream, waiting for ceph builds with the fix. VERIFICATION COMMENTS :- ===================== Steps to Reproduce: ------------------- 1. Configure RDR setup 2. Deploy an application containing 20 PVCs/Pods on C1 3. Wait for 10 minutes to run IOs 4. Scale down RBD mirror daemon deployment to 0 5. Initiate failover to C2 6. Check PVC and pod resources are created on C2 successfully. 7. Scale up RBD mirror daemon deployment back to 1 8. Check application and replication resources deleted from C1 9. Check mirroring status cluster: sagrawal-c1 {'daemon_health': 'OK', 'health': 'OK', 'image_health': 'OK', 'states': {'replaying': 20}} cluster: sagrawal-c2 {'daemon_health': 'OK', 'health': 'OK', 'image_health': 'OK', 'states': {'replaying': 20}} 10. Wait for 10 minutes to run IOs 11. Initiate Relocate to C1 12. Check mirroring status after relocate operation cluster: sagrawal-c1 {'daemon_health': 'OK', 'health': 'OK', 'image_health': 'OK', 'states': {'replaying': 20}} cluster: sagrawal-c2 {'daemon_health': 'OK', 'health': 'OK', 'image_health': 'OK', 'states': {'replaying': 20}} 13. Delete the application 14. Observe the mirroring status cluster: sagrawal-c1 {"daemon_health":"OK","health":"OK","image_health":"OK","states":{}} cluster: sagrawal-c2 {"daemon_health":"OK","health":"WARNING","image_health":"WARNING","states":{"unknown":15}} Automated test: tests/disaster-recovery/regional-dr/test_failover_and_relocate.py Actual results: After deleting the application workload, mirroring status in WARNING and RBD images left behind in the managed cluster Expected results: Mirroring status should be OK and all RBD images should be deleted after deleting the application workload. _____________________________________________________________________________________________________________ o/p after deleting the application ----------------------------------------------- On C1 sh-5.1$ rbd mirror pool status ocs-storagecluster-cephblockpool health: OK daemon health: OK image health: OK images: 0 total sh-5.1$ for i in $(rbd ls -p ocs-storagecluster-cephblockpool); do echo $i; rbd snap ls ocs-storagecluster-cephblockpool/$i --all 2>/dev/null; echo "##########################################";done csi-vol-e9ffb004-2730-4369-a916-dd45e29f2a41 ########################################## $ oc get pods No resources found in busybox-workloads-1 namespace. $ oc get pvc No resources found in busybox-workloads-1 namespace. On C2 sh-5.1$ rbd mirror pool status ocs-storagecluster-cephblockpool health: OK daemon health: OK image health: OK images: 0 total sh-5.1$ for i in $(rbd ls -p ocs-storagecluster-cephblockpool); do echo $i; rbd snap ls ocs-storagecluster-cephblockpool/$i --all 2>/dev/null; echo "##########################################";done csi-vol-7e98268e-2b2a-40c7-86cd-deeab148d0c9 ########################################## Verified on ------------ OCP - 4.14.0-0.nightly-2023-10-13-032002 ODF - 4.14.0-150 Ceph version - ceph version 17.2.6-146.el9cp (1d01c2b30b5fd39787bb8804707c4b2e52e30137) quincy (stable) Submariner - 0.16.0 ACM - 2.9.0 (Image - 2.9.0-DOWNSTREAM-2023-10-03-20-08-35) Must gather - http://rhsqe-repo.lab.eng.blr.redhat.com/OCS/ocs-qe-bugs/keerthana/bz-v/bz-2180329/ Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Important: Red Hat OpenShift Data Foundation 4.14.0 security, enhancement & bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2023:6832 |