Description of problem: Live image migration command "Abort" is not working as expected. when i use abort option i am getting the below message. [root@magna027 ~]# rbd migration abort libvirt-pool/librbd --cluster site-a 2019-11-18 10:00:37.966 7f8f21ffb700 -1 librbd::image::PreRemoveRequest: 0x55935d187080 handle_exclusive_lock: cannot obtain exclusive lock - not removing 2019-11-18 10:00:37.999 7f8f3d4b0040 -1 librbd::Migration: abort: failed removing destination image 'libvirt-pool/librbd (af59a0e341ed)': (16) Device or resource busy Abort image migration: 100% complete...done. [root@magna027 ~]# Version-Release number of selected component (if applicable): ceph-mon-14.2.4-59.el8cp.x86_64 ceph-common-14.2.4-59.el8cp.x86_64 ceph-ansible-4.0.4-1.el8cp.noarch How reproducible: always. Steps to Reproduce: 1. Deploy a cluster with 4.0 on rhel8.1 2. Create two rbd pools and one image in first pool. 3. Prepare rbd migration from source pool to destination pool 4. Start IO on image which was moved from source to destination. 5. start migration using "rbd migration execute migration_target" command while IO is in progress. 6. Abort migration using "rbd migration abort migration_target" command while IO is in progress. Actual results: Image is not removed from target and moved to source as well. [root@magna027 ~]# rbd migration abort libvirt-pool/librbd --cluster site-a 2019-11-18 10:00:37.966 7f8f21ffb700 -1 librbd::image::PreRemoveRequest: 0x55935d187080 handle_exclusive_lock: cannot obtain exclusive lock - not removing 2019-11-18 10:00:37.999 7f8f3d4b0040 -1 librbd::Migration: abort: failed removing destination image 'libvirt-pool/librbd (af59a0e341ed)': (16) Device or resource busy Abort image migration: 100% complete...done. [root@magna027 ~]# Expected results: Aborting the migration should remove the cross-links between the source and target images, and should remove the target image. Additional info: Stopped running IO and tried abort the image which is worked as expected.
Can we assign an engineer to triage this? Moving along 4.x is not the right way to get this into a release - it should be targeted to 5.0 first.
Tested the bug on latest build and it's working as expected. Hence moving the bug into verified state.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Important: Red Hat Ceph Storage 4.2 Security and Bug Fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:0081