Description of problem (please be detailed as possible and provide log snippests): [RDR] When cluster was upgraded from 4.16 to 4.17, osd migrated to new ceph osd id's Version of all relevant components (if applicable): OCP version:- 4.17.0-0.nightly-2024-09-02-044025 ODF version:- 4.17.0-90 CEPH version:- ceph version 19.1.0-42.el9cp (03ae7f7ffec5e7796d2808064c4766b35c4b5ffb) squid (rc) ACM version:- 2.11.2 SUBMARINER version:- v0.18.0 VOLSYNC version:- volsync-product.v0.10.0 OADP version:- 1.4.0 VOLSYNC method:- destinationCopyMethod: Direct Does this issue impact your ability to continue to work with the product (please explain in detail what is the user impact)? Is there any workaround available to the best of your knowledge? Rate from 1 - 5 the complexity of the scenario you performed that caused this bug (1 - very simple, 5 - very complex)? Can this issue be reproducible? yes Can this issue reproduce from the UI? If this is a regression, please provide more details to justify this: Steps to Reproduce: 1.Deploy RDR 4.16 2.Upgrade RDR cluster to 4.17 3. check ceph status 4. Migration for osd will be started automatically Actual results: rook-ceph-osd-3-ff888c5f5-wqjfx 2/2 Running 0 89m rook-ceph-osd-4-8c7886d68-n9mbj 2/2 Running 0 64m rook-ceph-osd-5-696975c56-sl4mj 2/2 Running 0 35m Expected results: We should not change osd id's openshift-storage rook-ceph-osd-1-7f59c986c7-hv55p 2/2 Running 0 6h42m openshift-storage rook-ceph-osd-2-866d4787bf-trnqp 2/2 Running 0 6h48m openshift-storage rook-ceph-osd-3-ff888c5f5-wqjfx 2/2 Running 0 11m Additional info: cephstatus cluster: id: d378191d-7fe9-469c-9385-7cb128679367 health: HEALTH_WARN 1 osds down Degraded data redundancy: 79462/564801 objects degraded (14.069%), 73 pgs degraded, 80 pgs undersized services: mon: 3 daemons, quorum d,e,f (age 92m) mgr: a(active, since 9m), standbys: b mds: 1/1 daemons up, 1 hot standby osd: 6 osds: 3 up (since 36m), 4 in (since 37m) rbd-mirror: 1 daemon active (1 hosts) rgw: 1 daemon active (1 hosts, 1 zones) data: volumes: 1/1 healthy pools: 12 pools, 169 pgs objects: 188.27k objects, 67 GiB usage: 163 GiB used, 5.8 TiB / 6 TiB avail pgs: 79462/564801 objects degraded (14.069%) 89 active+clean 73 active+undersized+degraded 7 active+undersized io: client: 20 MiB/s rd, 6.1 MiB/s wr, 1.73k op/s rd, 28 op/s wr