Bug 2203265
| Summary: | protection against splitting off mirror legs during up conversion appears to be missing now when sync is close to completion with fast nvme devices | ||
|---|---|---|---|
| Product: | Red Hat Enterprise Linux 9 | Reporter: | Corey Marthaler <cmarthal> |
| Component: | lvm2 | Assignee: | Heinz Mauelshagen <heinzm> |
| lvm2 sub component: | Mirroring and RAID | QA Contact: | cluster-qe <cluster-qe> |
| Status: | NEW --- | Docs Contact: | |
| Severity: | medium | ||
| Priority: | high | CC: | agk, heinzm, jbrassow, msnitzer, prajnoha, zkabelac |
| Version: | 9.3 | Keywords: | Triaged |
| Target Milestone: | rc | ||
| Target Release: | --- | ||
| Hardware: | x86_64 | ||
| OS: | Linux | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | Type: | Bug | |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Corey Marthaler
2023-05-11 16:49:40 UTC
This appears to only happen with fast nvme devices. I haven't been able to reproduce this with SSDs. Also, if I bump the size of the mirror way up with the nvme devices, then I will see the "Unable to split mirror mirror_sanity/fs_to_mirror that is not in-sync." messages again. So this appears to be a timing issue where LVM thinks the sync is complete, or close to complete, but it's not really. (In reply to Corey Marthaler from comment #1) > This appears to only happen with fast nvme devices. I haven't been able to > reproduce this with SSDs. Also, if I bump the size of the mirror way up with > the nvme devices, then I will see the "Unable to split mirror > mirror_sanity/fs_to_mirror that is not in-sync." messages again. > > So this appears to be a timing issue where LVM thinks the sync is complete, > or close to complete, but it's not really. So it seems. FYI: we've seen similar issues with dm-raid on (small) LVs backed by fast PVs whch have been fixed a few years ago. No reproducer in VM here (virtio-scsi -> host thin -> host PCIe4 NVMe): # lvconvert -y --ty mirror -m1 t/t&lvconvert -y --splitmirrors 1 --name n t/t [1] 5089 Logical volume t/t being converted. Unable to split mirror t/t that is not in-sync. t/t: Converted: 75.00% # lvs LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert root rhel_vm45 -wi-ao---- <17.00g swap rhel_vm45 -wi-ao---- 2.00g t t mwi-a-m--- 128.00m [t_mlog] 100.00 # lvconvert -m0 t t/t: Converted: 100.00% /t Logical volume t/t converted. [root@vm45 ~]# lvs LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert root rhel_vm45 -wi-ao---- <17.00g swap rhel_vm45 -wi-ao---- 2.00g t t -wi-a----- 128.00m # lvconvert -y --ty mirror -m1 t/t&sleep 0.1;lvconvert -y --splitmirrors 1 --name n t/t [2] 5404 Logical volume t/t being converted. [1] Done lvconvert -y --ty mirror -m1 t/t Unable to split mirror t/t that is not in-sync. t/t: Converted: 71.88% |