Bug 199498
Summary: | mirror leg failure during I/O causes I/O hang and apparent volume corruption | ||
---|---|---|---|
Product: | [Retired] Red Hat Cluster Suite | Reporter: | Corey Marthaler <cmarthal> |
Component: | cmirror | Assignee: | Jonathan Earl Brassow <jbrassow> |
Status: | CLOSED CURRENTRELEASE | QA Contact: | Cluster QE <mspqa-list> |
Severity: | high | Docs Contact: | |
Priority: | high | ||
Version: | 4 | CC: | agk, cfeist, dwysocha, mbroz |
Target Milestone: | --- | ||
Target Release: | --- | ||
Hardware: | All | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2008-08-05 21:32:19 UTC | Type: | --- |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Corey Marthaler
2006-07-19 21:20:31 UTC
Reproduced this issue with I/O on a non cmirror server. Trying again with I/O on the cmirror server... With I/O on the mirror server the device failure case worked after about 5 minutes of scsi and write errors before the device was finally converted to a linear. Re-attempting the case in comment #1 and will let is hang over night to see if we are just not waiting long enough. Attempted the case in comment #1 again (I/O on a non cmirror server and then fail one of the legs) and this time the volumes didn't get corrupted, however it didn't appear that it got properely converted to a linear either. After the failed leg: [root@taft-02 ~]# lvscan /dev/sdd1: read failed after 0 of 2048 at 0: Input/output error ACTIVE '/dev/vg/mirror' [1.00 GB] inherit [root@taft-02 ~]# lvs -a -o +devices /dev/sdd1: read failed after 0 of 2048 at 0: Input/output error LV VG Attr LSize Origin Snap% Move Log Copy% Devices mirror vg mwi-s- 1.00G mirror_mlog 100.00 mirror_mimage_0(0),mirror_mimage_1(0) [mirror_mimage_0] vg iwi-so 1.00G [mirror_mimage_1] vg iwi-so 1.00G /dev/sdb1(0) [mirror_mlog] vg lwi-so 4.00M /dev/sdc1(0) After a couple of minutes had passed since the failed leg, the I/O ended up hanging (as well as clvmd) and I let that hang overnight, it never came back. Hit this issue while attempting the same leg failure but with I/O from all nodes in the cluster to the mirror. CLVMD ended up hung as well as the I/O. clvmd should no longer hang, and no volume corruption should occur. as far as I/O hanging... that could be bug #199724, and could be a result of mirror reconfiguration taking too long. marking this verified, there are other leg failure bugs open for more specific cases. Closing as this has been fixed in the current (4.7) release. |