Bug 243878
Summary: | kernel BUG when rebuilding raid5 array with hardware failures | ||||||
---|---|---|---|---|---|---|---|
Product: | Red Hat Enterprise Linux 4 | Reporter: | Orion Poplawski <orion> | ||||
Component: | kernel | Assignee: | Doug Ledford <dledford> | ||||
Status: | CLOSED WONTFIX | QA Contact: | Martin Jenner <mjenner> | ||||
Severity: | low | Docs Contact: | |||||
Priority: | low | ||||||
Version: | 4.5 | CC: | jbaron, mgahagan | ||||
Target Milestone: | --- | ||||||
Target Release: | --- | ||||||
Hardware: | All | ||||||
OS: | Linux | ||||||
Whiteboard: | |||||||
Fixed In Version: | Doc Type: | Bug Fix | |||||
Doc Text: | Story Points: | --- | |||||
Clone Of: | Environment: | ||||||
Last Closed: | 2009-04-07 13:31:02 UTC | Type: | --- | ||||
Regression: | --- | Mount Type: | --- | ||||
Documentation: | --- | CRM: | |||||
Verified Versions: | Category: | --- | |||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
Cloudforms Team: | --- | Target Upstream Version: | |||||
Embargoed: | |||||||
Attachments: |
|
Description
Orion Poplawski
2007-06-12 14:55:26 UTC
Created attachment 156799 [details]
Full logs
This request was evaluated by Red Hat Product Management for inclusion in a Red Hat Enterprise Linux maintenance release. Product Management has requested further review of this request by Red Hat Engineering, for potential inclusion in a Red Hat Enterprise Linux Update release for currently deployed products. This request is not yet committed for inclusion in an Update release. I know this was quite a while ago, but can you possibly expand on what you mean by "I was rebuilding a raid5 array that encountered hardware failures during the rebuild". This report indicates that the oops happened at the end of the rebuild when the resync thread was shutting down, but raid5 is a single failure only raid subsystem, so if it experienced hardware failure during the rebuild, how did it ever complete in the first place? What might be more relevant is the time starting at Jun 11 17:40:42, when a sync of md2 completes and then immediately there are several disk failures on that array. This seems to send the md driver into some kind of loop attempting to sync the array, even though it doesn't have enough devices. Otherwise, not sure I can be of more help, and I haven't reproduced it (thankfully). I eventually got things back up I believe by using badblocks -w to remap the sectors on the various drives before re-assembling the array. I don't have the ability to reproduce this, and as it involves multiple drive failures in a raid level that's only tolerant of a single drive failure, it doesn't rank very high on the "needs fixing" list. I'm closing this bug out as WONTFIX. Please reopen if you think this needs further attention. |