Bug 620017
Summary: | GFS2 locks out entire cluster in case when mounted through fstab with ACL option | ||
---|---|---|---|
Product: | Red Hat Enterprise Linux 5 | Reporter: | Igor Smitran <viruslaki> |
Component: | gfs2-utils | Assignee: | Robert Peterson <rpeterso> |
Status: | CLOSED INSUFFICIENT_DATA | QA Contact: | Cluster QE <mspqa-list> |
Severity: | high | Docs Contact: | |
Priority: | low | ||
Version: | 5.7 | CC: | adas, bmarzins, edamato, swhiteho |
Target Milestone: | rc | ||
Target Release: | --- | ||
Hardware: | All | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2010-10-06 10:00:03 UTC | Type: | --- |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Igor Smitran
2010-07-31 12:16:48 UTC
In this call trace, the system is trying to unmount the gfs2 mount point, and that is waiting for dlm to send it a response to a lock request. Since it's hung, dlm is probably stuck for another reason, like a prior failure. Unfortunately, we don't have any information on any prior failure. I suspect that GFS2 is either hung due to another bug, or it encountered an error that caused it to panic due to errors=panic. In either case we may have already solved the problem but it hasn't made its way to your system yet. So here's what I recommend: 1. First, make sure you have a way to monitor the consoles of your nodes. 2. Adjust your "post_fail_delay" to a large value temporarily and reboot your cluster. 3. Recreate the hang. 4. Check to make sure both systems are still up and running after the hang. 5. Check dmesg to see if there are any indications of a failure on either node. 6. If there aren't any indications of failure on either node, use sysrq-t to collect complete call traces from both nodes. 7. Attach the call trace output or syslog from both nodes to the bugzilla. Igor, without further information we are unable to locate the source of this issue. If no further information is available we'll have to close this bug I'm afraid. |