Bug 1414119
Summary: | running out of disk space in /etc/lvm/ on one cluster node may cause lvm ops running exclusively on another node to deadlock | ||
---|---|---|---|
Product: | Red Hat Enterprise Linux 6 | Reporter: | Corey Marthaler <cmarthal> |
Component: | lvm2 | Assignee: | LVM and device-mapper development team <lvm-team> |
lvm2 sub component: | Clustering / clvmd (RHEL6) | QA Contact: | cluster-qe <cluster-qe> |
Status: | CLOSED WONTFIX | Docs Contact: | |
Severity: | low | ||
Priority: | unspecified | CC: | agk, cmarthal, heinzm, jbrassow, msnitzer, prajnoha, prockai, zkabelac |
Version: | 6.9 | ||
Target Milestone: | rc | ||
Target Release: | --- | ||
Hardware: | x86_64 | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | If docs needed, set a value | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2017-12-06 10:27:52 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Corey Marthaler
2017-01-17 19:57:35 UTC
To make myself clear what is tested here - So to make it clear - thin-pool (and all its thins) have to be active only on 1 node in cluster - unfortunately this is not 'enforced' properly by lvm2 ATM and depends on a user to not make a mistake (there is not yet upstream fix for this). So is this bug about 'overfilling' thin-pool on 1 node. Then kill thin-pool on this particular node (lvchange -an for all LVs) And then trying thin-pool stack activation on different node ? No. This has nothing to do with full thin pool testing. All "cluster thin lvm" testing was done exclusively on just one node (in this scenario host-124). Activation and all other operations happened on only this *one* node. This bug is about how while this testing was happening, the root file systems on the other two nodes in the cluster (host-122 and host-123) were filled to 100%. So lvm no longer had any space in /etc/lvm/* to do any backup, cache, or archive writing on those two nodes, and as such, caused the testing that was happening exclusively on host-124 to start failing. Host-124 did still have free space in /. Red Hat Enterprise Linux 6 is in the Production 3 Phase. During the Production 3 Phase, Critical impact Security Advisories (RHSAs) and selected Urgent Priority Bug Fix Advisories (RHBAs) may be released as they become available. The official life cycle policy can be reviewed here: http://redhat.com/rhel/lifecycle This issue does not meet the inclusion criteria for the Production 3 Phase and will be marked as CLOSED/WONTFIX. If this remains a critical requirement, please contact Red Hat Customer Support to request a re-evaluation of the issue, citing a clear business justification. Note that a strong business justification will be required for re-evaluation. Red Hat Customer Support can be contacted via the Red Hat Customer Portal at the following URL: https://access.redhat.com/ |