Bug 1333406
Summary: | [HC]: After bringing down and up of the bricks VM's are getting paused | |||
---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | RajeshReddy <rmekala> | |
Component: | replicate | Assignee: | Krutika Dhananjay <kdhananj> | |
Status: | CLOSED ERRATA | QA Contact: | SATHEESARAN <sasundar> | |
Severity: | high | Docs Contact: | ||
Priority: | high | |||
Version: | rhgs-3.1 | CC: | amukherj, mzywusko, pkarampu, rcyriac, rhinduja, rhs-bugs, sabose, sasundar | |
Target Milestone: | --- | |||
Target Release: | RHGS 3.2.0 | |||
Hardware: | Unspecified | |||
OS: | Unspecified | |||
Whiteboard: | ||||
Fixed In Version: | glusterfs-3.8.4-1 | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | ||
Clone Of: | ||||
: | 1363721 (view as bug list) | Environment: | ||
Last Closed: | 2017-03-23 05:29:33 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | ||||
Bug Blocks: | 1277939, 1351522, 1363721, 1367270, 1367272 |
Description
RajeshReddy
2016-05-05 12:32:06 UTC
sosreports are avilable @rhsqe-repo.lab.eng.blr.redhat.com:/home/repo/sosreports/bug.1333406 This bug is related to cyclic network outage test causing file to be in split brain. As this is not a likely scenario, removing from 3.1.3 target You are correct, we can't prevent VMs getting paused. We only need to make sure that split-brains won't happen. Please note that this case may lead to the VM image going extremely bad, but all we can guarantee is the file not going into split-brain. Upstream mainline patch http://review.gluster.org/15080 posted for review. Upstream mainline : http://review.gluster.org/15080 http://review.gluster.org/15145 Upstream 3.8 : http://review.gluster.org/15221 http://review.gluster.org/15164 And the fix is available in rhgs-3.2.0 as part of rebase to GlusterFS 3.8.4. Tested with RHGS 3.2.0 interim build ( glusterfs-3.8.4-12.el7rhgs ) with the following steps: 1. Created replica 3 volume and used it as data domain in RHV 2. When there are continuous I/O happening on the VMs, killed first brick 3. After some time brought up the down brick, and in few mins killed second brick 4. After some time brought up the down brick, and in another few mins killed third brick. 5. After some time brought up the down brick, and in another few mins killed first brick. After all this steps, I haven't seen any hiccups with VMs, VMs healthy post reboot, and there are no problems Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHSA-2017-0486.html |