Bug 1444861
Summary: | Brick Multiplexing: bricks of volume going offline possibly because the brick PID is associated with another volume which was brought down | ||
---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Nag Pavan Chilakam <nchilaka> |
Component: | core | Assignee: | Mohit Agrawal <moagrawa> |
Status: | CLOSED ERRATA | QA Contact: | Nag Pavan Chilakam <nchilaka> |
Severity: | urgent | Docs Contact: | |
Priority: | unspecified | ||
Version: | rhgs-3.3 | CC: | amukherj, nchilaka, rhs-bugs, storage-qa-internal |
Target Milestone: | --- | ||
Target Release: | RHGS 3.3.0 | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | brick-multiplexing | ||
Fixed In Version: | glusterfs-3.8.4-25 | Doc Type: | If docs needed, set a value |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2017-09-21 04:39:40 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | 1450630 | ||
Bug Blocks: | 1417151 |
Description
Nag Pavan Chilakam
2017-04-24 12:47:01 UTC
upstream patch : https://review.gluster.org/#/c/17101/ Upstream patches : https://review.gluster.org/#/q/topic:bug-1444596 Downstream patches: https://code.engineering.redhat.com/gerrit/#/c/105595/ https://code.engineering.redhat.com/gerrit/#/c/105596/ QA validation: Moving to failed_qa If i bring down a brick of one volume , it still is disconnecting all the bricks from the glusterfsd did below steps 1)have a cluster with brick mux enabled 2) created 10 vols of 1x3 type 3) brought down b1 of vol7 (using umount of lv) 4) now mount vol7 and vol1(base vol) and vol3(any other vol) 5)do IOs to all the above vols ==>you will see that all the bricks associated with the same glusterfsd as b1 of vol7 would not be receiving any IO, effectively losing the brick availability you can check even the heal info for the volume, it will show files as heal pending and checked the backend brick test version ==== 3.8.4-25 Nag, This is a known issue and currently it(scenario) is not handled completely.The issue will come only when brick has down in some ungraceful manner and as per bugzilla earlier brick was down in some graceful way(through the cli). So please verified to this bugzilla followed same procedure as you mentioned in comment 1 For specific to handle this kind of scenario fix in under progress from below patch https://review.gluster.org/17287 Regards Mohit Agrawal I agree with Mohit. The steps which were followed to file this bug and the steps which were followed to verify this bug are different. please follow the same steps and reconfirm. I cannot verify this until BZ#1450630 is fixed Patch for (BZ#1450630) is already merged in downstream from this bugzilla https://bugzilla.redhat.com/show_bug.cgi?id=1450806. Regards Mohit Agrawal Mohit - the current build doesn't have the fix, so Nag's comment is valid. Nag - as this bug has been moved to MODIFIED state, expect this fix to land in the next build. On_qa validation: 3.8.4-33 is the test build ran both the cases mentioned in 1)description 2)comment#8 not seeing the issue anymore hence moving to verified Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2017:2774 |