Bug 1120198
Summary: | Peers moved to rejected state because of mismatch in volume checksums after creating the volume. | ||||||
---|---|---|---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | spandura | ||||
Component: | glusterd | Assignee: | Kaushal <kaushal> | ||||
Status: | CLOSED ERRATA | QA Contact: | SATHEESARAN <sasundar> | ||||
Severity: | high | Docs Contact: | |||||
Priority: | high | ||||||
Version: | rhgs-3.0 | CC: | kaushal, nlevinki, nsathyan, sasundar, spandura, ssamanta, vagarwal, vbellur | ||||
Target Milestone: | --- | ||||||
Target Release: | RHGS 3.0.0 | ||||||
Hardware: | Unspecified | ||||||
OS: | Unspecified | ||||||
Whiteboard: | |||||||
Fixed In Version: | glusterfs-3.6.0.25-1 | Doc Type: | Bug Fix | ||||
Doc Text: | Story Points: | --- | |||||
Clone Of: | |||||||
: | 1120589 (view as bug list) | Environment: | |||||
Last Closed: | 2014-09-22 19:44:25 UTC | Type: | Bug | ||||
Regression: | --- | Mount Type: | --- | ||||
Documentation: | --- | CRM: | |||||
Verified Versions: | Category: | --- | |||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
Cloudforms Team: | --- | Target Upstream Version: | |||||
Embargoed: | |||||||
Bug Depends On: | 1120589 | ||||||
Bug Blocks: | |||||||
Attachments: |
|
Description
spandura
2014-07-16 11:44:21 UTC
Shwetha, Could you describe how you had the bricks setup (LVM/ThinP etc.)? I've not verified it yet, but I feel this is a part of the cause. ~kaushal Created attachment 918626 [details]
Script to create bricks
running the script : ./mkfs_snapshot1.sh "create"
Kaushal, I feel this bug is dependent on, https://bugzilla.redhat.com/show_bug.cgi?id=1116264 'gluster volume info' has got some extra capabilities listed and does that change the volume checksums ? Thanks Shwetha. Turns out the problem wasn't with the bricks. It was with the way BD xlator capabilities was being set, or rather being erased during volume create. During volume create, the caps are initially set to all enabled. This happens on all volumes irrespective of whether its a BD volume or not. And then based on the bricks capabilities, unsupported capabilities were removed or erased. But this removal was being done only on the peers which contained bricks for the volume. On the other peers, the capabilities were not erased, which was leading to checksums differring. This lead to peers being rejected when they were restarted. (In reply to SATHEESARAN from comment #4) > Kaushal, > > I feel this bug is dependent on, > https://bugzilla.redhat.com/show_bug.cgi?id=1116264 > > 'gluster volume info' has got some extra capabilities listed and does that > change the volume checksums ? Yes. Both of these are caused by the same problem. I would close this bug as a duplicate of the above, but that would probably cause some other procedural problems. Can you check and let me know if it is okay? *** Bug 1116264 has been marked as a duplicate of this bug. *** Verified with glusterfs-3.6.0.25-1.el6rhs Performed the following steps : 1. Created a 4 node cluster ( Trusted Storage Pool ) 2. Created thinp bricks as recommended for gluster volume snapshot feature 3. Created distributed volume, distributed-replicate volume 4. Checked the 'gluster volume status' on all the nodes Observation: No BD Xlator related capabilities are shown in the 'gluster volume info' output 5. Restarted glusterd on all nodes Observation: 'gluster peer status' shows all the nodes with "Peer in Cluster" 6. Started the volume 7. Checked 'gluster peer status' and 'gluster volume info' Observations remain the same Marking this bug as verified Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHEA-2014-1278.html |