Bug 865693
Summary: | volume information is out of sync in the cluster | ||||||
---|---|---|---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Rahul Hinduja <rhinduja> | ||||
Component: | glusterd | Assignee: | krishnan parthasarathi <kparthas> | ||||
Status: | CLOSED ERRATA | QA Contact: | spandura | ||||
Severity: | urgent | Docs Contact: | |||||
Priority: | medium | ||||||
Version: | 2.0 | CC: | amarts, grajaiya, nsathyan, rhs-bugs, shaines, spandura, vbellur | ||||
Target Milestone: | --- | ||||||
Target Release: | --- | ||||||
Hardware: | x86_64 | ||||||
OS: | Linux | ||||||
Whiteboard: | |||||||
Fixed In Version: | glusterfs-3.4.0qa6 | Doc Type: | Bug Fix | ||||
Doc Text: | Story Points: | --- | |||||
Clone Of: | Environment: | ||||||
Last Closed: | 2013-09-23 22:39:16 UTC | Type: | Bug | ||||
Regression: | --- | Mount Type: | --- | ||||
Documentation: | --- | CRM: | |||||
Verified Versions: | Category: | --- | |||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
Cloudforms Team: | --- | Target Upstream Version: | |||||
Embargoed: | |||||||
Attachments: |
|
Description
Rahul Hinduja
2012-10-12 07:21:13 UTC
Created attachment 625839 [details]
sosreports and /var/lib/glusterd directory files
general behavior/bug in glusterd. Not very specific to 2.0+ *** Bug 865406 has been marked as a duplicate of this bug. *** marking ON_QA for rhs-2.1.0 flag. Let us know which update of RHS.2.0.z do we need this fix. Updating summary since this is a general bug. Verified the fix on the build: ============================== glusterfs 3.4.0.22rhs built on Aug 23 2013 01:58:42 "volume sync command scenarios" ======================================== a. gluster volume sync <hostname> b. gluster volume sync <hostname> all c. gluster volume sync <hostname> <volume_name> d. gluster volume sync <localhost> Following cases tests the above 4 scenarios. ============================================================================== Case1 : ( 1 x 2 replicate volumes, 2 Storage nodes ) ============================================================================== 1. start glusterd on both the storage nodes. 2. peer probe storage_node2 ( from storage_node1 ) 3. On storage_node1 execute: for i in `seq 1 5`; do gluster v create "vol_rep_$i" replica 2 storage_node1:/rhs/bricks/vol_rep_${i}_b0 storage_node2:/rhs/bricks/vol_rep_${i}_b1 --mode=script gluster v set "vol_rep_$i" self-heal-daemon off gluster v info "vol_rep_$i" gluster v start "vol_rep_$i" done 4. killall glusterfsd glusterd glusterfs on storage_node2. 5. On storage_node1 execute: rm -rf /rhs/bricks/* for i in `seq 1 5`; do gluster v stop vol_rep_${i} --mode=script gluster v delete vol_rep_${i} --mode=script gluster v create vol_rep_${i} replica 2 storage_node1:/rhs/bricks/vol_rep_${i}_b0 storage_node1:/rhs/bricks/vol_rep_${i}_b1 --mode=script gluster v set vol_rep_${i} self-heal-daemon on gluster v info vol_rep_${i} gluster v start vol_rep_${i} done 6. restart glusterd on storage_node2 (service glusterd start) 7. check the peer status from both the nodes. ( Both the nodes will be in "Peer Rejected" state for each other ). 8. From storage_node2 execute : +++++++++++++++++++++++++++++++++++ a). "gluster volume sync <storage_node1> vol_rep_1" Expected: "volume 'vol_rep_1' information should be synced from storage_node1 to storage_node2." Actual : as expected b. "gluster volume sync <storage_node1> all" Expected: "All volumes information should be synced from storage_node1 to storage_node2." Actual : as expected c. "gluster volume sync <storage_node2>" Expected: "volume sync: failed: sync from localhost not allowed " Actual : as expected ============================================================================== Case2 : ( 1 x 2 replicate volumes, 2 Storage nodes ) ============================================================================== 1. start glusterd on both the storage nodes. 2. peer probe storage_node2 ( from storage_node1 ) 3. On storage_node1 execute: for i in `seq 1 5`; do gluster v create "vol_rep_$i" replica 2 storage_node1:/rhs/bricks/vol_rep_${i}_b0 storage_node2:/rhs/bricks/vol_rep_${i}_b1 --mode=script gluster v set "vol_rep_$i" self-heal-daemon off gluster v info "vol_rep_$i" gluster v start "vol_rep_$i" done 4. killall glusterfsd glusterd glusterfs on storage_node2. 5. On storage_node1 execute: rm -rf /rhs/bricks/* for i in `seq 1 5`; do gluster v stop vol_rep_${i} --mode=script gluster v delete vol_rep_${i} --mode=script gluster v create vol_rep_${i} replica 2 storage_node1:/rhs/bricks/vol_rep_${i}_b0 storage_node1:/rhs/bricks/vol_rep_${i}_b1 --mode=script gluster v set vol_rep_${i} self-heal-daemon on gluster v info vol_rep_${i} gluster v start vol_rep_${i} done 6. restart glusterd on storage_node2 (service glusterd start) 7. check the peer status from both the nodes. ( Both the nodes will be in "Peer Rejected" state for each other ). 8. From storage_node1 execute : ++++++++++++++++++++++++++++++++++++ a. "gluster volume sync <storage_node2> vol_rep_1" Expected:"volume vol_rep_1 information should be synced from storage_node2 to storage_node1. Actual : as expected b. "gluster volume sync <storage_node2>" Expected: "All volumes information should be synced from storage_node2 to storage_node1." Actual : as expected ============================================================================== Note: ============================================================================== The above 2 cases verifies this bug. However the Peers remains in "Peer Rejected" state and the volumes are not restarted. This issue is tracked in the bug : https://bugzilla.redhat.com/show_bug.cgi?id=865700 Moving this bug from ON_QA to Verified state. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHBA-2013-1262.html Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHBA-2013-1262.html |