Bug 1066130 - simultaneous start rebalance only starts rebalance for one volume for volumes made up of 16 hosts
Summary: simultaneous start rebalance only starts rebalance for one volume for volumes...
Keywords:
Status: CLOSED EOL
Alias: None
Product: Red Hat Gluster Storage
Classification: Red Hat Storage
Component: rhsc
Version: 2.1
Hardware: Unspecified
OS: Unspecified
low
unspecified
Target Milestone: ---
: ---
Assignee: Sahina Bose
QA Contact: RHS-C QE
URL:
Whiteboard:
Depends On:
Blocks: 1035040
TreeView+ depends on / blocked
 
Reported: 2014-02-17 19:23 UTC by Dustin Tsang
Modified: 2015-12-03 17:13 UTC (History)
7 users (show)

Fixed In Version:
Doc Type: Known Issue
Doc Text:
Simultaneous start of rebalance on volumes that span same set of hosts fails as glusterd lock is acquired on participating hosts. Workaround: Start Rebalance again on the other volume after the process starts on first volume.
Clone Of:
Environment:
Last Closed: 2015-12-03 17:13:38 UTC
Embargoed:


Attachments (Terms of Use)

Description Dustin Tsang 2014-02-17 19:23:39 UTC
Description of problem:

On volumes made up of bricks on 16 distinct hosts, starting rebalanace on 3 volumes simultaneously fails on 2 out 3 of the volumes.



Version-Release number of selected component (if applicable):
rhsc-cb17

How reproducible:
every time

Steps to Reproduce:
1. create 3 volumes which has bricks on 16 distinct hosts
2. ctrl click to select 3 hosts in the main volume tab
3. right click on on of the selected hosts and choose rebalance from the context menu


Actual results:
'could not start Gluster volume x rebalance' event reported for 2 of the volumes

rhsc-log-collector logs: http://rhsqe-repo.lab.eng.blr.redhat.com/dustin/sosreport-LogCollector-20140217141607.tar.xz

Expected results:
rebalance starts on all 3 volumes

Additional info:

Comment 2 Sahina Bose 2014-02-18 06:21:03 UTC
From vdsm logs, there's an exception

on vol_three
GlusterVolumeRebalanceStartFailedException: <unprintable GlusterVolumeRebalanceStartFailedException object>

on vol_dis
GlusterVolumeRebalanceStartFailedException: Volume rebalance start failed
error: Another transaction could be in progress. Please try again after sometime.
return code: 61

I'm not sure which prior operation on volume has obtained lock - which causes this. Will need help from gluster team to analyse.

Comment 3 Sahina Bose 2014-02-19 06:26:36 UTC
If all 3 volumes being started span the same set of hosts, simultaneous start fails as the gluster CLI obtains a lock on the host for the start rebalance command.

Comment 4 Shalaka 2014-02-19 10:57:36 UTC
Edited the doc text as discussed.

Comment 6 Vivek Agarwal 2015-12-03 17:13:38 UTC
Thank you for submitting this issue for consideration in Red Hat Gluster Storage. The release for which you requested us to review, is now End of Life. Please See https://access.redhat.com/support/policy/updates/rhs/

If you can reproduce this bug against a currently maintained version of Red Hat Gluster Storage, please feel free to file a new report against the current release.


Note You need to log in before you can comment on or make changes to this bug.