Description of problem: If the volume is in "start" state, rebalance with gdeploy fails. Version-Release number of selected component (if applicable): gdeploy-1.0-12.el7rhgs.noarch How reproducible: Always Steps to Reproduce: 1. Create a conf file for running rebalance over a volume: cat gconf: [hosts] rhshdp03.lab.eng.blr.redhat.com rhshdp04.lab.eng.blr.redhat.com rhshdp05.lab.eng.blr.redhat.com rhshdp06.lab.eng.blr.redhat.com [tune-profile] none [volume] action=rebalance volname=gluster_vol1 state=start 2. Run rebalance with gdeploy: gdeploy -c gconf Actual results: Rebalance fails Additional info: # gdeploy -c gconf INFO: Volume management(action: rebalance) triggered PLAY [master] ***************************************************************** TASK: [Starts a volume] ******************************************************* failed: [rhshdp04.lab.eng.blr.redhat.com] => {"failed": true} msg: volume start: gluster_vol1: failed: Volume gluster_vol1 already started FATAL: all hosts have already failed -- aborting PLAY RECAP ******************************************************************** to retry, use: --limit @/root/ansible_playbooks.retry rhshdp04.lab.eng.blr.redhat.com : ok=0 changed=0 unreachable=0 failed=1
This bug exists in the present release and the cause of the issue is that rebalance operations tries to start the gluster volume before doing the actual rebalance. So if the volume is already started, rebalance would fail. As a workaround, one could stop the volume and then try doing the rebalance. In short, in the present release, rebalance in gdeploy is possible only for stopped volumes.
Edited the doc text, please review and sign off to be included in Known Issues chapter.
Looks good, Anjana.
Fixed in branch: https://github.com/gluster/gdeploy/tree/1.1
Verified with gdeploy-2.0-2.el7rhgs.noarch
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2016:1250