Bug 1041857 - rest api: rebalance causes hosts to become Non Operational after adding 4 bricks sequentially to a 6 node dist volume
Summary: rest api: rebalance causes hosts to become Non Operational after adding 4 bri...
Keywords:
Status: CLOSED NOTABUG
Alias: None
Product: Red Hat Gluster Storage
Classification: Red Hat
Component: rhsc-sdk
Version: 2.1
Hardware: Unspecified
OS: Unspecified
unspecified
urgent
Target Milestone: ---
: ---
Assignee: Shubhendu Tripathi
QA Contact: Dustin Tsang
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2013-12-12 19:44 UTC by Dustin Tsang
Modified: 2015-07-13 04:39 UTC (History)
7 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2013-12-13 03:26:54 UTC
Target Upstream Version:


Attachments (Terms of Use)
automation log (517.78 KB, text/plain)
2013-12-12 19:44 UTC, Dustin Tsang
no flags Details
vdsm.log (1.29 MB, text/x-log)
2013-12-12 19:47 UTC, Dustin Tsang
no flags Details
engine.log (1.16 MB, text/x-log)
2013-12-12 19:51 UTC, Dustin Tsang
no flags Details

Description Dustin Tsang 2013-12-12 19:44:27 UTC
Created attachment 836010 [details]
automation log

Description of problem:

rest api: rebalance causes hosts to become Non Operational after adding 4 bricks sequentially to a 6 node dist volume. 

* glusterd is down on each of the nodes
* vdsmd is up on each of the nodes
* glusterd process terminates briefly after attempts to start glusterd


Version-Release number of selected component (if applicable):
rhsc-cb11

How reproducible:
100% of the time

Steps to Reproduce:
1. setup a 2 node cluster 
2. via rest add a 6 brick dist volume and start the volume
3. via rest add 4 bricks one at a time sequentially to the volume
=> hosts are up and volume is fine.
4. via rest or gui start rebalance on the volume

Actual results:
Rebalance fails to start. Hosts are Non Operational. 

response given --
HTTP 400 connection failed. please check if gluster daemon is operational.

Expected results:
Rebalance starts successfully.

Additional info:

Comment 1 Dustin Tsang 2013-12-12 19:47:33 UTC
Created attachment 836012 [details]
vdsm.log

Comment 2 Dustin Tsang 2013-12-12 19:51:00 UTC
Created attachment 836013 [details]
engine.log

Comment 4 Dustin Tsang 2013-12-12 22:04:10 UTC
adding all bricks at once for step 4 in comment#0 also results in hosts becoming Non Operational.

Comment 5 Dustin Tsang 2013-12-12 22:05:28 UTC
correction to comment#4 s/step 4/step 3/

Comment 6 Dustin Tsang 2013-12-12 23:38:57 UTC
step 3 doesn't need to run to cause the the hosts to go non operational.

Comment 7 Dustin Tsang 2013-12-13 03:26:54 UTC
The issue seems to be with the volume name "StartMigrationDuringRebalanceTest".
Reopening this bug as a glusterfs bug.


Note You need to log in before you can comment on or make changes to this bug.