Description of problem: Glusterfs rebalance xml output is null although gluster rebalance status returns rebalance status. Version-Release number of selected component (if applicable): How reproducible: Sometimes Steps to Reproduce: 1.Create a distribute volume with atleast 1 brick each from 4 hosts. 2.Start rebalance on the volume. 3.Before rebalance completes, stop glusterd on one of the hosts. 4.Check rebalance status "xml" returned from gluster cli on other hosts. Actual results: Displays/Returns nothing. Expected results: Should display/return the xml output Additional info: This is required by vdsm and hence rhsc.
This issue is blocking two of the RHSC bugs... It needs to be fixed.
Asking Aravinda to look at this, because Kaushal is already looking at bunch of other issues.
Occurs for remove-brick status also.
This was an regression introduced by a change done to introduce consistent ordering of rebalance status results. Prior to the change, the index was incremented sequentially on each response to a rebalance status received from a peer. This meant that there were no holes in the indices, and the cli output code was written to handle the indices in this manner. With the change, each peer is given a consistent index, which results in the indices having holes when one or more of the peers are down. Since the cli output code was not changed to match this, we have the issue observed in this bug. The difference seen between the normal and xml output is caused because the xml output code is written to display all the available information or nothing at all, whereas the normal cli output displays the available information. I have informed Aravinda of this and he has agreed to make the required cli changes to get the outputs working again.
Aravinda, Can you provide patch url ?
@Satheesaran, patch link is already updated in External Trackers section. let me know if you need upstream patch url.
Thanks Aravinda for comment 7 Verified this bug with glusterfs-3.4.0.51rhs.el6rhs Performed the following steps: 1. Created trusted storage pool of 3 RHSS Nodes (i.e) gluster peer probe <host-ip> 2. Created a plain distribute volume with 1 brick, with 1 brick per RHSS Node (ie) gluster volume create <vol-name> <server1>:<brick1> 3. Started the volume (i.e) gluster volume start <vol-name> 4. Fuse mounted the volume on a RHEL 6.5 client with glusterfs-3.4.0.51rhs.el6_4 (i.e) mount.glusterfs <RHSS-IP>:<vol-name> <mount-point> 5. Write some 200 files with 41 MB on the mount point (i.e) for i in {1..200}; do dd if=/dev/urandom of=file$i bs=4k count=10000;done 6. Add 2 more bricks to the volume ( one brick per RHSS Node ) (i.e) gluster volume add-brick <vol-name> <server2>:<brick2> <server3>:<brick3> 7. Start rebalance on the volume (i.e) gluster volume start rebalance <vol-name> start 8. While step 7 is in progress, stop glusterd on the third RHSS Node (i.e) service glusted stop 9. Try to xml dump of gluster volume status from other nodes, where glusterd is UP (i.e) gluster volume status all --xml xml dump was successful even when glusterd was down in a node
Can you please verify the doc text for technical accuracy?
Doc Text looks good to me.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHEA-2014-0208.html