Bug 1024725

Summary: For the nodes which does not participate in remove brick , remove brick status gives the ouptut of rebalance.
Product: [Red Hat Storage] Red Hat Gluster Storage Reporter: RamaKasturi <knarra>
Component: glusterfsAssignee: Krutika Dhananjay <kdhananj>
Status: CLOSED ERRATA QA Contact: RamaKasturi <knarra>
Severity: high Docs Contact:
Priority: high    
Version: 2.1CC: dpati, dtsang, kdhananj, mmahoney, pprakash, psriniva, ssampat, vagarwal, vbellur
Target Milestone: ---Keywords: ZStream
Target Release: RHGS 2.1.2   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: glusterfs-3.4.0.50rhs-1 Doc Type: Bug Fix
Doc Text:
Previously, status from a previous remove-brick or Rebalance operation was not reset before starting a new remove-brick or Rebalance operation. As a result, remove-brick status displayed the output of a previous Rebalance operation on those nodes which did not participate in an ongoing remove-brick operation. With this update, the status of the remove-brick or Rebalance operation is set to NOT-STARTED before starting remove-brick or Rebalance operations again on all the nodes in the cluster.
Story Points: ---
Clone Of:
: 1040371 (view as bug list) Environment:
Last Closed: 2014-02-25 07:58:03 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 1040371    
Attachments:
Description Flags
Attaching screenshot
none
Attaching screenshot none

Description RamaKasturi 2013-10-30 10:30:42 UTC
Description of problem:
For the nodes which does not participate in remove brick , remove brick status gives the ouptut of rebalance.

Version-Release number of selected component (if applicable):
glusterfs-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-geo-replication-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-rdma-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-debuginfo-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-libs-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-fuse-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-server-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-api-3.4.0.35.1u2rhs-1.el6rhs.x86_64
samba-glusterfs-3.6.9-160.3.el6rhs.x86_64


How reproducible:
Always

Steps to Reproduce:
1. Create a distribute volume with 2 bricks.
2. start rebalance on the volume and stop it.
3. Now rebalance status shows stopped for the nodes where rebalance was running. Following is the ouput for the same.

[root@localhost ~]# gluster vol rebalance vol_dis_rep status
                                    Node Rebalanced-files          size       scanned      failures       skipped               status   run time in secs
                               ---------      -----------   -----------   -----------   -----------   -----------         ------------     --------------
                               localhost                2         2.0GB            61             0            20            completed              66.00
                            10.70.37.140                0        0Bytes            60             0             0            completed               0.00
                             10.70.37.75                0        0Bytes             0             0             0          not started               0.00
                             10.70.37.43                0        0Bytes             0             0             0              stopped               0.00
volume rebalance: vol_dis_rep: success: 

4. Now start remove brick.
5. Once started check the ouput. The following is what it displays.

[root@localhost ~]# gluster vol remove-brick vol_dis_rep 10.70.37.108:/rhs/brick3/b5 10.70.37.140:/rhs/brick3/b6  status
                                    Node Rebalanced-files          size       scanned      failures       skipped         status run-time in secs
                               ---------      -----------   -----------   -----------   -----------   -----------   ------------   --------------
                               localhost                2         2.0GB            61             0             0      completed            66.00
                            10.70.37.140                0        0Bytes            60             0             0      completed             0.00
                             10.70.37.75                0        0Bytes             0             0             0    not started             0.00
                             10.70.37.43                0        0Bytes             0             0             0        stopped             0.00



Actual results:
For the nodes on which remove brick is not started, it shows ouput of rebalance status.

Expected results:
For the nodes on which remove brick is not started it should show status as "notstarted" or the nodes which does not participate in the remove brick should not be shown in the status.

Additional info:

Comment 2 Dusmant 2013-10-30 11:08:51 UTC
RHSC remove brick status is showing wrong data... This needs to be fixed...

Comment 4 RamaKasturi 2013-12-03 13:05:23 UTC
Remove-brick status output still have the nodes where rebalance was stopped.

Attaching the screenshot for the same.

From screenshot 8, it is clear that rebalance was stopped on localhost and 10.70.37.182.

From screenshot 9, it is clear that remove-brick was started on 10.70.37.177 and 10.70.37.109.

10.70.37.182 was not at all participating in remove-brick but it was still shown in the remove-brick status which is incorrect.

So moving this back to assgined.

Comment 5 RamaKasturi 2013-12-03 13:05:57 UTC
Created attachment 832041 [details]
Attaching screenshot

Comment 6 RamaKasturi 2013-12-03 13:06:24 UTC
Created attachment 832042 [details]
Attaching screenshot

Comment 9 RamaKasturi 2013-12-19 05:08:48 UTC
verified and works fine with glusterfs-server-3.4.0.50rhs-1.el6rhs.x86_64.

Now remove-brick does not give the output of rebalance. Remove-brick status shows only the nodes which participates in remove-brick.

Comment 10 Pavithra 2014-01-07 07:02:54 UTC
Can you please review the doc text for technical accuracy?

Comment 11 Krutika Dhananjay 2014-01-07 07:05:17 UTC
LGTM, Pavithra.

Comment 13 errata-xmlrpc 2014-02-25 07:58:03 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

http://rhn.redhat.com/errata/RHEA-2014-0208.html