Bug 1024725 - For the nodes which does not participate in remove brick , remove brick status gives the ouptut of rebalance.
Summary: For the nodes which does not participate in remove brick , remove brick statu...
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Gluster Storage
Classification: Red Hat Storage
Component: glusterfs
Version: 2.1
Hardware: Unspecified
OS: Unspecified
high
high
Target Milestone: ---
: RHGS 2.1.2
Assignee: Krutika Dhananjay
QA Contact: RamaKasturi
URL:
Whiteboard:
Depends On:
Blocks: 1040371
TreeView+ depends on / blocked
 
Reported: 2013-10-30 10:30 UTC by RamaKasturi
Modified: 2015-05-13 16:26 UTC (History)
9 users (show)

Fixed In Version: glusterfs-3.4.0.50rhs-1
Doc Type: Bug Fix
Doc Text:
Previously, status from a previous remove-brick or Rebalance operation was not reset before starting a new remove-brick or Rebalance operation. As a result, remove-brick status displayed the output of a previous Rebalance operation on those nodes which did not participate in an ongoing remove-brick operation. With this update, the status of the remove-brick or Rebalance operation is set to NOT-STARTED before starting remove-brick or Rebalance operations again on all the nodes in the cluster.
Clone Of:
: 1040371 (view as bug list)
Environment:
Last Closed: 2014-02-25 07:58:03 UTC
Embargoed:


Attachments (Terms of Use)
Attaching screenshot (69.60 KB, image/png)
2013-12-03 13:05 UTC, RamaKasturi
no flags Details
Attaching screenshot (114.35 KB, image/png)
2013-12-03 13:06 UTC, RamaKasturi
no flags Details


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHEA-2014:0208 0 normal SHIPPED_LIVE Red Hat Storage 2.1 enhancement and bug fix update #2 2014-02-25 12:20:30 UTC

Description RamaKasturi 2013-10-30 10:30:42 UTC
Description of problem:
For the nodes which does not participate in remove brick , remove brick status gives the ouptut of rebalance.

Version-Release number of selected component (if applicable):
glusterfs-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-geo-replication-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-rdma-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-debuginfo-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-libs-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-fuse-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-server-3.4.0.35.1u2rhs-1.el6rhs.x86_64
glusterfs-api-3.4.0.35.1u2rhs-1.el6rhs.x86_64
samba-glusterfs-3.6.9-160.3.el6rhs.x86_64


How reproducible:
Always

Steps to Reproduce:
1. Create a distribute volume with 2 bricks.
2. start rebalance on the volume and stop it.
3. Now rebalance status shows stopped for the nodes where rebalance was running. Following is the ouput for the same.

[root@localhost ~]# gluster vol rebalance vol_dis_rep status
                                    Node Rebalanced-files          size       scanned      failures       skipped               status   run time in secs
                               ---------      -----------   -----------   -----------   -----------   -----------         ------------     --------------
                               localhost                2         2.0GB            61             0            20            completed              66.00
                            10.70.37.140                0        0Bytes            60             0             0            completed               0.00
                             10.70.37.75                0        0Bytes             0             0             0          not started               0.00
                             10.70.37.43                0        0Bytes             0             0             0              stopped               0.00
volume rebalance: vol_dis_rep: success: 

4. Now start remove brick.
5. Once started check the ouput. The following is what it displays.

[root@localhost ~]# gluster vol remove-brick vol_dis_rep 10.70.37.108:/rhs/brick3/b5 10.70.37.140:/rhs/brick3/b6  status
                                    Node Rebalanced-files          size       scanned      failures       skipped         status run-time in secs
                               ---------      -----------   -----------   -----------   -----------   -----------   ------------   --------------
                               localhost                2         2.0GB            61             0             0      completed            66.00
                            10.70.37.140                0        0Bytes            60             0             0      completed             0.00
                             10.70.37.75                0        0Bytes             0             0             0    not started             0.00
                             10.70.37.43                0        0Bytes             0             0             0        stopped             0.00



Actual results:
For the nodes on which remove brick is not started, it shows ouput of rebalance status.

Expected results:
For the nodes on which remove brick is not started it should show status as "notstarted" or the nodes which does not participate in the remove brick should not be shown in the status.

Additional info:

Comment 2 Dusmant 2013-10-30 11:08:51 UTC
RHSC remove brick status is showing wrong data... This needs to be fixed...

Comment 4 RamaKasturi 2013-12-03 13:05:23 UTC
Remove-brick status output still have the nodes where rebalance was stopped.

Attaching the screenshot for the same.

From screenshot 8, it is clear that rebalance was stopped on localhost and 10.70.37.182.

From screenshot 9, it is clear that remove-brick was started on 10.70.37.177 and 10.70.37.109.

10.70.37.182 was not at all participating in remove-brick but it was still shown in the remove-brick status which is incorrect.

So moving this back to assgined.

Comment 5 RamaKasturi 2013-12-03 13:05:57 UTC
Created attachment 832041 [details]
Attaching screenshot

Comment 6 RamaKasturi 2013-12-03 13:06:24 UTC
Created attachment 832042 [details]
Attaching screenshot

Comment 9 RamaKasturi 2013-12-19 05:08:48 UTC
verified and works fine with glusterfs-server-3.4.0.50rhs-1.el6rhs.x86_64.

Now remove-brick does not give the output of rebalance. Remove-brick status shows only the nodes which participates in remove-brick.

Comment 10 Pavithra 2014-01-07 07:02:54 UTC
Can you please review the doc text for technical accuracy?

Comment 11 Krutika Dhananjay 2014-01-07 07:05:17 UTC
LGTM, Pavithra.

Comment 13 errata-xmlrpc 2014-02-25 07:58:03 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

http://rhn.redhat.com/errata/RHEA-2014-0208.html


Note You need to log in before you can comment on or make changes to this bug.