Bug 1059237 - [RHSC] Skipped file count in remove-brick status dialog not shown
Summary: [RHSC] Skipped file count in remove-brick status dialog not shown
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Gluster Storage
Classification: Red Hat Storage
Component: rhsc
Version: 2.1
Hardware: Unspecified
OS: Unspecified
medium
high
Target Milestone: ---
: RHGS 2.1.2
Assignee: Susant Kumar Palai
QA Contact: Shruti Sampat
URL:
Whiteboard:
Depends On:
Blocks: 1060208 1060209
TreeView+ depends on / blocked
 
Reported: 2014-01-29 13:29 UTC by Shruti Sampat
Modified: 2015-05-15 18:15 UTC (History)
10 users (show)

Fixed In Version: glusterfs-3.4.0.59rhs-1.el6rhs
Doc Type: Bug Fix
Doc Text:
Clone Of:
: 1060208 1060209 (view as bug list)
Environment:
Last Closed: 2014-02-25 08:15:33 UTC
Embargoed:


Attachments (Terms of Use)
screenshot (101.77 KB, image/png)
2014-01-29 13:29 UTC, Shruti Sampat
no flags Details
engine logs (6.42 MB, text/x-log)
2014-01-29 13:31 UTC, Shruti Sampat
no flags Details


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHEA-2014:0208 0 normal SHIPPED_LIVE Red Hat Storage 2.1 enhancement and bug fix update #2 2014-02-25 12:20:30 UTC

Description Shruti Sampat 2014-01-29 13:29:32 UTC
Created attachment 857054 [details]
screenshot

Description of problem:
-------------------------

Skipped file count is not displayed in the remove-brick status dialog, even though 'gluster volume remove-brick status' shows skipped files.

See below - 

[root@rhs glusterfs_58]# gluster v remove-brick dis_rep_vol 10.70.37.70:/rhs/brick3/b1 10.70.37.162:/rhs/brick4/b1 10.70.37.70:/rhs/brick4/b1 10.70.37.162:/rhs/brick3/b1 status
                                    Node Rebalanced-files          size       scanned      failures       skipped               status   run time in secs
                               ---------      -----------   -----------   -----------   -----------   -----------         ------------     --------------
                               localhost                5         2.9GB            36             0            30          in progress             161.00
                             10.70.37.70                0        0Bytes           650             0             0            completed               5.00

See screenshot for status dialog on the Console.

Version-Release number of selected component (if applicable):
Red Hat Storage Console Version: 2.1.2-0.35.el6rhs
glusterfs 3.4.0.58rhs

How reproducible:
Always

Steps to Reproduce:
1. Start remove-brick on a distribute-replicate volume such that there would be skipped files ( could be lack of space in the destination bricks )
2. See gluster CLI remove-brick status output for skipped file count.
3. Check the status dialog in the UI.

Actual results:
Skipped file count not displayed in the UI.

Expected results:
All the data shown in the UI should match the gluster CLI output.

Additional info:

Comment 1 Shruti Sampat 2014-01-29 13:31:47 UTC
Created attachment 857055 [details]
engine logs

Comment 5 Ramesh N 2014-01-31 04:30:30 UTC
There is difference in gluster cli output with --xml flag.

Skipped count is not returned in xml output.

[root@rhs ~]# gluster volume remove-brick dis_rep_vol 10.70.37.162:/rhs/brick3/b1 status
                                    Node Rebalanced-files          size       scanned      failures       skipped               status   run time in secs
                               ---------      -----------   -----------   -----------   -----------   -----------         ------------     --------------
                               localhost              104        33.2GB           751             0            30            completed             772.00
                             10.70.37.70                0        0Bytes           650             0             0            completed               4.00


[root@rhs ~]# gluster volume remove-brick dis_rep_vol 10.70.37.162:/rhs/brick3/b1 status --xml
<?xml version="1.0" encoding="UTF-8" standalone="yes"?>
<cliOutput>
  <opRet>0</opRet>
  <opErrno>0</opErrno>
  <opErrstr/>
  <volRemoveBrick>
    <task-id>97498624-6c35-4ab4-a878-2ce45a52a79d</task-id>
    <nodeCount>4</nodeCount>
    <node>
      <nodeName>localhost</nodeName>
      <id>706a5135-4737-48ee-9577-300d54b60ff6</id>
      <files>104</files>
      <size>35683081216</size>
      <lookups>751</lookups>
      <failures>0</failures>
      <skipped>0</skipped>
      <status>3</status>
      <statusStr>completed</statusStr>
      <runtime>772.00</runtime>
    </node>
    <node>
      <nodeName>10.70.37.70</nodeName>
      <id>bf570c26-c148-498a-9772-5d943ba81418</id>
      <files>0</files>
      <size>0</size>
      <lookups>650</lookups>
      <failures>0</failures>
      <skipped>0</skipped>
      <status>3</status>
      <statusStr>completed</statusStr>
      <runtime>4.00</runtime>
    </node>
    <aggregate>
      <files>104</files>
      <size>35683081216</size>
      <lookups>1401</lookups>
      <failures>0</failures>
      <skipped>0</skipped>
      <status>3</status>
      <statusStr>completed</statusStr>
      <runtime>772.00</runtime>
    </aggregate>
  </volRemoveBrick>
</cliOutput>
[root@rhs ~]#

Comment 6 Dusmant 2014-01-31 05:47:38 UTC
KP from gluster team is looking into this bug now.

Comment 7 Susant Kumar Palai 2014-01-31 16:59:56 UTC
Upstream Patch Sent at http://review.gluster.org/#/c/6882/


Result of cli-xml output after the change: 

[root@vm1 home]# gluster v remove-brick test1 192.168.122.240:/brick1/1 status
                                    Node Rebalanced-files          size       scanned      failures       skipped               status   run time in secs
                               ---------      -----------   -----------   -----------   -----------   -----------         ------------     --------------
                               localhost                0        0Bytes             2             0             1            completed               0.00
[root@vm1 home]# gluster v remove-brick test1 192.168.122.240:/brick1/1 status --xml
<?xml version="1.0" encoding="UTF-8" standalone="yes"?>
<cliOutput>
  <opRet>0</opRet>
  <opErrno>0</opErrno>
  <opErrstr/>
  <volRemoveBrick>
    <task-id>add66841-b53e-4a85-b8ae-ceceffe35b95</task-id>
    <nodeCount>1</nodeCount>
    <node>
      <nodeName>localhost</nodeName>
      <id>539073ad-e77c-44ee-bef7-84e3ac232a29</id>
      <files>0</files>
      <size>0</size>
      <lookups>2</lookups>
      <failures>0</failures>
      <skipped>1</skipped>
      <status>3</status>
      <statusStr>completed</statusStr>
      <runtime>0.00</runtime>
    </node>
    <aggregate>
      <files>0</files>
      <size>0</size>
      <lookups>2</lookups>
      <failures>0</failures>
      <skipped>1</skipped>
      <status>3</status>
      <statusStr>completed</statusStr>
      <runtime>0.00</runtime>
    </aggregate>
  </volRemoveBrick>
</cliOutput>
[root@vm1 home]#

Comment 9 RamaKasturi 2014-02-11 10:40:53 UTC
I see the following in the patch

cli/cli-xml :  skipped files should be treated as failures for
remove-brick operation.
Fix: For remove-brick operation skipped count is included into
failure count.
clixml-output : skipped count would be zero always for remove-brick
status.

If the above is the case, why do we need to have the skipped file count field? we can simply hide it right?

Comment 10 Susant Kumar Palai 2014-02-11 11:03:16 UTC
RamaKasturi ,
     The cli (non-xml) part uses the same function for remove-brick as well as rebalance status(To keep the code simple). And the intention was to keep the status fields for rebalance and remove-brick same.

Comment 11 RamaKasturi 2014-02-12 06:56:14 UTC
Hi Dusmant ,

  I see that the following change with this fix.

cli/cli-xml :  skipped files should be treated as failures for
remove-brick operation.
Fix: For remove-brick operation skipped count is included into
failure count.
clixml-output : skipped count would be zero always for remove-brick
status.

  Is it required to show the skipped file count field if that does not get updated at all in the status dialog for remove-brick.

Comment 12 RamaKasturi 2014-02-13 07:02:00 UTC
As per the patch submitted and comment 10, it works fine with the build

RHSC : rhsc-2.1.2-0.36.el6rhs.noarch

glusterfs : glusterfs-server-3.4.0.59rhs-1.el6rhs.x86_64

vdsm : vdsm-4.13.0-24.el6rhs.x86_64

Skipped file count is always displayed as zero, and for remove-brick operation skipped count is included into failure count.

Raised a new bug to remove the skipped file count field from the remove-brick status dialog.

https://bugzilla.redhat.com/show_bug.cgi?id=1064712

Comment 13 Dusmant 2014-02-13 10:13:47 UTC
We already discussed this and provided the info.

Comment 15 errata-xmlrpc 2014-02-25 08:15:33 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

http://rhn.redhat.com/errata/RHEA-2014-0208.html


Note You need to log in before you can comment on or make changes to this bug.