Bug 1223338 - glusterd could crash in remove-brick-status when local remove-brick process has just completed
Summary: glusterd could crash in remove-brick-status when local remove-brick process h...
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: GlusterFS
Classification: Community
Component: glusterd
Version: mainline
Hardware: Unspecified
OS: Unspecified
unspecified
unspecified
Target Milestone: ---
Assignee: krishnan parthasarathi
QA Contact:
URL:
Whiteboard:
Depends On:
Blocks: 1223634 1225318
TreeView+ depends on / blocked
 
Reported: 2015-05-20 11:20 UTC by krishnan parthasarathi
Modified: 2016-06-16 13:03 UTC (History)
4 users (show)

Fixed In Version: glusterfs-3.8rc2
Clone Of:
: 1223634 1225318 (view as bug list)
Environment:
Last Closed: 2016-06-16 13:03:20 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Embargoed:


Attachments (Terms of Use)

Description krishnan parthasarathi 2015-05-20 11:20:37 UTC
Description of problem:
Glusterd process could crash while executing remove-brick-status command around the time when the local remove-brick process (i.e, rebalance process) has completed migrating data. 

Version-Release number of selected component (if applicable):
mainline

How reproducible:
Intermittent

Steps to Reproduce:
1. Create and start a volume.
2. Add files/directories as required.
3. Remove one or more bricks using remove-brick-start command.
4. Issue remove-brick-status command around the time when the local remove-brick
   process is completed.

Actual results:
glusterd process crashes.

Expected results:
glusterd shouldn't crash. It would be helpful if the remove-brick-status command failed saying that the rebalance process may have just completed with
migration of data from the bricks being removed.

Additional info:
The above steps are representative of when the issue can be seen but not really helpful if you wish to automate this. The following link leads to the regression test, as part of GlusterFS regression test suite, that has hit this problem more often. This could help those interested in automation.

https://github.com/gluster/glusterfs/blob/master/tests/bugs/glusterd/bug-974007.t

Comment 1 Anand Avati 2015-05-21 07:11:46 UTC
REVIEW: http://review.gluster.org/10872 (glusterd: fix double-free of rebalance process' rpc object) posted (#1) for review on master by Krishnan Parthasarathi (kparthas)

Comment 2 Anand Avati 2015-05-22 03:13:13 UTC
REVIEW: http://review.gluster.org/10872 (glusterd: fix double-free of rebalance process' rpc object) posted (#2) for review on master by Krishnan Parthasarathi (kparthas)

Comment 3 Anand Avati 2015-05-26 09:58:45 UTC
REVIEW: http://review.gluster.org/10872 (glusterd: fix double-free of rebalance process' rpc object) posted (#3) for review on master by Krishnan Parthasarathi (kparthas)

Comment 4 Anand Avati 2015-05-26 13:42:02 UTC
COMMIT: http://review.gluster.org/10872 committed in master by Kaushal M (kaushal) 
------
commit f59a1431e38ad644b6bdd4adbdf94ff028c9c60d
Author: Krishnan Parthasarathi <kparthas>
Date:   Wed May 20 18:23:58 2015 +0530

    glusterd: fix double-free of rebalance process' rpc object
    
    Change-Id: I0c79c4de47a160b1ecf3a8994eedc02e3f5002a9
    BUG: 1223338
    Signed-off-by: Krishnan Parthasarathi <kparthas>
    Reviewed-on: http://review.gluster.org/10872
    Tested-by: NetBSD Build System
    Tested-by: Gluster Build System <jenkins.com>
    Reviewed-by: Atin Mukherjee <amukherj>
    Reviewed-by: Kaushal M <kaushal>

Comment 5 Anand Avati 2015-05-27 07:06:31 UTC
REVIEW: http://review.gluster.org/10932 (glusterd: fix double-free of rebalance process' rpc object) posted (#1) for review on release-3.7 by Krishnan Parthasarathi (kparthas)

Comment 6 Nagaprasad Sathyanarayana 2015-10-25 14:53:03 UTC
Fix for this BZ is already present in a GlusterFS release. You can find clone of this BZ, fixed in a GlusterFS release and closed. Hence closing this mainline BZ as well.

Comment 7 Niels de Vos 2016-06-16 13:03:20 UTC
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.8.0, please open a new bug report.

glusterfs-3.8.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution.

[1] http://blog.gluster.org/2016/06/glusterfs-3-8-released/
[2] http://thread.gmane.org/gmane.comp.file-systems.gluster.user


Note You need to log in before you can comment on or make changes to this bug.