Bug 1325841 - Volume stop is failing when one of brick is down due to underlying filesystem crash
Summary: Volume stop is failing when one of brick is down due to underlying filesystem...
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: GlusterFS
Classification: Community
Component: glusterd
Version: mainline
Hardware: x86_64
OS: Linux
unspecified
high
Target Milestone: ---
Assignee: Atin Mukherjee
QA Contact:
URL:
Whiteboard:
Depends On: 1325750
Blocks: 1326174
TreeView+ depends on / blocked
 
Reported: 2016-04-11 10:42 UTC by Atin Mukherjee
Modified: 2016-06-16 14:03 UTC (History)
2 users (show)

Fixed In Version: glusterfs-3.8rc2
Doc Type: Bug Fix
Doc Text:
Clone Of: 1325750
: 1326174 (view as bug list)
Environment:
Last Closed: 2016-06-16 14:03:04 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Embargoed:


Attachments (Terms of Use)

Description Atin Mukherjee 2016-04-11 10:42:12 UTC
+++ This bug was initially created as a clone of Bug #1325750 +++

Description of problem:
=======================
When one of brick is down due to underlying filesystem (here xfs) crash, volume stop is failing with error message "volume stop: Dis: failed: brick operations failed"


Version-Release number of selected component (if applicable):
=============================================================
glusterfs-3.7.9-1

How reproducible:
=================
Always


Steps to Reproduce:
===================
1. Have one node cluster
2. Create 1*2 volume and start it.
3. crash underlying filesystem for one of the volume brick using "godown tool"
4. Check brick is down using "volume status"
5. Try to stop the volume // it will fail.


Actual results:
===============
Volume stop is failing with error message "volume stop: Dis: failed: brick operations failed"

Expected results:
=================
Volume stop should work properly.


Additional info:
================

--- Additional comment from Red Hat Bugzilla Rules Engine on 2016-04-11 01:46:29 EDT ---

This bug is automatically being proposed for the current z-stream release of Red Hat Gluster Storage 3 by setting the release flag 'rhgs‑3.1.z' to '?'. 

If this bug should be proposed for a different release, please manually change the proposed release flag.

--- Additional comment from RHEL Product and Program Management on 2016-04-11 02:02:20 EDT ---

This bug report has Keywords: Regression or TestBlocker.

Since no regressions or test blockers are allowed between releases,
it is also being identified as a blocker for this release.

Please resolve ASAP.

Comment 1 Vijay Bellur 2016-04-11 10:45:07 UTC
REVIEW: http://review.gluster.org/13965 (glusterd: populate brickinfo->real_path conditionally) posted (#1) for review on master by Atin Mukherjee (amukherj)

Comment 2 Vijay Bellur 2016-04-11 18:39:35 UTC
COMMIT: http://review.gluster.org/13965 committed in master by Jeff Darcy (jdarcy) 
------
commit d129d4eea33aae5db24dba17adcb04e9d4829817
Author: Atin Mukherjee <amukherj>
Date:   Mon Apr 11 16:07:40 2016 +0530

    glusterd: populate brickinfo->real_path conditionally
    
    glusterd_brickinfo_new_from_brick () is called from multiple places and one of
    them is glusterd_brick_rpc_notify where its very well possible that an
    underlying brick's file system has crashed and a disconnect event has been
    received. In this case glusterd tries to build the brickinfo from the brickid in
    the RPC request, however the same fails as glusterd_brickinfo_new_from_brick ()
    fails from realpath.
    
    Fix is to skip populating real_path if its a disconnect event.
    
    Change-Id: I9d9149c64a9cf2247abb731f219c1b1eef037960
    BUG: 1325841
    Signed-off-by: Atin Mukherjee <amukherj>
    Reviewed-on: http://review.gluster.org/13965
    Smoke: Gluster Build System <jenkins.com>
    CentOS-regression: Gluster Build System <jenkins.com>
    NetBSD-regression: NetBSD Build System <jenkins.org>
    Reviewed-by: Jeff Darcy <jdarcy>

Comment 3 Niels de Vos 2016-06-16 14:03:04 UTC
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.8.0, please open a new bug report.

glusterfs-3.8.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution.

[1] http://blog.gluster.org/2016/06/glusterfs-3-8-released/
[2] http://thread.gmane.org/gmane.comp.file-systems.gluster.user


Note You need to log in before you can comment on or make changes to this bug.