Bug 1233036

Summary: [AFR-V2] - Fix shd coredump from tests/bugs/glusterd/bug-948686.t
Product: [Community] GlusterFS Reporter: Krutika Dhananjay <kdhananj>
Component: replicateAssignee: Krutika Dhananjay <kdhananj>
Status: CLOSED CURRENTRELEASE QA Contact:
Severity: unspecified Docs Contact:
Priority: unspecified    
Version: 3.6.3CC: bugs, rabhat
Target Milestone: ---Keywords: Triaged
Target Release: ---   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: glusterfs-v3.6.4 Doc Type: Bug Fix
Doc Text:
Story Points: ---
Clone Of: 1229172 Environment:
Last Closed: 2016-02-04 15:27:21 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 1229172, 1233144    
Bug Blocks: 1216965, 1229550    

Description Krutika Dhananjay 2015-06-18 06:29:40 UTC
+++ This bug was initially created as a clone of Bug #1229172 +++

Description of problem:
http://www.gluster.org/pipermail/gluster-devel/2015-June/045499.html


Version-Release number of selected component (if applicable):


How reproducible:


Steps to Reproduce:
1.
2.
3.

Actual results:


Expected results:


Additional info:

--- Additional comment from Krutika Dhananjay on 2015-06-08 04:47:32 EDT ---

http://review.gluster.org/#/c/11119/

--- Additional comment from Anand Avati on 2015-06-08 23:24:33 EDT ---

COMMIT: http://review.gluster.org/11119 committed in master by Vijay Bellur (vbellur) 
------
commit 7ca78f7a6466a0f2ff19caff526f6560b5275f69
Author: Krutika Dhananjay <kdhananj>
Date:   Mon Jun 8 11:36:12 2015 +0530

    cluster/afr: Do not attempt entry self-heal if the last lookup on entry failed on src
    
    Test bug-948686.t was causing shd to dump core due to gfid being NULL.
    This was due to the volume being stopped while index heal's in progress,
    causing afr_selfheal_unlocked_lookup_on() to fail sometimes on the src brick
    with ENOTCONN. And when afr_selfheal_newentry_mark() copies the gfid off the
    src iatt, it essentially copies null gfid. This was causing the assertion
    as part of xattrop in protocol/client to fail.
    
    Change-Id: I237a0d6b1849e4c48d7645a2cc16d9bc1441ef95
    BUG: 1229172
    Signed-off-by: Krutika Dhananjay <kdhananj>
    Reviewed-on: http://review.gluster.org/11119
    Tested-by: Gluster Build System <jenkins.com>
    Reviewed-by: Pranith Kumar Karampuri <pkarampu>
    Reviewed-by: Vijay Bellur <vbellur>

Comment 1 Anand Avati 2015-06-18 07:03:39 UTC
REVIEW: http://review.gluster.org/11309 (cluster/afr: Do not attempt entry self-heal if the last lookup on entry failed on src) posted (#1) for review on release-3.6 by Krutika Dhananjay (kdhananj)

Comment 2 Anand Avati 2015-06-19 06:38:13 UTC
COMMIT: http://review.gluster.org/11309 committed in release-3.6 by Raghavendra Bhat (raghavendra) 
------
commit d86a238f29c1519bad37bd38d12227bd69d1947f
Author: Krutika Dhananjay <kdhananj>
Date:   Mon Jun 8 11:36:12 2015 +0530

    cluster/afr: Do not attempt entry self-heal if the last lookup on entry failed on src
    
            Backport of: http://review.gluster.org/11119
    
    Test bug-948686.t was causing shd to dump core due to gfid being NULL.
    This was due to the volume being stopped while index heal's in progress,
    causing afr_selfheal_unlocked_lookup_on() to fail sometimes on the src brick
    with ENOTCONN. And when afr_selfheal_newentry_mark() copies the gfid off the
    src iatt, it essentially copies null gfid. This was causing the assertion
    as part of xattrop in protocol/client to fail.
    
    Change-Id: I81723567af824ce4a9fa37e309eeeab8404ac71e
    BUG: 1233036
    Signed-off-by: Krutika Dhananjay <kdhananj>
    Reviewed-on: http://review.gluster.org/11309
    Tested-by: Gluster Build System <jenkins.com>
    Reviewed-by: Pranith Kumar Karampuri <pkarampu>
    Reviewed-by: Raghavendra Bhat <raghavendra>

Comment 3 Kaushal 2016-02-04 15:27:21 UTC
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-v3.6.4, please open a new bug report.

glusterfs-v3.6.4 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution.

[1] https://www.gluster.org/pipermail/gluster-users/2015-July/022826.html
[2] http://thread.gmane.org/gmane.comp.file-systems.gluster.user