Bug 765430 (GLUSTER-3698)

Summary: [gluserfs-3.3.0qa14]: nfs client hangs
Product: [Community] GlusterFS Reporter: Raghavendra Bhat <rabhat>
Component: replicateAssignee: Junaid <junaid>
Status: CLOSED CURRENTRELEASE QA Contact: Raghavendra Bhat <rabhat>
Severity: medium Docs Contact:
Priority: medium    
Version: pre-releaseCC: amarts, gluster-bugs, junaid, vagarwal, vijay
Target Milestone: ---   
Target Release: ---   
Hardware: x86_64   
OS: Linux   
Whiteboard:
Fixed In Version: glusterfs-3.4.0 Doc Type: Bug Fix
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2013-07-24 17:16:36 UTC Type: ---
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: glusterfs-3.3.0qa43 Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 817967    
Attachments:
Description Flags
nfs server statedump none

Description Raghavendra Bhat 2011-10-05 06:53:51 UTC
Created attachment 694

Comment 1 Raghavendra Bhat 2011-10-05 09:36:42 UTC
Created replicate volume with replica count.

2 fuse clients, 1 nfs client.

1 fuse client executing sanity script, nfs client executing the sanity script and the other fuse client executing iozone.

Added 2 more bricks to the volume thus making it 2x2 distributed replicate setup. enabled quota and profile.

nfs client hung.

Triggered the self-heal by bringing a brick down and up.

This is the statedump.


 [global.callpool.stack.1.frame.1]
ref_count=1
translator=nfs-server
complete=0

[global.callpool.stack.1.frame.2]
ref_count=0
translator=mirror-replicate-0
complete=0
parent=mirror-dht
wind_from=dht_writev
wind_to=subvol->fops->writev
unwind_to=dht_writev_cbk

[global.callpool.stack.1.frame.3]
ref_count=1
translator=mirror-dht
complete=0
parent=mirror-quota
wind_from=quota_writev_helper
wind_to=FIRST_CHILD(this)->fops->writev
unwind_to=quota_writev_cbk

[global.callpool.stack.1.frame.4]
ref_count=0
translator=mirror-client-0
complete=1
parent=mirror-replicate-0
wind_from=afr_getxattr
wind_to=children[call_child]->fops->getxattr
unwind_from=client3_1_getxattr_cbk
unwind_to=afr_getxattr_cbk

[global.callpool.stack.1.frame.5]
ref_count=0
translator=mirror-replicate-0
complete=1
parent=mirror-dht
wind_from=dht_getxattr
wind_to=subvol->fops->getxattr
unwind_from=afr_getxattr_cbk
unwind_to=dht_getxattr_cbk

Comment 2 Krishna Srinivas 2011-10-10 07:42:25 UTC
statedump indicates lot of pending frames from replicate. Pranith can you take a look? Assign it back to me in case it is an NFS bug.

Comment 3 Anand Avati 2011-12-22 13:10:21 UTC
CHANGE: http://review.gluster.com/807 (cluster/afr: Fix meta data lock range) merged in master by Vijay Bellur (vijay)

Comment 4 Anand Avati 2011-12-22 13:11:31 UTC
CHANGE: http://review.gluster.com/808 (features/locks: Fix release of locks on disconnect) merged in master by Vijay Bellur (vijay)

Comment 5 Anand Avati 2011-12-22 13:13:13 UTC
CHANGE: http://review.gluster.com/809 (protocol/server: Do connection cleanup if reply fails) merged in master by Vijay Bellur (vijay)

Comment 6 Anand Avati 2011-12-22 13:14:46 UTC
CHANGE: http://review.gluster.com/810 (libglusterfs: set inode path to be NULL in case of errors) merged in master by Vijay Bellur (vijay)

Comment 7 Raghavendra Bhat 2012-05-25 06:00:52 UTC
Checked with glsuterfs-3.3.0qa43. Did not find hang.