Bug 765430 (GLUSTER-3698) - [gluserfs-3.3.0qa14]: nfs client hangs
Summary: [gluserfs-3.3.0qa14]: nfs client hangs
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: GLUSTER-3698
Product: GlusterFS
Classification: Community
Component: replicate
Version: pre-release
Hardware: x86_64
OS: Linux
medium
medium
Target Milestone: ---
Assignee: Junaid
QA Contact: Raghavendra Bhat
URL:
Whiteboard:
Depends On:
Blocks: 817967
TreeView+ depends on / blocked
 
Reported: 2011-10-05 09:36 UTC by Raghavendra Bhat
Modified: 2013-08-06 22:37 UTC (History)
5 users (show)

Fixed In Version: glusterfs-3.4.0
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2013-07-24 17:16:36 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions: glusterfs-3.3.0qa43
Embargoed:


Attachments (Terms of Use)
nfs server statedump (22.25 KB, application/x-bzip)
2011-10-05 06:53 UTC, Raghavendra Bhat
no flags Details

Description Raghavendra Bhat 2011-10-05 06:53:51 UTC
Created attachment 694

Comment 1 Raghavendra Bhat 2011-10-05 09:36:42 UTC
Created replicate volume with replica count.

2 fuse clients, 1 nfs client.

1 fuse client executing sanity script, nfs client executing the sanity script and the other fuse client executing iozone.

Added 2 more bricks to the volume thus making it 2x2 distributed replicate setup. enabled quota and profile.

nfs client hung.

Triggered the self-heal by bringing a brick down and up.

This is the statedump.


 [global.callpool.stack.1.frame.1]
ref_count=1
translator=nfs-server
complete=0

[global.callpool.stack.1.frame.2]
ref_count=0
translator=mirror-replicate-0
complete=0
parent=mirror-dht
wind_from=dht_writev
wind_to=subvol->fops->writev
unwind_to=dht_writev_cbk

[global.callpool.stack.1.frame.3]
ref_count=1
translator=mirror-dht
complete=0
parent=mirror-quota
wind_from=quota_writev_helper
wind_to=FIRST_CHILD(this)->fops->writev
unwind_to=quota_writev_cbk

[global.callpool.stack.1.frame.4]
ref_count=0
translator=mirror-client-0
complete=1
parent=mirror-replicate-0
wind_from=afr_getxattr
wind_to=children[call_child]->fops->getxattr
unwind_from=client3_1_getxattr_cbk
unwind_to=afr_getxattr_cbk

[global.callpool.stack.1.frame.5]
ref_count=0
translator=mirror-replicate-0
complete=1
parent=mirror-dht
wind_from=dht_getxattr
wind_to=subvol->fops->getxattr
unwind_from=afr_getxattr_cbk
unwind_to=dht_getxattr_cbk

Comment 2 Krishna Srinivas 2011-10-10 07:42:25 UTC
statedump indicates lot of pending frames from replicate. Pranith can you take a look? Assign it back to me in case it is an NFS bug.

Comment 3 Anand Avati 2011-12-22 13:10:21 UTC
CHANGE: http://review.gluster.com/807 (cluster/afr: Fix meta data lock range) merged in master by Vijay Bellur (vijay)

Comment 4 Anand Avati 2011-12-22 13:11:31 UTC
CHANGE: http://review.gluster.com/808 (features/locks: Fix release of locks on disconnect) merged in master by Vijay Bellur (vijay)

Comment 5 Anand Avati 2011-12-22 13:13:13 UTC
CHANGE: http://review.gluster.com/809 (protocol/server: Do connection cleanup if reply fails) merged in master by Vijay Bellur (vijay)

Comment 6 Anand Avati 2011-12-22 13:14:46 UTC
CHANGE: http://review.gluster.com/810 (libglusterfs: set inode path to be NULL in case of errors) merged in master by Vijay Bellur (vijay)

Comment 7 Raghavendra Bhat 2012-05-25 06:00:52 UTC
Checked with glsuterfs-3.3.0qa43. Did not find hang.


Note You need to log in before you can comment on or make changes to this bug.