Description of problem: ---------------------- 4 node cluster.Mounted the volume via v4 on 4 clients.Created a huge deep dir data set. Ran rm -rf <mount-point>/* from multiple clients.It should have cleaned the mount point completely.But that wasn't the case. It threw a lot of "Stale File Handle: messages on the application side,which I understand is expected from BZ#1396776. But it left 143 directories as it is ,without deleting.There were no files though : [root@gqac010 gluster-mount]# ll total 12 drwxr-xr-x 3 root root 4096 Nov 26 07:02 d1 drwxr-xr-x 3 root root 4096 Nov 26 03:40 d2 drwxrwxr-x 3 root root 4096 Nov 26 07:09 linux-4.8.9 [root@gqac010 gluster-mount]# [root@gqac010 gluster-mount]# find . -mindepth 1 -type f [root@gqac010 gluster-mount]# find . -mindepth 1 -type d | wc -l 143 [root@gqac010 gluster-mount]# I see this on SSL as well as non SSL environment. Version-Release number of selected component (if applicable): ------------------------------------------------------------ nfs-ganesha-gluster-2.4.1-1.el7rhgs.x86_64 glusterfs-ganesha-3.8.4-5.el7rhgs.x86_64 How reproducible: ----------------- 2/2 Steps to Reproduce: ------------------- Given in description. Actual results: -------------- rm -rf * does not clean up properly. Expected results: ----------------- rm -rf * should clean up everything. Additional info: ---------------- OS : RHEL 7.3 *Vol Config* : Volume Name: testvol Type: Distributed-Replicate Volume ID: efe39b14-0eed-498c-b3cd-3946e7f9769c Status: Started Snapshot Count: 0 Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: gqas014.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick0 Brick2: gqas008.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick1 Brick3: gqas016.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick2 Brick4: gqas009.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick3 Options Reconfigured: ganesha.enable: on features.cache-invalidation: on nfs.disable: on performance.readdir-ahead: on transport.address-family: inet performance.stat-prefetch: off server.allow-insecure: on nfs-ganesha: enable cluster.enable-shared-storage: enable
Quick update. I ran rm -rf from multiple FUSE mounts(2*2 volume),and it cleared everything from the mount point.
Ravi and myself have taken a look at the setup and logs provided. We do not see any obvious errors logged or issues. Maybe we should try to reproduce and take tcpdump and observer traffic between NFS-client/NFS-ganesha and Gluster server.
Hi jiffin, I have updated the doc text for the release notes. Let me know if this looks ok, and if I got the workaround correctly.
Known Issue. Behaviour comes from the client side
The needinfo request[s] on this closed bug have been removed as they have been unresolved for 1000 days