Description of problem: cd to .snaps fails with "transport endpoint not connected" after force start of the volume Version-Release number of selected component (if applicable): glusterfs-3.7.5-19 How reproducible: Always Steps to Reproduce: 1.Create a volume and start it and mount it on the client. 2.Enable USS on the volume. 3.cd to .snaps and make sure it works fine without any issues. 4.kill snapd on one node and try to cd .snaps and make sure it works fine. 5.kill snapd on second node and try to cd .snaps and make sure it works fine. 6.kill snapd on third node and try to cd .snaps and make sure it works fine. 7.kill snapd on fourth node from which the volume is mounted and try to cd .snaps and observe that if fails with "transport endpoint no connected" which is expected. 8. Now force start the volume. 9. Observe that uss is enabled on the volume and snapd is running on all the nodes after force start, however cd to .snaps fails with "transport endpoint not connected". Actual results: After force start of the volume, even though snapd is running on all the nodes and uss is enabled, cd to .snaps fails with "transport endpoint not connected" Expected results: After force restart of the volume, cd to .snaps should work when snapd is running on all the nodes. Additional info:
Merged in master and 3.7
Master URL: http://review.gluster.org/#/c/13420/ (MERGED) Release 3.7 URL: http://review.gluster.org/#/c/13457/ (MERGED)
When all the snapd processes are down on each node [root@dhcp46-4 ~]# gluster v status Status of volume: vol0 Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick 10.70.46.4:/run/gluster/snaps/c7fafe9 d394b4feaae6fd35391ae192c/brick1/b1 49168 0 Y 22010 Brick 10.70.47.46:/run/gluster/snaps/c7fafe 9d394b4feaae6fd35391ae192c/brick2/b2 49168 0 Y 22424 Brick 10.70.46.213:/run/gluster/snaps/c7faf e9d394b4feaae6fd35391ae192c/brick3/b3 49168 0 Y 19382 Brick 10.70.46.148:/run/gluster/snaps/c7faf e9d394b4feaae6fd35391ae192c/brick4/b4 49168 0 Y 16248 Snapshot Daemon on localhost N/A N/A N N/A NFS Server on localhost 2049 0 Y 14297 Self-heal Daemon on localhost N/A N/A Y 14307 Snapshot Daemon on 10.70.46.148 N/A N/A N N/A NFS Server on 10.70.46.148 2049 0 Y 7211 Self-heal Daemon on 10.70.46.148 N/A N/A Y 7221 Snapshot Daemon on 10.70.47.46 N/A N/A N N/A NFS Server on 10.70.47.46 2049 0 Y 14901 Self-heal Daemon on 10.70.47.46 N/A N/A Y 14911 Snapshot Daemon on 10.70.46.213 N/A N/A N N/A NFS Server on 10.70.46.213 2049 0 Y 11111 Self-heal Daemon on 10.70.46.213 N/A N/A Y 11123 On client ===================================== [root@dhcp46-47 fuse]# cd .test -bash: cd: .test: Transport endpoint is not connected [root@dhcp46-4 ~]# gluster v start vol0 force volume start: vol0: success [root@dhcp46-4 ~]# gluster v status Status of volume: vol0 Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick 10.70.46.4:/run/gluster/snaps/c7fafe9 d394b4feaae6fd35391ae192c/brick1/b1 49168 0 Y 22010 Brick 10.70.47.46:/run/gluster/snaps/c7fafe 9d394b4feaae6fd35391ae192c/brick2/b2 49168 0 Y 22424 Brick 10.70.46.213:/run/gluster/snaps/c7faf e9d394b4feaae6fd35391ae192c/brick3/b3 49168 0 Y 19382 Brick 10.70.46.148:/run/gluster/snaps/c7faf e9d394b4feaae6fd35391ae192c/brick4/b4 49168 0 Y 16248 Snapshot Daemon on localhost 49172 0 Y 14485 NFS Server on localhost 2049 0 Y 14493 Self-heal Daemon on localhost N/A N/A Y 14501 Snapshot Daemon on 10.70.47.46 49172 0 Y 15052 NFS Server on 10.70.47.46 2049 0 Y 15061 Self-heal Daemon on 10.70.47.46 N/A N/A Y 15070 Snapshot Daemon on 10.70.46.213 49172 0 Y 11243 NFS Server on 10.70.46.213 2049 0 Y 11251 Self-heal Daemon on 10.70.46.213 N/A N/A Y 11260 Snapshot Daemon on 10.70.46.148 49172 0 Y 7369 NFS Server on 10.70.46.148 2049 0 Y 7378 Self-heal Daemon on 10.70.46.148 N/A N/A Y 7386 [root@dhcp46-47 fuse]# cd .test [root@dhcp46-47 .test]# Bug verified on build glusterfs-3.7.9-1.el7rhgs.x86_64.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2016:1240