Bug 1560877

Summary: Bricks offline
Product: [Community] GlusterFS Reporter: david <davidpsv17>
Component: upcallAssignee: bugs <bugs>
Status: CLOSED EOL QA Contact:
Severity: unspecified Docs Contact:
Priority: unspecified    
Version: 3.10CC: amukherj, bugs
Target Milestone: ---   
Target Release: ---   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Description david 2018-03-27 07:58:52 UTC
Hi!

I have a Distributed-Replica volume with glusterfs, it has 32 bricks distributed in 16 machines (2 bricks per machine).  Almost every day, random bricks get offline and I have to go to the machine that has the brick offline and restart the glusterfs-server. After restart the service, the brick affected go back online.

At the moment that one brick gets offline I see in the affected machines these logs:

glustershd.log:
W [socket.c:590:__socket_rwv] 0-volume1-client-27: readv on 10.0.10.58:49155 failed (No data available)
I [MSGID: 114018] [client.c:2280:client_rpc_notify] 0-volume1-client-27: disconnected from volume1-client-27. Client process will keep trying to connect to glusterd until brick's port is available
I [rpc-clnt.c:2001:rpc_clnt_reconfig] 0-volume1-client-27: changing port to 49155 (from 0)
E [socket.c:2309:socket_connect_finish] 0-volume1-client-27: connection to 10.0.10.58:49155 failed (Connection refused)


storage-pool.log -> same errors as glusterhd.log

etc-glusterfs-glusterd.vol.log:
W [socket.c:590:__socket_rwv] 0-management: readv on /var/run/gluster/17f9df49125385b6fcafaf23966f47d3.socket failed (No data available)
I [MSGID: 106005] [glusterd-handler.c:5055:__glusterd_brick_rpc_notify] 0-management: Brick ubuntu15:/work/work2/gluster-storage has disconnected from glusterd.


work-work2-gluster-storage.log:

pending frames:
patchset: git://git.gluster.com/glusterfs.git
signal received: 11
time of crash: 
2018-03-27 05:04:23
configuration details:
argp 1
backtrace 1
dlfcn 1
libpthread 1
llistxattr 1
setfsid 1
spinlock 1
epoll.h 1
xattr.h 1
st_atim.tv_nsec 1
package-string: glusterfs 3.8.15
/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_msg_backtrace_nomem+0x7e)[0x7f6af107705e]
/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(gf_print_trace+0x324)[0x7f6af1080784]
/lib/x86_64-linux-gnu/libc.so.6(+0x354b0)[0x7f6af046c4b0]
/usr/lib/x86_64-linux-gnu/glusterfs/3.8.15/xlator/features/upcall.so(+0xe9da)[0x7f6ae8fab9da]
/usr/lib/x86_64-linux-gnu/glusterfs/3.8.15/xlator/features/upcall.so(+0xeaac)[0x7f6ae8fabaac]
/usr/lib/x86_64-linux-gnu/glusterfs/3.8.15/xlator/features/upcall.so(+0x1759)[0x7f6ae8f9e759]
/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(+0x2f349)[0x7f6af1084349]
/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(+0x3042a)[0x7f6af108542a]
/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(inode_unref+0x34)[0x7f6af10856e4]
/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(loc_wipe+0x12)[0x7f6af1074eb2]
/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(args_wipe+0x12)[0x7f6af10e1ed2]
/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(call_stub_destroy+0x39)[0x7f6af1097a19]
/usr/lib/x86_64-linux-gnu/glusterfs/3.8.15/xlator/performance/io-threads.so(+0x44e7)[0x7f6ae8d964e7]
/lib/x86_64-linux-gnu/libpthread.so.0(+0x76ba)[0x7f6af08086ba]
/lib/x86_64-linux-gnu/libc.so.6(clone+0x6d)[0x7f6af053e41d]
---------


Version-Release number of selected component (if applicable):
glusterfs 3.8.15

Comment 1 Shyamsundar 2018-06-20 18:28:13 UTC
This bug reported is against a version of Gluster that is no longer maintained (or has been EOL'd). See https://www.gluster.org/release-schedule/ for the versions currently maintained.

As a result this bug is being closed.

If the bug persists on a maintained version of gluster or against the mainline gluster repository, request that it be reopened and the Version field be marked appropriately.

Comment 2 Shyamsundar 2018-06-20 18:28:15 UTC
This bug reported is against a version of Gluster that is no longer maintained
(or has been EOL'd). See https://www.gluster.org/release-schedule/ for the
versions currently maintained.

As a result this bug is being closed.

If the bug persists on a maintained version of gluster or against the mainline
gluster repository, request that it be reopened and the Version field be marked
appropriately.