Hide Forgot
glusterfsd.log (similar on both machines) ============== pending frames: frame : type(1) op(LK) patchset: v2.0.7 signal received: 11 time of crash: 2009-10-23 13:11:07 configuration details: argp 1 backtrace 1 db.h 1 dlfcn 1 fdatasync 1 libpthread 1 llistxattr 1 setfsid 1 spinlock 1 epoll.h 1 xattr.h 1 st_atim.tv_nsec 1 package-string: glusterfs 2.0.7 /lib64/libc.so.6[0x3350830280] /usr/lib64/glusterfs/2.0.7/xlator/features/locks.so(__delete_lock+0x7)[0x2b5c2bd2fd07] /usr/lib64/glusterfs/2.0.7/xlator/features/locks.so[0x2b5c2bd2ff69] /usr/lib64/glusterfs/2.0.7/xlator/features/locks.so(pl_setlk+0x79)[0x2b5c2bd307c9] /usr/lib64/glusterfs/2.0.7/xlator/features/locks.so(pl_lk+0x15c)[0x2b5c2bd30dec] /usr/lib64/glusterfs/2.0.7/xlator/protocol/server.so(server_lk+0x1eb)[0x2b5c2bf43b9b] /usr/lib64/glusterfs/2.0.7/xlator/protocol/server.so(protocol_server_pollin+0x90)[0x2b5c2bf3ccd0] /usr/lib64/glusterfs/2.0.7/xlator/protocol/server.so(notify+0xcb)[0x2b5c2bf3cdab] /usr/lib64/glusterfs/2.0.7/transport/socket.so(socket_event_handler+0xd3)[0x2aaaaaaafdf3] /usr/lib64/libglusterfs.so.0[0x2b5c2b093755] /usr/sbin/glusterfsd(main+0x9e8)[0x403fa8] /lib64/libc.so.6(__libc_start_main+0xf4)[0x335081d974] /usr/sbin/glusterfsd[0x4025c9] --------- glusterfsd.vol (server) ============== volume posix type storage/posix option directory /mnt/sdb1 end-volume volume brick type features/locks subvolumes posix end-volume volume server type protocol/server option transport-type tcp option auth.addr.brick.allow * subvolumes brick end-volume glusterfs.vol (client) ============= volume client1 type protocol/client option transport-type tcp option remote-host 192.168.25.31 option remote-subvolume brick end-volume volume client2 type protocol/client option transport-type tcp option remote-host 192.168.25.32 option remote-subvolume brick end-volume volume replicate type cluster/replicate subvolumes client1 client2 end-volume
Mark, can you send us the log files and the backtrace of this crash?
Created attachment 103 [details] patch replacing the fpregset patch, fixing the described problems
Created attachment 104 [details] Slovenian XKB - experimental
(In reply to comment #1) > Mark, can you send us the log files and the backtrace of this crash? how do I get the backtrace?
The backtrace can be got from the core file using gdb: gdb -c <core-file> <glusterfs binary> and then a 'bt' command on the gdb prompt. I tried reproducing this issue and was unable to crash glusterfs when it's mount point is re-exported as a Samba share. I saw your client logs and there are plenty of messages indicating a possible 'spilt brain' (files from the backend directories are modified, but not from the mount point). Was the backend accessed and files modified to result in a split brain?
(In reply to comment #5) > The backtrace can be got from the core file using gdb: > gdb -c <core-file> <glusterfs binary> and then a 'bt' command on the gdb > prompt. > I tried reproducing this issue and was unable to crash glusterfs when it's > mount point is re-exported as a Samba share. I saw your client logs and there > are plenty of messages indicating a possible 'spilt brain' (files from the > backend directories are modified, but not from the mount point). Was the > backend accessed and files modified to result in a split brain? Node1 Back Trace: (gdb) bt #0 0x00002b5c2bd2fd07 in ?? () #1 0x00002b5c2bd2ff69 in ?? () #2 0x0000000000000098 in ?? () #3 0x000000104ae19d5a in ?? () #4 0x0000000011696fb0 in ?? () #5 0x000000001165b550 in ?? () #6 0x0000000000000000 in ?? () Node2 Back Trace: (gdb) bt #0 0x00002b9f7e775d07 in ?? () #1 0x00002b9f7e775f69 in ?? () #2 0x0000000000000098 in ?? () #3 0x000000104ae15adc in ?? () #4 0x0000000012f43e20 in ?? () #5 0x0000000012fc0900 in ?? () #6 0x0000000000000000 in ?? ()
The logs show a slew of 'split brain' messages. Were there any files modified directly from the backend? I am unable to reproduce this crash, infact, glusterfs mounts being re-exported over samba has been well tested. Can we have remote access to the core if possible, since the backtrace provided is not much use.
Closing this bug due to lack of sufficient data. Please re-open this bug if it surfaces again.