Hide Forgot
In a 2*2 distributed-replicated setup, replace brick fails with following output. [root@FC-4 Bricks]# gluster volume replace-brick hosdu 10.1.12.127:/data/vishwa/brick3/ 10.1.12.126:/data/vishwa/brick3 start replace-brick started successfully [root@FC-4 Bricks]# gluster volume replace-brick hosdu 10.1.12.127:/data/vishwa/brick3/ 10.1.12.126:/data/vishwa/brick3 status 10.1.12.126, is not connected at the moment Both the machines were connected before replace-brick and after replace-brick they get disconnected. I have attached the glusterd log.
Can you please confirm if this still happens?
yes...This issue is still there in 3.2.0beta1 code.
test comment
test comment2
Created attachment 490
Created attachment 491
With the same set-up, the issue still persists with DEBUG build. root@gluster-Ubuntu1:/etc/glusterd# /opt/320qa7/sbin/gluster volume info Volume Name: mattondu Type: Distributed-Replicate Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: 192.168.1.224:/data/export5 Brick2: 192.168.1.223:/data/export6 Brick3: 192.168.1.224:/data/export7 Brick4: 192.168.1.223:/data/export8 root@gluster-Ubuntu1:/etc/glusterd# /opt/320qa7/sbin/gluster volume replace-brick mattondu 192.168.1.223:/data/export8/ 192.168.1.224:/data/export9/ start replace-brick started successfully root@gluster-Ubuntu1:/etc/glusterd# /opt/320qa7/sbin/gluster volume replace-brick mattondu 192.168.1.223:/data/export8/192.168.1.224:/data/export9/ status Number of files migrated = 0 Current file= root@gluster-Ubuntu1:/etc/glusterd# /opt/320qa7/sbin/gluster volume replace-brick mattondu 192.168.1.223:/data/export8/ 192.168.1.224:/data/export9/ status 192.168.1.224, is not connected at the moment Here glusterd at 224 machine crashed. I couldn't attach the core file since it's too big(16M). I have attached the glusterd logs from both the machines.
*** This bug has been marked as a duplicate of bug 2489 ***