Description of problem: [root@rhs-client19 ~]# gluster volume info Volume Name: quo Type: Distributed-Replicate Volume ID: 96852dd0-e8f6-48f8-94e2-ef80e8c70778 Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: rhs-client19.lab.eng.blr.redhat.com:/home/A Brick2: rhs-client20.lab.eng.blr.redhat.com:/home/B Brick3: rhs-client21.lab.eng.blr.redhat.com:/home/C Brick4: rhs-client23.lab.eng.blr.redhat.com:/home/D Options Reconfigured: cluster.server-quorum-type: server cluster.server-quorum-ratio: 100% [root@rhs-client19 ~]# [root@rhs-client19 ~]# gluster volume status Status of volume: quo Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick rhs-client19.lab.eng.blr.redhat.com:/home/A 24010 Y 7374 Brick rhs-client20.lab.eng.blr.redhat.com:/home/B 24011 Y 6834 NFS Server on localhost 38467 Y 7380 Self-heal Daemon on localhost N/A Y 7385 NFS Server on rhs-client22.lab.eng.blr.redhat.com 38467 Y 18244 Self-heal Daemon on rhs-client22.lab.eng.blr.redhat.com N/A Y 18249 NFS Server on rhs-client20.lab.eng.blr.redhat.com 38467 Y 6840 Self-heal Daemon on rhs-client20.lab.eng.blr.redhat.com N/A Y 6845 [root@rhs-client19 ~]# gluster peer status Number of Peers: 3 Hostname: rhs-client22.lab.eng.blr.redhat.com Uuid: 8c743ecc-d9aa-4cb0-a7a7-3c45c5e1284d State: Peer in Cluster (Connected) Hostname: rhs-client20.lab.eng.blr.redhat.com Uuid: b7f33530-25c1-406c-8c76-2c5feabaf7b0 State: Peer in Cluster (Connected) Hostname: rhs-client21.lab.eng.blr.redhat.com Uuid: 5b315725-90dd-41f9-abe8-827d27db8210 State: Peer in Cluster (Disconnected) [root@rhs-client19 ~]#
And when the server comes up the brick does not come up: [root@rhs-client21 ~]# gluster volume status Status of volume: quo Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick rhs-client19.lab.eng.blr.redhat.com:/home/A 24010 Y 7374 Brick rhs-client20.lab.eng.blr.redhat.com:/home/B 24011 Y 6834 Brick rhs-client21.lab.eng.blr.redhat.com:/home/C 24010 N 4625 NFS Server on localhost 38467 Y 3165 Self-heal Daemon on localhost N/A Y 3171 NFS Server on rhs-client22.lab.eng.blr.redhat.com 38467 Y 18244 Self-heal Daemon on rhs-client22.lab.eng.blr.redhat.com N/A Y 18249 NFS Server on rhs-client20.lab.eng.blr.redhat.com 38467 Y 6840 Self-heal Daemon on rhs-client20.lab.eng.blr.redhat.com N/A Y 6845 NFS Server on 10.70.36.43 38467 Y 7380 Self-heal Daemon on 10.70.36.43 N/A Y 7385 [root@rhs-client21 ~]# gluster volume info Volume Name: quo Type: Distributed-Replicate Volume ID: 96852dd0-e8f6-48f8-94e2-ef80e8c70778 Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: rhs-client19.lab.eng.blr.redhat.com:/home/A Brick2: rhs-client20.lab.eng.blr.redhat.com:/home/B Brick3: rhs-client21.lab.eng.blr.redhat.com:/home/C Brick4: rhs-client23.lab.eng.blr.redhat.com:/home/D Options Reconfigured: cluster.server-quorum-type: server cluster.server-quorum-ratio: 100% [root@rhs-client21 ~]# [root@rhs-client21 ~]# gluster peer status Number of Peers: 4 Hostname: rhs-client23.lab.eng.blr.redhat.com Uuid: 230ae9f2-310e-49a6-b9f6-440bb5962da3 State: Peer Rejected (Connected) Hostname: rhs-client22.lab.eng.blr.redhat.com Uuid: 8c743ecc-d9aa-4cb0-a7a7-3c45c5e1284d State: Peer in Cluster (Connected) Hostname: 10.70.36.43 Uuid: 772396e0-ccae-4b64-99f9-84f7e836d101 State: Peer in Cluster (Connected) Hostname: rhs-client20.lab.eng.blr.redhat.com Uuid: b7f33530-25c1-406c-8c76-2c5feabaf7b0 State: Peer in Cluster (Connected) [root@rhs-client21 ~]#
seems like the release 3.3.0.3-32rhs should work fine.
Pranith, can you please help Kaushal on debugging these issues? (if it is still relevant)
Already sent the fix for this.
Tested with latest update, the server now is up when the machine is brought back up.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHBA-2013-1262.html