Description of problem: **************************************** In a 1X2 CTDB setup, if both the nodes are rebooted it doesn't mount the /gluster/lock on one of the node once that node is up but ctdb status shows ok for that node. On the other node where /gluster/lock is mounted the ctdb status shows unhealthy. once the node is rebooted and /gluster/lock is mounted the ctdb status should be consistent. As per BZ https://bugzilla.redhat.com/show_bug.cgi?id=1164222 , if both nodes are down then Version-Release number of selected component (if applicable): ************************************************************** glusterfs-3.6.0.40-1.el6rhs.x86_64 samba-glusterfs-3.6.509-169.4.el6rhs.x86_64 How reproducible: ************************************************************* Always Steps to Reproduce: 1.Create 1x2 volume.Do ctdb setup 2.Reboot one node in the cluster, check ctdb status, check mount point 3.Reboot the other node, check ctdb status, check mount point Actual results: ************************************************************** Once the nodes are up, on the node where gluster mount didn't happen shows ctdb status as OK, and the other node where the mount happened shows ctdb status as UNHEALTHY. Expected results: **************************************************************** After coming up, the /gluster/lock mount should happen on both the nodes and ctdb status should show the correct status. Additional info:
After doing the setup with configuring public and private interfaces separately, where the gluster volume is present on public IP and ctdb nodes on private ip's, rebooting one of the nodes doesn't cause ctdb node to stay in unhealthy state once the gluster lock is mounted. Also as I learned if the gluster lock is not mounted then ctdb will create and access it's own lock and will show status as OK. As per discussion with glusterd team , on a two node gluster cluster if one node goes down then the gluster services may not come up once the node comes up (because of 50% node down in a cluster)and so lock doesn't get mounted. Closing this BZ as works for me. If I see the similar issue in 4 node setup then we open the BZ again.
As mentioned in above comment, ctdb creates its own lock and and ctdb status shows ok even if gluster lock is not mounted , that happens because the lockfile is created on the root. We may want to have lock-file being created in the sub-directory of mount point and not on the root. Reopening the BZ to get this fix.
*** Bug 1202328 has been marked as a duplicate of this bug. ***
Updated the doc text. Kindly review it for technical accuracy.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2019:0261