I ran into this when adding and deleting nodes from a cluster in conga:
I had a 3 node cluster and deleted one of the nodes (successfully), then tried
to add it back. The cluster I'm trying to add it to is named duck. Here's the
skeletal cluster.conf file in /etc/cluster/ on louey (the node I'm adding):
<cluster config_version="1" name="duck">
When i run 'service cman start', ccsd starts, but doesn't pick up the current
configuration from the one of the other two cluster nodes, causing cman to error
out because it can't find the local node name in the cluster.conf.
Here's the current configuration from the other two nodes:
<?xml version="1.0" ?>
<cluster alias="duck" config_version="8" name="duck">
<fence_daemon clean_start="0" post_fail_delay="0" post_join_delay="3"/>
<clusternode name="huey.lab.boston.redhat.com" nodeid="2" votes="1">
<device name="donald" port="1"/>
<device name="donald" port="5"/>
<clusternode name="dewey.lab.boston.redhat.com" nodeid="3"
<device name="donald" port="2"/>
<clusternode name="louey.lab.boston.redhat.com" nodeid="4"
ipaddr="donald.lab.boston.redhat.com" login="X" name="donald" passwd="X"/>
I restarted the two nodes that remained in the cluster to rule out any possible
problems with lingering cman two_node="1" state.
[root@huey ~]# cman_tool status
Config Version: 8
Cluster Name: duck
Cluster Id: 1573
Cluster Member: Yes
Cluster Generation: 188
Membership state: Cluster-Member
Expected votes: 3
Total votes: 2
Active subsystems: 6
Ports Bound: 0
Node name: huey.lab.boston.redhat.com
Node ID: 2
Multicast addresses: 22.214.171.124
Node addresses: 192.168.77.141
[root@huey ~]# cman_tool nodes
Node Sts Inc Joined Name
2 M 184 2007-07-26 12:38:39 huey.lab.boston.redhat.com
3 M 188 2007-07-26 12:38:39 dewey.lab.boston.redhat.com
4 X 0 louey.lab.boston.redhat.com
Forgot to mention above, I'm running the 20070725.0 RHEL5.1-Server tree with
cman package cman-2.0.70-1.el5
Did the cluster.conf file on "louey" get updated to match the config file on the
other two nodes? I'm just wondering if the file was correctly updated on the new
Nope, the new node never received the updated file, and still has the
version="1" cluster.conf file.
I'm not sure that this was ever intended to work. I think the best way to add a
node is to add the node to copy the cluster.conf to the new node and start from
there. That said, it seems like this should be doable.
I've added Dave to the CC list for this bug. He had some comments about this
method of adding a node.
Note that this has been reported to work when running a 3 node cluster and
adding a fourth node. Potential that this bug may only exist when going from a 2
node cluster to a 3 node cluster, since the 2 node cluster is a special case.
Is this still an issue for Conga? I'm assuming that Conga actually distributes
the full config file to all nodes rather than create a skeleton config file and
them depend on ccsd to update. That was the solution that was proposed.
If this is stil causing problems, let met know. Otherwise I'll close it.
Since there is a workaround, closing this WONTFIX. Also, ccsd won't be around too much longer, and ricci has ability to distibute cluster.conf file across cluster. All of these are better options.
this is still causing problems and breaks cluster administration as described in the manual with luci. I'm using the current packages provided by RHN.