Description of problem: - `rabbitmq cluster_status` shows nodedown alerts - list_queues / list_connections hang - `rabbitmqctl node_health_check` fails with an error. * There is no any issue while performing activity on RHOS setup(From horizon/cli). i.e. RHOS environment is functioning as expected. <snip> sudo rabbitmqctl node_health_check -n rabbit@node1 Checking health of node 'rabbit@node1' ... Heath check failed: health check of node 'rabbit@node1' fails: nodedown </snip> - api network is on ipv6. Version-Release number of selected component (if applicable): RHOS9 rabbitmq-server-3.6.3-5.el7ost.noarch How reproducible: 100% Steps to Reproduce: 1. 2. 3. Actual results: rabbitmq cluster_status shows nodedown alerts, list_queues / list_connections hang Expected results: there should not be any issue with rabbitmqctl commands. Additional info:
Pratik, Are the nodedown alerts are shown after deployment finished or after some time(or operations/workload)?
Verified using openstack-tripleo-heat-templates-liberty-2.0.0-41.el7ost.noarch Tested the status of rabbitmq using following scenarios: -after deployment finished. -after cluster operations -after/during opnestack operations using Rally( create and delete instances x20times) Results=all passed [root@overcloud-controller-1 ~]# rabbitmqctl cluster_status Cluster status of node 'rabbit@overcloud-controller-1' ... [{nodes,[{disc,['rabbit@overcloud-controller-0', 'rabbit@overcloud-controller-1', 'rabbit@overcloud-controller-2']}]}, {running_nodes,['rabbit@overcloud-controller-2', 'rabbit@overcloud-controller-0', 'rabbit@overcloud-controller-1']}, {cluster_name,<<"rabbit">>}, {partitions,[]}, {alarms,[{'rabbit@overcloud-controller-2',[]}, {'rabbit@overcloud-controller-0',[]}, {'rabbit@overcloud-controller-1',[]}]}] [root@overcloud-controller-1 ~]# rabbitmqctl node_health_check Checking health of node 'rabbit@overcloud-controller-1' ... Health check passed [root@overcloud-controller-1 ~]#
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHBA-2016-2983.html