Bug 1641944
| Summary: | [UPGRADES][10]RMQ resource-agent should handle stopped node [rhel-7.6.z] | ||
|---|---|---|---|
| Product: | Red Hat Enterprise Linux 7 | Reporter: | Oneata Mircea Teodor <toneata> |
| Component: | resource-agents | Assignee: | Oyvind Albrigtsen <oalbrigt> |
| Status: | CLOSED ERRATA | QA Contact: | pkomarov |
| Severity: | urgent | Docs Contact: | |
| Priority: | urgent | ||
| Version: | 7.5 | CC: | agk, aherr, apevec, augol, ccamacho, cchen, cfeist, cluster-maint, ctowsley, fdinitto, jeckersb, lhh, lmanasko, lmiccini, michele, mkrcmari, morazi, oalbrigt, pkomarov, rscarazz, sbradley, sgolovat, srevivo, toneata, yprokule |
| Target Milestone: | rc | Keywords: | ReleaseNotes, Triaged, ZStream |
| Target Release: | --- | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | resource-agents-4.1.1-12.el7_6.6 | Doc Type: | If docs needed, set a value |
| Doc Text: |
Previously, the rabbitmqctl cluster_status command read cached cluster status from disk and returned 0 when the mnesia service was not running. For example, this happened if rabbitmqctl stop_app was called, or the service paused during partition due to the pause_minority strategy. As a consequence, RabbitMQ sometimes returned cached status from disk. With this update, RabbitMQ now gets cluster status from mnesia during the monitor action. As a result, the described problem no longer occurs.
|
Story Points: | --- |
| Clone Of: | 1595753 | Environment: | |
| Last Closed: | 2018-11-27 01:19:45 UTC | Type: | --- |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | 1595753 | ||
| Bug Blocks: | |||
|
Description
Oneata Mircea Teodor
2018-10-23 08:29:36 UTC
Note that issuing 'rabbitmqctl shutdown' - did cause the correct restart of rabbitmq resource - the difference is that the earlang node was shutdown along side the rabbitmq app. https://www.rabbitmq.com/rabbitmqctl.8.html#shutdown Veriried,
#tested with resource-agents-4.1.1-12.el7_6.6, bumped the version in "Fixed in..."
[root@controller-1 ~]# rabbitmqctl stop_app
Stopping node 'rabbit@controller-1' ...
#pacemaker sees rabbitmq failing on controller-1
[root@controller-0 ~]#
Clone Set: rabbitmq-clone [rabbitmq]
rabbitmq (ocf::heartbeat:rabbitmq-cluster): FAILED controller-1
Started: [ controller-0 controller-2 ]
Master/Slave Set: redis-master [redis]
#pacemaker restarts rabbitmq on controller-1 succesfully:
rabbitmq (ocf::heartbeat:rabbitmq-cluster): Started controller-1 (Monitoring)
...
ip-10.0.0.106 (ocf::heartbeat:IPaddr2): Started controller-1
Clone Set: rabbitmq-clone [rabbitmq]
Started: [ controller-0 controller-1 controller-2 ]
#mnesia is reporting all rabbitmq apps are up:
[root@controller-0 ~]# rabbitmqctl eval "rabbit_mnesia:cluster_status_from_mnesia()."
{ok,{['rabbit@controller-1','rabbit@controller-2','rabbit@controller-0'],
['rabbit@controller-0','rabbit@controller-1','rabbit@controller-2'],
['rabbit@controller-1','rabbit@controller-2','rabbit@controller-0']}}
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2018:3661 |