Bug 492948
| Summary: | Ghost cluster member after restarting openais. | ||
|---|---|---|---|
| Product: | Red Hat Enterprise MRG | Reporter: | Alan Conway <aconway> |
| Component: | qpid-cpp | Assignee: | Alan Conway <aconway> |
| Status: | CLOSED DUPLICATE | QA Contact: | MRG Quality Engineering <mrgqe-bugs> |
| Severity: | medium | Docs Contact: | |
| Priority: | medium | ||
| Version: | 1.1 | ||
| Target Milestone: | 1.3 | ||
| Target Release: | --- | ||
| Hardware: | All | ||
| OS: | Linux | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | Bug Fix | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2009-03-31 21:34:37 UTC | Type: | --- |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
Caused by 489451, which is fixed in openais-0.80.3-22.el5_3.4. I confirmed that qpidd behaves as expected with the updated openais. *** This bug has been marked as a duplicate of bug 489451 *** |
Description of problem: From Nick.Hall Ran into one issue on my first test - steps to reproduce: 1) start 4 node OpenAIS cluster 2) start qpidd on all 4 nodes 3) qpid-cluster shows 4 nodes 4) pick one node and fail it by running service openais stop 5) qpid-cluster still shows 4 nodes 6) restart OpenAIS and qpidd on failed node 7) qpid-cluster now shows 5 nodes with one repeated: [root@lnaiqlv24 ~]# qpid-cluster Cluster Name: TESTCLUSTER Cluster Status: ACTIVE Cluster Size: 5 Members: ID=10.229.20.196:3936 URL=amqp:tcp:10.229.20.196:5672 : ID=10.229.20.198:3661 URL=amqp:tcp:10.229.20.198:5672 : ID=10.229.20.199:4649 URL=amqp:tcp:10.229.20.199:5672 : ID=10.229.20.203:4104 URL=amqp:tcp:10.229.20.203:5672 : ID=10.229.20.203:4256 URL=amqp:tcp:10.229.20.203:5672 OpenAIS log correctly shows 4 nodes: Mar 30 16:31:21.912122 [CLM ] CLM CONFIGURATION CHANGE Mar 30 16:31:21.912135 [CLM ] New Configuration: Mar 30 16:31:21.912145 [CLM ] r(0) ip(10.229.20.196) Mar 30 16:31:21.912153 [CLM ] r(0) ip(10.229.20.198) Mar 30 16:31:21.912160 [CLM ] r(0) ip(10.229.20.199) Mar 30 16:31:21.912166 [CLM ] Members Left: Mar 30 16:31:21.912172 [CLM ] Members Joined: Mar 30 16:31:21.912188 [CLM ] CLM CONFIGURATION CHANGE Mar 30 16:31:21.912194 [CLM ] New Configuration: Mar 30 16:31:21.912202 [CLM ] r(0) ip(10.229.20.196) Mar 30 16:31:21.912209 [CLM ] r(0) ip(10.229.20.198) Mar 30 16:31:21.912217 [CLM ] r(0) ip(10.229.20.199) Mar 30 16:31:21.912224 [CLM ] r(0) ip(10.229.20.203) Mar 30 16:31:21.912230 [CLM ] Members Left: Mar 30 16:31:21.912237 [CLM ] Members Joined: Mar 30 16:31:21.912244 [CLM ] r(0) ip(10.229.20.203) Mar 30 16:31:21.912253 [SYNC ] This node is within the primary component and will provide service. Mar 30 16:31:21.912276 [TOTEM] entering OPERATIONAL state. Mar 30 16:31:21.913668 [CLM ] got nodejoin message 10.229.20.196 Mar 30 16:31:21.913775 [CLM ] got nodejoin message 10.229.20.198 Mar 30 16:31:21.913838 [CLM ] got nodejoin message 10.229.20.199 Mar 30 16:31:21.914079 [CLM ] got nodejoin message 10.229.20.203 Mar 30 16:31:21.921216 [CPG ] got joinlist message from node -954931958 Mar 30 16:31:21.921526 [CPG ] got joinlist message from node -1005263606 Mar 30 16:31:21.921628 [CPG ] got joinlist message from node -971709174