Bug 1225423
| Summary: | pcs should allow to remove a dead node from a cluster | ||||||||
|---|---|---|---|---|---|---|---|---|---|
| Product: | Red Hat Enterprise Linux 7 | Reporter: | Tomas Jelinek <tojeline> | ||||||
| Component: | pcs | Assignee: | Tomas Jelinek <tojeline> | ||||||
| Status: | CLOSED ERRATA | QA Contact: | cluster-qe <cluster-qe> | ||||||
| Severity: | urgent | Docs Contact: | |||||||
| Priority: | urgent | ||||||||
| Version: | 7.2 | CC: | amaumene, cfeist, cluster-maint, idevat, j_t_williams, michele, mlisik, rsteiger, sankarshan, tojeline, vcojot | ||||||
| Target Milestone: | rc | Keywords: | ZStream | ||||||
| Target Release: | --- | ||||||||
| Hardware: | Unspecified | ||||||||
| OS: | Unspecified | ||||||||
| Whiteboard: | |||||||||
| Fixed In Version: | pcs-0.9.152-5.el7 | Doc Type: | Bug Fix | ||||||
| Doc Text: |
Cause:
User wants to remove a powered off node from a cluster.
Consequence:
Pcs does not remove the node as it cannot connect to it and remove the cluster configuration files from it.
Fix:
Skip removing configuration files from the node if the user used --force flag.
Result:
It is possible to remove a powered off node from the cluster.
|
Story Points: | --- | ||||||
| Clone Of: | |||||||||
| : | 1382633 (view as bug list) | Environment: | |||||||
| Last Closed: | 2016-11-03 20:54:10 UTC | Type: | Bug | ||||||
| Regression: | --- | Mount Type: | --- | ||||||
| Documentation: | --- | CRM: | |||||||
| Verified Versions: | Category: | --- | |||||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||||
| Embargoed: | |||||||||
| Bug Depends On: | |||||||||
| Bug Blocks: | 1305654, 1382633 | ||||||||
| Attachments: |
|
||||||||
|
Description
Tomas Jelinek
2015-05-27 10:41:36 UTC
Hi, I also hit the bug, thanks for the workaround. Regards, Created attachment 1181676 [details] proposed fix Test: > Let's have a three node cluster [root@rh72-node1:~]# pcs status nodes both Corosync Nodes: Online: rh72-node1 rh72-node2 rh72-node3 Offline: Pacemaker Nodes: Online: rh72-node1 rh72-node2 rh72-node3 Standby: Maintenance: Offline: Pacemaker Remote Nodes: Online: Standby: Maintenance: Offline: > Power off one node ... [root@rh72-node1:~]# pcs status nodes both Corosync Nodes: Online: rh72-node1 rh72-node2 Offline: rh72-node3 Pacemaker Nodes: Online: rh72-node1 rh72-node2 Standby: Maintenance: Offline: rh72-node3 Pacemaker Remote Nodes: Online: Standby: Maintenance: Offline: > ... and remove it from the cluster [root@rh72-node1:~]# pcs cluster node remove rh72-node3 Error: pcsd is not running on rh72-node3, use --force to override [root@rh72-node1:~]# pcs cluster node remove rh72-node3 --force rh72-node3: Unable to connect to rh72-node3 ([Errno 113] No route to host) rh72-node3: Unable to connect to rh72-node3 ([Errno 113] No route to host) Warning: unable to destroy cluster rh72-node3: Unable to connect to rh72-node3 ([Errno 113] No route to host) rh72-node2: Corosync updated rh72-node1: Corosync updated [root@rh72-node1:~]# pcs status nodes both Corosync Nodes: Online: rh72-node1 rh72-node2 Offline: Pacemaker Nodes: Online: rh72-node1 rh72-node2 Standby: Maintenance: Offline: Pacemaker Remote Nodes: Online: Standby: Maintenance: Offline: Created attachment 1181958 [details]
proposed fix web UI
fix for web UI
Setup:
[vm-rhel72-1 ~] $ pcs status nodes both
Corosync Nodes:
Online: vm-rhel72-1 vm-rhel72-2 vm-rhel72-3
Offline:
Pacemaker Nodes:
Online: vm-rhel72-1 vm-rhel72-2 vm-rhel72-3
Standby:
Maintenance:
Offline:
Pacemaker Remote Nodes:
Online:
Standby:
Maintenance:
Offline:
> Power off one node ...
[vm-rhel72-1 ~] $ pcs status nodes both
Corosync Nodes:
Online: vm-rhel72-1 vm-rhel72-3
Offline: vm-rhel72-2
Pacemaker Nodes:
Online: vm-rhel72-1 vm-rhel72-3
Standby:
Maintenance:
Offline: vm-rhel72-2
Pacemaker Remote Nodes:
Online:
Standby:
Maintenance:
Offline:
Before Fix:
[vm-rhel72-1 ~] $ rpm -q pcs
pcs-0.9.152-4.el7.x86_64
[vm-rhel72-1 ~] $ pcs cluster node remove vm-rhel72-2
Error: pcsd is not running on vm-rhel72-2
After Fix:
[vm-rhel72-1 ~] $ rpm -q pcs
pcs-0.9.152-5.el7.x86_64
[vm-rhel72-1 ~] $ pcs cluster node remove vm-rhel72-2
Error: pcsd is not running on vm-rhel72-2, use --force to override
[vm-rhel72-1 ~] $ pcs cluster node remove vm-rhel72-2 --force
vm-rhel72-2: Unable to connect to vm-rhel72-2 ([Errno 111] Connection refused)
vm-rhel72-2: Unable to connect to vm-rhel72-2 ([Errno 111] Connection refused)
Warning: unable to destroy cluster
vm-rhel72-2: Unable to connect to vm-rhel72-2 ([Errno 111] Connection refused)
vm-rhel72-1: Corosync updated
vm-rhel72-3: Corosync updated
[vm-rhel72-1 ~] $ pcs status nodes both
Corosync Nodes:
Online: vm-rhel72-1 vm-rhel72-3
Offline:
Pacemaker Nodes:
Online: vm-rhel72-1 vm-rhel72-3
Standby:
Maintenance:
Offline:
Pacemaker Remote Nodes:
Online:
Standby:
Maintenance:
Offline:
*** Bug 1376209 has been marked as a duplicate of this bug. *** Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHSA-2016-2596.html |