Bug 1225423
Summary: | pcs should allow to remove a dead node from a cluster | ||||||||
---|---|---|---|---|---|---|---|---|---|
Product: | Red Hat Enterprise Linux 7 | Reporter: | Tomas Jelinek <tojeline> | ||||||
Component: | pcs | Assignee: | Tomas Jelinek <tojeline> | ||||||
Status: | CLOSED ERRATA | QA Contact: | cluster-qe <cluster-qe> | ||||||
Severity: | urgent | Docs Contact: | |||||||
Priority: | urgent | ||||||||
Version: | 7.2 | CC: | amaumene, cfeist, cluster-maint, idevat, j_t_williams, michele, mlisik, rsteiger, sankarshan, tojeline, vcojot | ||||||
Target Milestone: | rc | Keywords: | ZStream | ||||||
Target Release: | --- | ||||||||
Hardware: | Unspecified | ||||||||
OS: | Unspecified | ||||||||
Whiteboard: | |||||||||
Fixed In Version: | pcs-0.9.152-5.el7 | Doc Type: | Bug Fix | ||||||
Doc Text: |
Cause:
User wants to remove a powered off node from a cluster.
Consequence:
Pcs does not remove the node as it cannot connect to it and remove the cluster configuration files from it.
Fix:
Skip removing configuration files from the node if the user used --force flag.
Result:
It is possible to remove a powered off node from the cluster.
|
Story Points: | --- | ||||||
Clone Of: | |||||||||
: | 1382633 (view as bug list) | Environment: | |||||||
Last Closed: | 2016-11-03 20:54:10 UTC | Type: | Bug | ||||||
Regression: | --- | Mount Type: | --- | ||||||
Documentation: | --- | CRM: | |||||||
Verified Versions: | Category: | --- | |||||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||||
Cloudforms Team: | --- | Target Upstream Version: | |||||||
Embargoed: | |||||||||
Bug Depends On: | |||||||||
Bug Blocks: | 1305654, 1382633 | ||||||||
Attachments: |
|
Description
Tomas Jelinek
2015-05-27 10:41:36 UTC
Hi, I also hit the bug, thanks for the workaround. Regards, Created attachment 1181676 [details] proposed fix Test: > Let's have a three node cluster [root@rh72-node1:~]# pcs status nodes both Corosync Nodes: Online: rh72-node1 rh72-node2 rh72-node3 Offline: Pacemaker Nodes: Online: rh72-node1 rh72-node2 rh72-node3 Standby: Maintenance: Offline: Pacemaker Remote Nodes: Online: Standby: Maintenance: Offline: > Power off one node ... [root@rh72-node1:~]# pcs status nodes both Corosync Nodes: Online: rh72-node1 rh72-node2 Offline: rh72-node3 Pacemaker Nodes: Online: rh72-node1 rh72-node2 Standby: Maintenance: Offline: rh72-node3 Pacemaker Remote Nodes: Online: Standby: Maintenance: Offline: > ... and remove it from the cluster [root@rh72-node1:~]# pcs cluster node remove rh72-node3 Error: pcsd is not running on rh72-node3, use --force to override [root@rh72-node1:~]# pcs cluster node remove rh72-node3 --force rh72-node3: Unable to connect to rh72-node3 ([Errno 113] No route to host) rh72-node3: Unable to connect to rh72-node3 ([Errno 113] No route to host) Warning: unable to destroy cluster rh72-node3: Unable to connect to rh72-node3 ([Errno 113] No route to host) rh72-node2: Corosync updated rh72-node1: Corosync updated [root@rh72-node1:~]# pcs status nodes both Corosync Nodes: Online: rh72-node1 rh72-node2 Offline: Pacemaker Nodes: Online: rh72-node1 rh72-node2 Standby: Maintenance: Offline: Pacemaker Remote Nodes: Online: Standby: Maintenance: Offline: Created attachment 1181958 [details]
proposed fix web UI
fix for web UI
Setup:
[vm-rhel72-1 ~] $ pcs status nodes both
Corosync Nodes:
Online: vm-rhel72-1 vm-rhel72-2 vm-rhel72-3
Offline:
Pacemaker Nodes:
Online: vm-rhel72-1 vm-rhel72-2 vm-rhel72-3
Standby:
Maintenance:
Offline:
Pacemaker Remote Nodes:
Online:
Standby:
Maintenance:
Offline:
> Power off one node ...
[vm-rhel72-1 ~] $ pcs status nodes both
Corosync Nodes:
Online: vm-rhel72-1 vm-rhel72-3
Offline: vm-rhel72-2
Pacemaker Nodes:
Online: vm-rhel72-1 vm-rhel72-3
Standby:
Maintenance:
Offline: vm-rhel72-2
Pacemaker Remote Nodes:
Online:
Standby:
Maintenance:
Offline:
Before Fix:
[vm-rhel72-1 ~] $ rpm -q pcs
pcs-0.9.152-4.el7.x86_64
[vm-rhel72-1 ~] $ pcs cluster node remove vm-rhel72-2
Error: pcsd is not running on vm-rhel72-2
After Fix:
[vm-rhel72-1 ~] $ rpm -q pcs
pcs-0.9.152-5.el7.x86_64
[vm-rhel72-1 ~] $ pcs cluster node remove vm-rhel72-2
Error: pcsd is not running on vm-rhel72-2, use --force to override
[vm-rhel72-1 ~] $ pcs cluster node remove vm-rhel72-2 --force
vm-rhel72-2: Unable to connect to vm-rhel72-2 ([Errno 111] Connection refused)
vm-rhel72-2: Unable to connect to vm-rhel72-2 ([Errno 111] Connection refused)
Warning: unable to destroy cluster
vm-rhel72-2: Unable to connect to vm-rhel72-2 ([Errno 111] Connection refused)
vm-rhel72-1: Corosync updated
vm-rhel72-3: Corosync updated
[vm-rhel72-1 ~] $ pcs status nodes both
Corosync Nodes:
Online: vm-rhel72-1 vm-rhel72-3
Offline:
Pacemaker Nodes:
Online: vm-rhel72-1 vm-rhel72-3
Standby:
Maintenance:
Offline:
Pacemaker Remote Nodes:
Online:
Standby:
Maintenance:
Offline:
*** Bug 1376209 has been marked as a duplicate of this bug. *** Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHSA-2016-2596.html |