Bug 2166249
Summary: | Commands `pcs stonith sbd enable|disable` do not work properly when cluster is not running | ||
---|---|---|---|
Product: | Red Hat Enterprise Linux 9 | Reporter: | Miroslav Lisik <mlisik> |
Component: | pcs | Assignee: | Miroslav Lisik <mlisik> |
Status: | CLOSED ERRATA | QA Contact: | cluster-qe <cluster-qe> |
Severity: | high | Docs Contact: | |
Priority: | high | ||
Version: | 9.2 | CC: | cluster-maint, cluster-qe, idevat, mlisik, mmazoure, mpospisi, nhostako, omular, tojeline |
Target Milestone: | rc | Keywords: | Regression, Triaged |
Target Release: | 9.2 | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | pcs-0.11.4-5.el9 | Doc Type: | No Doc Update |
Doc Text: |
The affected packages have not been released.
|
Story Points: | --- |
Clone Of: | 2166243 | Environment: | |
Last Closed: | 2023-05-09 07:18:58 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | 2166243 | ||
Bug Blocks: |
Description
Miroslav Lisik
2023-02-01 09:49:37 UTC
Affected version of pcs in RHEL-.9.2.0 is pcs-0.11.4-1.el9. Upstream patch: https://github.com/ClusterLabs/pcs/commit/dd83d026340f21dc60f1bd9d572eaec8a5377bca Test: (pcs) [root@r92-1 pcs]# pcs/pcs host auth -u hacluster -p $PASSWORD r92-{1..2}.vm r92-2.vm: Authorized r92-1.vm: Authorized (pcs) [root@r92-1 pcs]# pcs/pcs cluster setup HACluster r92-{1..2}.vm <snip> >>> Enabling sbd before first cluster start (pcs) [root@r92-1 pcs]# pcs/pcs stonith sbd enable <snip> (pcs) [root@r92-1 pcs]# ls -la /var/lib/pacemaker/cib total 8 drwxr-x---. 2 hacluster haclient 4096 Feb 1 17:21 . drwxr-x---. 6 hacluster haclient 4096 Jan 30 10:48 .. (pcs) [root@r92-2 pcs]# ls -la /var/lib/pacemaker/cib total 8 drwxr-x---. 2 hacluster haclient 4096 Feb 1 17:21 . drwxr-x---. 6 hacluster haclient 4096 Jan 30 10:48 .. (pcs) [root@r92-1 pcs]# pcs/pcs cluster start --all --wait <snip> (pcs) [root@r92-1 pcs]# pcs/pcs status | grep -A1 "Node List" Node List: * Online: [ r92-1.vm r92-2.vm ] >>> Cluster started without issues. >>> Set stonith-watchdog-timeout and stop cluster. (pcs) [root@r92-1 pcs]# pcs/pcs property set stonith-watchdog-timeout=10 (pcs) [root@r92-1 pcs]# pcs/pcs cluster cib | grep stonith-watchdog-timeout <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> (pcs) [root@r92-1 pcs]# pcs/pcs cluster stop --all <snip> >>> Check cib file before and after enabling sbd. (pcs) [root@r92-1 pcs]# grep stonith-watchdog-timeout -R /var/lib/pacemaker/cib; ls -la /var/lib/pacemaker/cib/cib.xml* /var/lib/pacemaker/cib/cib-5.raw: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> /var/lib/pacemaker/cib/cib.xml: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> -rw-------. 1 hacluster haclient 1268 Feb 1 17:47 /var/lib/pacemaker/cib/cib.xml -rw-------. 1 hacluster haclient 32 Feb 1 17:47 /var/lib/pacemaker/cib/cib.xml.sig (pcs) [root@r92-2 pcs]# grep stonith-watchdog-timeout -R /var/lib/pacemaker/cib; ls -la /var/lib/pacemaker/cib/cib.xml* /var/lib/pacemaker/cib/cib.xml: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> -rw-------. 1 hacluster haclient 1268 Feb 1 17:46 /var/lib/pacemaker/cib/cib.xml -rw-------. 1 hacluster haclient 32 Feb 1 17:46 /var/lib/pacemaker/cib/cib.xml.sig (pcs) [root@r92-1 pcs]# pcs/pcs stonith sbd enable <snip> (pcs) [root@r92-1 pcs]# grep stonith-watchdog-timeout -R /var/lib/pacemaker/cib; ls -la /var/lib/pacemaker/cib/cib.xml* /var/lib/pacemaker/cib/cib-5.raw: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> -rw-------. 1 hacluster haclient 1171 Feb 1 17:48 /var/lib/pacemaker/cib/cib.xml (pcs) [root@r92-2 pcs]# grep stonith-watchdog-timeout -R /var/lib/pacemaker/cib; ls -la /var/lib/pacemaker/cib/cib.xml* -rw-------. 1 hacluster haclient 1171 Feb 1 17:48 /var/lib/pacemaker/cib/cib.xml >>> Cib files were updated on all nodes. >>> Start cluster and check stonith-watchdog-timeout property. (pcs) [root@r92-1 pcs]# pcs/pcs cluster start --all --wait <snip> (pcs) [root@r92-1 pcs]# pcs/pcs cluster cib | grep stonith-watchdog-timeout >>> Cluster property stonith-watchdog-timeout is not set as expected. >>> Set stonith-watchdog-timeout again and stop the cluster. (pcs) [root@r92-1 pcs]# pcs/pcs property set stonith-watchdog-timeout=10 (pcs) [root@r92-1 pcs]# pcs/pcs cluster cib | grep stonith-watchdog-timeout <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> (pcs) [root@r92-1 pcs]# pcs/pcs cluster stop --all <snip> >>> Check cib file before and after disabling sbd. (pcs) [root@r92-1 pcs]# grep stonith-watchdog-timeout -R /var/lib/pacemaker/cib; ls -la /var/lib/pacemaker/cib/cib.xml* /var/lib/pacemaker/cib/cib-5.raw: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> /var/lib/pacemaker/cib/cib.xml: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> -rw-------. 1 hacluster haclient 1268 Feb 1 17:55 /var/lib/pacemaker/cib/cib.xml -rw-------. 1 hacluster haclient 32 Feb 1 17:55 /var/lib/pacemaker/cib/cib.xml.sig (pcs) [root@r92-2 pcs]# grep stonith-watchdog-timeout -R /var/lib/pacemaker/cib; ls -la /var/lib/pacemaker/cib/cib.xml* /var/lib/pacemaker/cib/cib.xml: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> -rw-------. 1 hacluster haclient 1268 Feb 1 17:55 /var/lib/pacemaker/cib/cib.xml -rw-------. 1 hacluster haclient 32 Feb 1 17:55 /var/lib/pacemaker/cib/cib.xml.sig (pcs) [root@r92-1 pcs]# pcs/pcs stonith sbd disable <snip> (pcs) [root@r92-1 pcs]# grep stonith-watchdog-timeout -R /var/lib/pacemaker/cib; ls -la /var/lib/pacemaker/cib/cib.xml* /var/lib/pacemaker/cib/cib-5.raw: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="10"/> /var/lib/pacemaker/cib/cib.xml: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="0"/> -rw-------. 1 hacluster haclient 1279 Feb 1 17:58 /var/lib/pacemaker/cib/cib.xml (pcs) [root@r92-2 pcs]# grep stonith-watchdog-timeout -R /var/lib/pacemaker/cib; ls -la /var/lib/pacemaker/cib/cib.xml* /var/lib/pacemaker/cib/cib.xml: <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="0"/> -rw-------. 1 hacluster haclient 1279 Feb 1 17:58 /var/lib/pacemaker/cib/cib.xml >>> Cib files were updated on all nodes. >>> Start cluster and check stonith-watchdog-timeout property. (pcs) [root@r92-1 pcs]# pcs/pcs cluster start --all --wait <snip> (pcs) [root@r92-1 pcs]# pcs/pcs cluster cib | grep stonith-watchdog-timeout <nvpair id="cib-bootstrap-options-stonith-watchdog-tim" name="stonith-watchdog-timeout" value="0"/> (pcs) [root@r92-1 pcs]# pcs/pcs status | grep -A1 "Node List" Node List: * Online: [ r92-1.vm r92-2.vm ] >>> Cluster property stonith-watchdog-timeout is set to 0 as expected. DevTestResults: [root@r09-02-a ~]# rpm -q pcs pcs-0.11.4-5.el9.x86_64 [root@r09-02-a ~]# pcs cluster setup thefinalcountdown r09-02-a.vm r09-02-b.vm r09-02-c.vm No addresses specified for host 'r09-02-a.vm', using 'r09-02-a.vm' No addresses specified for host 'r09-02-b.vm', using 'r09-02-b.vm' No addresses specified for host 'r09-02-c.vm', using 'r09-02-c.vm' Destroying cluster on hosts: 'r09-02-a.vm', 'r09-02-b.vm', 'r09-02-c.vm'... r09-02-b.vm: Successfully destroyed cluster r09-02-a.vm: Successfully destroyed cluster r09-02-c.vm: Successfully destroyed cluster Requesting remove 'pcsd settings' from 'r09-02-a.vm', 'r09-02-b.vm', 'r09-02-c.vm' r09-02-b.vm: successful removal of the file 'pcsd settings' r09-02-a.vm: successful removal of the file 'pcsd settings' r09-02-c.vm: successful removal of the file 'pcsd settings' Sending 'corosync authkey', 'pacemaker authkey' to 'r09-02-a.vm', 'r09-02-b.vm', 'r09-02-c.vm' r09-02-a.vm: successful distribution of the file 'corosync authkey' r09-02-a.vm: successful distribution of the file 'pacemaker authkey' r09-02-b.vm: successful distribution of the file 'corosync authkey' r09-02-b.vm: successful distribution of the file 'pacemaker authkey' r09-02-c.vm: successful distribution of the file 'corosync authkey' r09-02-c.vm: successful distribution of the file 'pacemaker authkey' Sending 'corosync.conf' to 'r09-02-a.vm', 'r09-02-b.vm', 'r09-02-c.vm' r09-02-a.vm: successful distribution of the file 'corosync.conf' r09-02-b.vm: successful distribution of the file 'corosync.conf' r09-02-c.vm: successful distribution of the file 'corosync.conf' Cluster has been successfully set up. [root@r09-02-a ~]# pcs stonith sbd enable Running SBD pre-enabling checks... r09-02-a.vm: SBD pre-enabling checks done r09-02-b.vm: SBD pre-enabling checks done r09-02-c.vm: SBD pre-enabling checks done Distributing SBD config... r09-02-a.vm: SBD config saved r09-02-b.vm: SBD config saved r09-02-c.vm: SBD config saved Enabling sbd... r09-02-a.vm: sbd enabled r09-02-b.vm: sbd enabled r09-02-c.vm: sbd enabled Warning: Cluster restart is required in order to apply these changes. [root@r09-02-a ~]# pcs status Cluster name: thefinalcountdown Status of pacemakerd: 'Pacemaker is running' (last updated 2023-02-03 18:31:23 +01:00) Cluster Summary: * Stack: corosync * Current DC: r09-02-c.vm (version 2.1.5-4.el9-a3f44794f94) - partition with quorum * Last updated: Fri Feb 3 18:31:24 2023 * Last change: Fri Feb 3 18:30:32 2023 by hacluster via crmd on r09-02-c.vm * 3 nodes configured * 0 resource instances configured Node List: * Online: [ r09-02-a.vm r09-02-b.vm r09-02-c.vm ] Full List of Resources: * No resources Daemon Status: corosync: active/disabled pacemaker: active/disabled pcsd: active/enabled sbd: active/enabled [root@r09-02-a ~]# pcs stonith sbd status SBD STATUS <node name>: <installed> | <enabled> | <running> r09-02-b.vm: YES | YES | YES r09-02-c.vm: YES | YES | YES r09-02-a.vm: YES | YES | YES Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (pcs bug fix and enhancement update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2023:2151 |