Bug 2176490
Summary: | Command 'pcs config checkpoint diff' does not show configuration differences between checkpoints | |||
---|---|---|---|---|
Product: | Red Hat Enterprise Linux 8 | Reporter: | Tomas Jelinek <tojeline> | |
Component: | pcs | Assignee: | Miroslav Lisik <mlisik> | |
Status: | CLOSED ERRATA | QA Contact: | cluster-qe <cluster-qe> | |
Severity: | high | Docs Contact: | Steven J. Levine <slevine> | |
Priority: | high | |||
Version: | 8.2 | CC: | cfeist, cluster-maint, idevat, mlisik, mmazoure, mpospisi, nhostako, omular, slevine, tojeline | |
Target Milestone: | rc | Keywords: | Regression, Triaged, ZStream | |
Target Release: | 8.9 | |||
Hardware: | Unspecified | |||
OS: | Unspecified | |||
Whiteboard: | ||||
Fixed In Version: | pcs-0.10.16-1.el8 | Doc Type: | Bug Fix | |
Doc Text: |
.The `pcs config checkpoint diff` command now works correctly for all configuration sections
When it was introduced in the the RHEL 8.1 release, the `pcs config checkpoint diff` command did not show the differences for the Fencing Levels, Ordering Constraints, Colocation Constraints, and Ticket Constraints configuration sections. As of the RHEL 8.4 release, the `pcs config checkpoint diff` command had stopped showing the differences for the the Resources Defaults and Operations Defaults configuration sections. As of the RHEL 8.7 release, the `pcs config checkpoint diff` command had stopped showing the differences for the Resources, and Stonith devices configuration sections. This is because as the code responsible for displaying each of the different configuration sections switched to a new mechanism for loading CIB files, the loaded content was cached. The second file used for the difference comparison was not loaded and the cached content of the first file was used instead. As a result, the `diff` command yielded no output. With this fix, the CIB file content is no longer cached and the `pcs config checkpoint diff` command shows differences for all configuration sections.
|
Story Points: | --- | |
Clone Of: | 2175881 | |||
: | 2180700 2180701 2180702 2180703 (view as bug list) | Environment: | ||
Last Closed: | 2023-11-14 15:22:35 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | ||||
Bug Blocks: | 2180700, 2180701, 2180702, 2180703 |
Description
Tomas Jelinek
2023-03-08 14:13:15 UTC
Upstream commit: https://github.com/ClusterLabs/pcs/commit/2e225d1e46f962fddb1d6e6bd1d13a2348dc5c41 Updated commands: * pcs config checkpoint diff Test: Prepare configuration for `pcs config`: export NODELIST=(r8-node-01 r8-node-02 r8-node-03) pcs host auth -u hacluster -p $PASSWORD ${NODELIST[*]} pcs cluster setup HACluster ${NODELIST[*]} --start --wait for node in ${NODELIST[*]}; do pcs stonith create fence-1-$node fence_xvm; done for node in ${NODELIST[*]}; do pcs stonith create fence-2-$node fence_xvm; done for node in ${NODELIST[*]}; do pcs stonith level add 1 $node fence-1-$node; pcs stonith level add 2 $node fence-2-$node; done pcs resource create p-1 ocf:pacemaker:Dummy --no-default-ops pcs resource create p-2 ocf:pacemaker:Dummy --no-default-ops pcs constraint location p-1 prefers ${NODELIST[0]} pcs constraint location p-2 avoids ${NODELIST[0]} pcs resource create s-1 ocf:pacemaker:Stateful promotable --no-default-ops pcs constraint location s-1-clone rule role=master "#uname" eq ${NODELIST[0]} pcs resource create oc-1 ocf:pacemaker:Dummy --no-default-ops pcs resource create oc-2 ocf:pacemaker:Dummy --no-default-ops pcs constraint order oc-1 then oc-2 pcs constraint colocation add oc-2 with oc-1 pcs resource create oc-set-1 ocf:pacemaker:Dummy --no-default-ops pcs resource create oc-set-2 ocf:pacemaker:Dummy --no-default-ops pcs constraint order set oc-set-1 oc-set-2 pcs constraint colocation set oc-set-2 oc-set-1 pcs resource create t ocf:pacemaker:Dummy --no-default-ops pcs constraint ticket add Ticket t pcs constraint ticket set p-1 p-2 setoptions ticket=Ticket-set pcs alert create path=/usr/bin/true id=Alert pcs alert recipient add Alert value=recipient-value pcs resource defaults resource-stickiness=2 pcs resource op defaults timeout=90 pcs property set maintenance-mode=false pcs tag create TAG p-1 p-2 pcs resource defaults set create id=set-1 meta target-role=Started pcs resource op defaults set create id=op-set-1 score=10 meta interval=30s Test `pcs config checkpoint diff` commmand: pcs config checkpoint diff 1 live pcs config checkpoint diff live 1 DevTestResults: [root@r08-09-a ~]# rpm -q pcs pcs-0.10.16-1.el8.x86_64 [root@r08-09-a ~]# pcs config checkpoint diff 1 live Differences between checkpoint 1 (-) and live configuration (+): Resources: + Resource: r1 (class=ocf provider=pacemaker type=Dummy) + Operations: + migrate_from: r1-migrate_from-interval-0s + interval=0s + timeout=20s + migrate_to: r1-migrate_to-interval-0s + interval=0s + timeout=20s + monitor: r1-monitor-interval-10s + interval=10s + timeout=20s + reload: r1-reload-interval-0s + interval=0s + timeout=20s + reload-agent: r1-reload-agent-interval-0s + interval=0s + timeout=20s + start: r1-start-interval-0s + interval=0s + timeout=20s + stop: r1-stop-interval-0s + interval=0s + timeout=20s + Resource: p-1 (class=ocf provider=pacemaker type=Dummy) + Operations: + monitor: p-1-monitor-interval-10s + interval=10s + timeout=20s + Resource: p-2 (class=ocf provider=pacemaker type=Dummy) + Operations: + monitor: p-2-monitor-interval-10s + interval=10s + timeout=20s + Resource: oc-1 (class=ocf provider=pacemaker type=Dummy) + Operations: + monitor: oc-1-monitor-interval-10s + interval=10s + timeout=20s + Resource: oc-2 (class=ocf provider=pacemaker type=Dummy) + Operations: + monitor: oc-2-monitor-interval-10s + interval=10s + timeout=20s + Resource: oc-set-1 (class=ocf provider=pacemaker type=Dummy) + Operations: + monitor: oc-set-1-monitor-interval-10s + interval=10s + timeout=20s + Resource: oc-set-2 (class=ocf provider=pacemaker type=Dummy) + Operations: + monitor: oc-set-2-monitor-interval-10s + interval=10s + timeout=20s + Resource: t (class=ocf provider=pacemaker type=Dummy) + Operations: + monitor: t-monitor-interval-10s + interval=10s + timeout=20s + Clone: s-1-clone + Meta Attributes: s-1-clone-meta_attributes + promotable=true + Resource: s-1 (class=ocf provider=pacemaker type=Stateful) + Operations: + monitor: s-1-monitor-interval-10s + interval=10s + timeout=20s + role=Master + monitor: s-1-monitor-interval-11s + interval=11s + timeout=20s + role=Slave Stonith Devices: + Resource: fence-1-r8-09-a.vm (class=stonith type=fence_xvm) + Operations: + monitor: fence-1-r8-09-a.vm-monitor-interval-60s + interval=60s + Resource: fence-1-r08-09-b.vm (class=stonith type=fence_xvm) + Operations: + monitor: fence-1-r08-09-b.vm-monitor-interval-60s + interval=60s + Resource: fence-2-r8-09-a.vm (class=stonith type=fence_xvm) + Operations: + monitor: fence-2-r8-09-a.vm-monitor-interval-60s + interval=60s + Resource: fence-2-r08-09-b.vm (class=stonith type=fence_xvm) + Operations: + monitor: fence-2-r08-09-b.vm-monitor-interval-60s + interval=60s + Resource: fence-1-r08-09-a.vm (class=stonith type=fence_xvm) + Operations: + monitor: fence-1-r08-09-a.vm-monitor-interval-60s + interval=60s + Resource: fence-2-r08-09-a.vm (class=stonith type=fence_xvm) + Operations: + monitor: fence-2-r08-09-a.vm-monitor-interval-60s + interval=60s Fencing Levels: + Target: r08-09-a.vm + Level 1 - fence-1-r08-09-a.vm + Level 2 - fence-2-r08-09-a.vm + Target: r08-09-b.vm + Level 1 - fence-1-r08-09-b.vm + Level 2 - fence-2-r08-09-b.vm Location Constraints: + Resource: p-1 + Enabled on: + Node: r08-09-a.vm (score:INFINITY) (id:location-p-1-r08-09-a.vm-INFINITY) + Resource: p-2 + Disabled on: + Node: r08-09-a.vm (score:-INFINITY) (id:location-p-2-r08-09-a.vm--INFINITY) + Resource: s-1-clone + Constraint: location-s-1-clone + Rule: role=Master score=INFINITY (id:location-s-1-clone-rule) Expression: #uname eq r08-09-a.vm (id:location-s-1-clone-rule-expr) Ordering Constraints: + start oc-1 then start oc-2 (kind:Mandatory) (id:order-oc-1-oc-2-mandatory) + Resource Sets: + set oc-set-1 oc-set-2 (id:order_set_o1o2_set) (id:order_set_o1o2) Colocation Constraints: + oc-2 with oc-1 (score:INFINITY) (id:colocation-oc-2-oc-1-INFINITY) + Resource Sets: + set oc-set-2 oc-set-1 (id:colocation_set_o2o1_set) setoptions score=INFINITY (id:colocation_set_o2o1) Ticket Constraints: + t ticket=Ticket (id:ticket-Ticket-t) + Resource Sets: + set p-1 p-2 (id:ticket_set_p1p2_set) setoptions ticket=Ticket-set (id:ticket_set_p1p2) Alerts: - No alerts defined + Alert: Alert (path=/usr/bin/true) + Recipients: + Recipient: Alert-recipient (value=recipient-value) Resources Defaults: - No defaults set + Meta Attrs: rsc_defaults-meta_attributes + resource-stickiness=2 + Meta Attrs: set-1 + target-role=Started Operations Defaults: - No defaults set + Meta Attrs: op_defaults-meta_attributes + timeout=90 + Meta Attrs: op-set-1 score=10 + interval=30s - Cluster Properties: + Cluster Properties: cib-bootstrap-options + cluster-infrastructure=corosync + cluster-name=pre1 + dc-version=2.1.5-8.el8-a3f44794f94 + have-watchdog=false + maintenance-mode=false + placement-strategy=minimal Tags: - No tags defined + TAG + p-1 + p-2 [root@virt-500 ~]# rpm -q pcs pcs-0.10.16-1.el8.x86_64 Starting tests, logging all output to /tmp/vedder.CHERRY.STSRHTS11062.202306291158 Starting XMLRPC server.... DEBUG:STSXMLRPC_DAEMON:Debugging enabled INFO:STSXMLRPC:Server started. Logging to /tmp/vedder.CHERRY.STSRHTS11062.202306291158/stsxmlrpc.log XMLRPC server started at http://virt-482:36909/ <start name="pcs,cli,CheckpointDiff" id="pcs,cli,CheckpointDiff" pid="256418" time="Thu Jun 29 11:58:17 2023 +0200" type="cmd" /> [pcs,cli,CheckpointDiff] 2023-06-29 11:58:17 INFO: PCS_CLUSTER_DESTROY [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: cluster destroyed. [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: CHECK_CLUSTER_HEALTH [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: SETUP_PCSD_PORTS [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: CLEAR_DAEMON_CONFIGS [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:58:22 INFO: [pcs,cli,CheckpointDiff] 2023-06-29 11:58:22 INFO: SET_PASSWORD [pcs,cli,CheckpointDiff] 2023-06-29 11:58:23 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:58:24 INFO: ENABLE_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:58:24 INFO: PCS_HOST_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 11:58:24 INFO: running auth from virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:58:26 INFO: PCS_STATUS_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 11:58:27 INFO: PCS_CLUSTER_SETUP [pcs,cli,CheckpointDiff] 2023-06-29 11:58:27 INFO: running: pcs cluster setup STSRHTS11062 virt-500 virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:58:32 INFO: cluster created [pcs,cli,CheckpointDiff] 2023-06-29 11:58:32 INFO: PCS_CLUSTER_START [pcs,cli,CheckpointDiff] 2023-06-29 11:58:32 INFO: starting cluster from virt-500 with --all [pcs,cli,CheckpointDiff] 2023-06-29 11:58:58 INFO: cluster started [pcs,cli,CheckpointDiff] 2023-06-29 11:58:58 INFO: == SUBTEST: StonithCheckpointDiff == [pcs,cli,CheckpointDiff] 2023-06-29 11:58:58 INFO: PCS_STONITH_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 11:58:58 INFO: running: pcs stonith create fence-virt-500 fence_xvm pcmk_host_list=virt-500 on virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:58:59 INFO: Stonith device fence-virt-500 created. [pcs,cli,CheckpointDiff] 2023-06-29 11:58:59 INFO: GET_DEVICES_BY_NODE [pcs,cli,CheckpointDiff] 2023-06-29 11:58:59 INFO: PCS_STONITH_LEVEL_ADD [pcs,cli,CheckpointDiff] 2023-06-29 11:58:59 INFO: running: pcs stonith level add 1 virt-500 fence-virt-500 on virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:00 INFO: PCS_STONITH_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:00 INFO: running: pcs stonith create fence-virt-501 fence_xvm pcmk_host_list=virt-501 on virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:01 INFO: Stonith device fence-virt-501 created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:01 INFO: GET_DEVICES_BY_NODE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:01 INFO: PCS_STONITH_LEVEL_ADD [pcs,cli,CheckpointDiff] 2023-06-29 11:59:01 INFO: running: pcs stonith level add 2 virt-501 fence-virt-501 on virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:02 INFO: PCS_CONFIG_CHECKPOINT_DIFF [pcs,cli,CheckpointDiff] 2023-06-29 11:59:02 INFO: running: pcs config checkpoint diff 1 live [pcs,cli,CheckpointDiff] 2023-06-29 11:59:04 INFO: Differences between checkpoint 1 (-) and live configuration (+): [pcs,cli,CheckpointDiff] Resources: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Stonith Devices: [pcs,cli,CheckpointDiff] + Resource: fence-virt-500 (class=stonith type=fence_xvm) [pcs,cli,CheckpointDiff] + Attributes: fence-virt-500-instance_attributes [pcs,cli,CheckpointDiff] + pcmk_host_list=virt-500 [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: fence-virt-500-monitor-interval-60s [pcs,cli,CheckpointDiff] + interval=60s [pcs,cli,CheckpointDiff] + Resource: fence-virt-501 (class=stonith type=fence_xvm) [pcs,cli,CheckpointDiff] + Attributes: fence-virt-501-instance_attributes [pcs,cli,CheckpointDiff] + pcmk_host_list=virt-501 [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: fence-virt-501-monitor-interval-60s [pcs,cli,CheckpointDiff] + interval=60s [pcs,cli,CheckpointDiff] Fencing Levels: [pcs,cli,CheckpointDiff] + Target: virt-500 [pcs,cli,CheckpointDiff] + Level 1 - fence-virt-500 [pcs,cli,CheckpointDiff] + Target: virt-501 [pcs,cli,CheckpointDiff] + Level 2 - fence-virt-501 [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Location Constraints: [pcs,cli,CheckpointDiff] Ordering Constraints: [pcs,cli,CheckpointDiff] Colocation Constraints: [pcs,cli,CheckpointDiff] Ticket Constraints: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Alerts: [pcs,cli,CheckpointDiff] No alerts defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Resources Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] Operations Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] - Cluster Properties: [pcs,cli,CheckpointDiff] + Cluster Properties: cib-bootstrap-options [pcs,cli,CheckpointDiff] + cluster-infrastructure=corosync [pcs,cli,CheckpointDiff] + cluster-name=STSRHTS11062 [pcs,cli,CheckpointDiff] + dc-version=2.1.6-1.el8-6fdc9deea29 [pcs,cli,CheckpointDiff] + have-watchdog=false [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Tags: [pcs,cli,CheckpointDiff] No tags defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] 2023-06-29 11:59:05 INFO: PCS_STONITH_LEVEL_REMOVE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:05 INFO: running: pcs stonith level remove 1 target virt-500 stonith fence-virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:06 INFO: PCS_STONITH_LEVEL_REMOVE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:06 INFO: running: pcs stonith level remove 2 target virt-501 stonith fence-virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:08 INFO: PCS_CLUSTER_DESTROY [pcs,cli,CheckpointDiff] 2023-06-29 11:59:11 INFO: cluster destroyed. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:11 INFO: CHECK_CLUSTER_HEALTH [pcs,cli,CheckpointDiff] 2023-06-29 11:59:12 INFO: SETUP_PCSD_PORTS [pcs,cli,CheckpointDiff] 2023-06-29 11:59:12 INFO: CLEAR_DAEMON_CONFIGS [pcs,cli,CheckpointDiff] 2023-06-29 11:59:12 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:59:14 INFO: [pcs,cli,CheckpointDiff] 2023-06-29 11:59:14 INFO: SET_PASSWORD [pcs,cli,CheckpointDiff] 2023-06-29 11:59:14 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:59:16 INFO: ENABLE_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:59:16 INFO: PCS_HOST_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 11:59:16 INFO: running auth from virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:18 INFO: PCS_STATUS_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 11:59:19 INFO: PCS_CLUSTER_SETUP [pcs,cli,CheckpointDiff] 2023-06-29 11:59:19 INFO: running: pcs cluster setup STSRHTS11062 virt-500 virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:24 INFO: cluster created [pcs,cli,CheckpointDiff] 2023-06-29 11:59:24 INFO: PCS_CLUSTER_START [pcs,cli,CheckpointDiff] 2023-06-29 11:59:24 INFO: starting cluster from virt-500 with --all [pcs,cli,CheckpointDiff] 2023-06-29 11:59:51 INFO: cluster started [pcs,cli,CheckpointDiff] 2023-06-29 11:59:51 INFO: == SUBTEST: ResourceConstraintLocationCheckpointDiff == [pcs,cli,CheckpointDiff] 2023-06-29 11:59:51 INFO: PCS_RESOURCE_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:51 INFO: running: pcs resource create d-0 ocf:heartbeat:Dummy --no-default-ops [pcs,cli,CheckpointDiff] 2023-06-29 11:59:52 INFO: Resource created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:52 INFO: PCS_LOCATION_PREFERS [pcs,cli,CheckpointDiff] 2023-06-29 11:59:52 INFO: running: pcs constraint location d-0 prefers virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:53 INFO: Location constraint created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:53 INFO: PCS_RESOURCE_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:53 INFO: running: pcs resource create d-1 ocf:heartbeat:Dummy --no-default-ops [pcs,cli,CheckpointDiff] 2023-06-29 11:59:54 INFO: Resource created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:54 INFO: PCS_LOCATION_PREFERS [pcs,cli,CheckpointDiff] 2023-06-29 11:59:54 INFO: running: pcs constraint location d-1 prefers virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:55 INFO: Location constraint created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:55 INFO: PCS_CONFIG_CHECKPOINT_DIFF [pcs,cli,CheckpointDiff] 2023-06-29 11:59:55 INFO: running: pcs config checkpoint diff 1 live [pcs,cli,CheckpointDiff] 2023-06-29 11:59:56 INFO: Differences between checkpoint 1 (-) and live configuration (+): [pcs,cli,CheckpointDiff] Resources: [pcs,cli,CheckpointDiff] + Resource: d-0 (class=ocf provider=heartbeat type=Dummy) [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: d-0-monitor-interval-10s [pcs,cli,CheckpointDiff] + interval=10s [pcs,cli,CheckpointDiff] + timeout=20s [pcs,cli,CheckpointDiff] + Resource: d-1 (class=ocf provider=heartbeat type=Dummy) [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: d-1-monitor-interval-10s [pcs,cli,CheckpointDiff] + interval=10s [pcs,cli,CheckpointDiff] + timeout=20s [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Stonith Devices: [pcs,cli,CheckpointDiff] Fencing Levels: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Location Constraints: [pcs,cli,CheckpointDiff] + Resource: d-0 [pcs,cli,CheckpointDiff] + Enabled on: [pcs,cli,CheckpointDiff] + Node: virt-500 (score:INFINITY) (id:location-d-0-virt-500-INFINITY) [pcs,cli,CheckpointDiff] + Resource: d-1 [pcs,cli,CheckpointDiff] + Enabled on: [pcs,cli,CheckpointDiff] + Node: virt-501 (score:INFINITY) (id:location-d-1-virt-501-INFINITY) [pcs,cli,CheckpointDiff] Ordering Constraints: [pcs,cli,CheckpointDiff] Colocation Constraints: [pcs,cli,CheckpointDiff] Ticket Constraints: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Alerts: [pcs,cli,CheckpointDiff] No alerts defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Resources Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] Operations Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] - Cluster Properties: [pcs,cli,CheckpointDiff] + Cluster Properties: cib-bootstrap-options [pcs,cli,CheckpointDiff] + cluster-infrastructure=corosync [pcs,cli,CheckpointDiff] + cluster-name=STSRHTS11062 [pcs,cli,CheckpointDiff] + dc-version=2.1.6-1.el8-6fdc9deea29 [pcs,cli,CheckpointDiff] + have-watchdog=false [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Tags: [pcs,cli,CheckpointDiff] No tags defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] 2023-06-29 11:59:57 INFO: PCS_RESOURCE_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:57 INFO: running: pcs resource delete d-0 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:58 INFO: Resource deleted. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:58 INFO: PCS_RESOURCE_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:58 INFO: running: pcs resource delete d-1 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:00 INFO: Resource deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:01 INFO: PCS_CLUSTER_DESTROY [pcs,cli,CheckpointDiff] 2023-06-29 12:00:04 INFO: cluster destroyed. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:04 INFO: CHECK_CLUSTER_HEALTH [pcs,cli,CheckpointDiff] 2023-06-29 12:00:04 INFO: SETUP_PCSD_PORTS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:04 INFO: CLEAR_DAEMON_CONFIGS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:05 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 12:00:07 INFO: [pcs,cli,CheckpointDiff] 2023-06-29 12:00:07 INFO: SET_PASSWORD [pcs,cli,CheckpointDiff] 2023-06-29 12:00:07 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 12:00:09 INFO: ENABLE_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 12:00:09 INFO: PCS_HOST_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 12:00:09 INFO: running auth from virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:10 INFO: PCS_STATUS_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 12:00:11 INFO: PCS_CLUSTER_SETUP [pcs,cli,CheckpointDiff] 2023-06-29 12:00:11 INFO: running: pcs cluster setup STSRHTS11062 virt-500 virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:16 INFO: cluster created [pcs,cli,CheckpointDiff] 2023-06-29 12:00:16 INFO: PCS_CLUSTER_START [pcs,cli,CheckpointDiff] 2023-06-29 12:00:16 INFO: starting cluster from virt-500 with --all [pcs,cli,CheckpointDiff] 2023-06-29 12:00:42 INFO: cluster started [pcs,cli,CheckpointDiff] 2023-06-29 12:00:42 INFO: == SUBTEST: TagCheckpointDiff == [pcs,cli,CheckpointDiff] 2023-06-29 12:00:42 INFO: PCS_RESOURCE_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:42 INFO: running: pcs resource create d-0 ocf:heartbeat:Dummy --no-default-ops [pcs,cli,CheckpointDiff] 2023-06-29 12:00:43 INFO: Resource created. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:43 INFO: PCS_TAG_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:43 INFO: running: pcs tag create tag-d-0 d-0 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:44 INFO: Tag created. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:44 INFO: PCS_RESOURCE_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:44 INFO: running: pcs resource create d-1 ocf:heartbeat:Dummy --no-default-ops [pcs,cli,CheckpointDiff] 2023-06-29 12:00:45 INFO: Resource created. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:45 INFO: PCS_TAG_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:45 INFO: running: pcs tag create tag-d-1 d-1 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:46 INFO: Tag created. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:46 INFO: PCS_CONFIG_CHECKPOINT_DIFF [pcs,cli,CheckpointDiff] 2023-06-29 12:00:46 INFO: running: pcs config checkpoint diff 1 live [pcs,cli,CheckpointDiff] 2023-06-29 12:00:48 INFO: Differences between checkpoint 1 (-) and live configuration (+): [pcs,cli,CheckpointDiff] Resources: [pcs,cli,CheckpointDiff] + Resource: d-0 (class=ocf provider=heartbeat type=Dummy) [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: d-0-monitor-interval-10s [pcs,cli,CheckpointDiff] + interval=10s [pcs,cli,CheckpointDiff] + timeout=20s [pcs,cli,CheckpointDiff] + Resource: d-1 (class=ocf provider=heartbeat type=Dummy) [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: d-1-monitor-interval-10s [pcs,cli,CheckpointDiff] + interval=10s [pcs,cli,CheckpointDiff] + timeout=20s [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Stonith Devices: [pcs,cli,CheckpointDiff] Fencing Levels: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Location Constraints: [pcs,cli,CheckpointDiff] Ordering Constraints: [pcs,cli,CheckpointDiff] Colocation Constraints: [pcs,cli,CheckpointDiff] Ticket Constraints: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Alerts: [pcs,cli,CheckpointDiff] No alerts defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Resources Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] Operations Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] - Cluster Properties: [pcs,cli,CheckpointDiff] + Cluster Properties: cib-bootstrap-options [pcs,cli,CheckpointDiff] + cluster-infrastructure=corosync [pcs,cli,CheckpointDiff] + cluster-name=STSRHTS11062 [pcs,cli,CheckpointDiff] + dc-version=2.1.6-1.el8-6fdc9deea29 [pcs,cli,CheckpointDiff] + have-watchdog=false [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Tags: [pcs,cli,CheckpointDiff] - No tags defined [pcs,cli,CheckpointDiff] + tag-d-0 [pcs,cli,CheckpointDiff] + d-0 [pcs,cli,CheckpointDiff] + tag-d-1 [pcs,cli,CheckpointDiff] + d-1 [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] 2023-06-29 12:00:49 INFO: PCS_TAG_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:49 INFO: running: pcs tag delete tag-d-0 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:50 INFO: Tag deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:50 INFO: PCS_RESOURCE_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:50 INFO: running: pcs resource delete d-0 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:51 INFO: Resource deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:51 INFO: PCS_TAG_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:51 INFO: running: pcs tag delete tag-d-1 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:52 INFO: Tag deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:52 INFO: PCS_RESOURCE_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:52 INFO: running: pcs resource delete d-1 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:53 INFO: Resource deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:54 INFO: CHECK_LOGS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:54 INFO: CHECK_LOGS_FOR_CIB_REPLACE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:54 INFO: PCS_CLUSTER_DESTROY [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: cluster destroyed. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: CHECK_CLUSTER_HEALTH [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: SETUP_PCSD_PORTS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: CLEAR_DAEMON_CONFIGS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: RESTART_DAEMON <pass name="pcs,cli,CheckpointDiff" id="pcs,cli,CheckpointDiff" pid="256418" time="Thu Jun 29 12:01:00 2023 +0200" type="cmd" duration="163" /> ------------------- Summary --------------------- Testcase Result -------- ------ pcs,cli,CheckpointDiff PASS ================================================= Total Tests Run: 1 Total PASS: 1 Total FAIL: 0 Total TIMEOUT: 0 Total KILLED: 0 Total STOPPED: 0 Test output in /tmp/vedder.CHERRY.STSRHTS11062.202306291158 Killing XMLRPC server... DEBUG:STSXMLRPC_DAEMON:Debugging enabled DEBUG:STSXMLRPC:Killing server with PID 256416 (SIGTERM) INFO:STSXMLRPC:Server terminated. > Marking as VERIFIED for pcs-0.10.16-1.el8.x86_64 Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (pcs bug fix and enhancement update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2023:6903 |