Bug 2176490
| Summary: | Command 'pcs config checkpoint diff' does not show configuration differences between checkpoints | |||
|---|---|---|---|---|
| Product: | Red Hat Enterprise Linux 8 | Reporter: | Tomas Jelinek <tojeline> | |
| Component: | pcs | Assignee: | Miroslav Lisik <mlisik> | |
| Status: | VERIFIED --- | QA Contact: | cluster-qe <cluster-qe> | |
| Severity: | high | Docs Contact: | ||
| Priority: | high | |||
| Version: | 8.2 | CC: | cfeist, cluster-maint, idevat, mlisik, mmazoure, mpospisi, nhostako, omular, tojeline | |
| Target Milestone: | rc | Keywords: | Regression, Triaged, ZStream | |
| Target Release: | 8.9 | |||
| Hardware: | Unspecified | |||
| OS: | Unspecified | |||
| Whiteboard: | ||||
| Fixed In Version: | pcs-0.10.16-1.el8 | Doc Type: | Bug Fix | |
| Doc Text: |
Cause: Loading checkpoint CIB files did not work for code in the new architecture.
Consequence: Command `pcs config checkpoint diff` stopped to show differences for configuration sections after re-writing relevant code parts to the new architecture.
Fix: Load properly checkpoint CIB files in the new code architecture.
Result: Command `pcs config checkpoint diff` works again and shows differences for all configuration sections.
|
Story Points: | --- | |
| Clone Of: | 2175881 | |||
| : | 2180700 2180701 2180702 2180703 (view as bug list) | Environment: | ||
| Last Closed: | Type: | Bug | ||
| Regression: | --- | Mount Type: | --- | |
| Documentation: | --- | CRM: | ||
| Verified Versions: | Category: | --- | ||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
| Cloudforms Team: | --- | Target Upstream Version: | ||
| Embargoed: | ||||
| Bug Depends On: | ||||
| Bug Blocks: | 2180700, 2180701, 2180702, 2180703 | |||
|
Description
Tomas Jelinek
2023-03-08 14:13:15 UTC
Upstream commit: https://github.com/ClusterLabs/pcs/commit/2e225d1e46f962fddb1d6e6bd1d13a2348dc5c41 Updated commands: * pcs config checkpoint diff Test: Prepare configuration for `pcs config`: export NODELIST=(r8-node-01 r8-node-02 r8-node-03) pcs host auth -u hacluster -p $PASSWORD ${NODELIST[*]} pcs cluster setup HACluster ${NODELIST[*]} --start --wait for node in ${NODELIST[*]}; do pcs stonith create fence-1-$node fence_xvm; done for node in ${NODELIST[*]}; do pcs stonith create fence-2-$node fence_xvm; done for node in ${NODELIST[*]}; do pcs stonith level add 1 $node fence-1-$node; pcs stonith level add 2 $node fence-2-$node; done pcs resource create p-1 ocf:pacemaker:Dummy --no-default-ops pcs resource create p-2 ocf:pacemaker:Dummy --no-default-ops pcs constraint location p-1 prefers ${NODELIST[0]} pcs constraint location p-2 avoids ${NODELIST[0]} pcs resource create s-1 ocf:pacemaker:Stateful promotable --no-default-ops pcs constraint location s-1-clone rule role=master "#uname" eq ${NODELIST[0]} pcs resource create oc-1 ocf:pacemaker:Dummy --no-default-ops pcs resource create oc-2 ocf:pacemaker:Dummy --no-default-ops pcs constraint order oc-1 then oc-2 pcs constraint colocation add oc-2 with oc-1 pcs resource create oc-set-1 ocf:pacemaker:Dummy --no-default-ops pcs resource create oc-set-2 ocf:pacemaker:Dummy --no-default-ops pcs constraint order set oc-set-1 oc-set-2 pcs constraint colocation set oc-set-2 oc-set-1 pcs resource create t ocf:pacemaker:Dummy --no-default-ops pcs constraint ticket add Ticket t pcs constraint ticket set p-1 p-2 setoptions ticket=Ticket-set pcs alert create path=/usr/bin/true id=Alert pcs alert recipient add Alert value=recipient-value pcs resource defaults resource-stickiness=2 pcs resource op defaults timeout=90 pcs property set maintenance-mode=false pcs tag create TAG p-1 p-2 pcs resource defaults set create id=set-1 meta target-role=Started pcs resource op defaults set create id=op-set-1 score=10 meta interval=30s Test `pcs config checkpoint diff` commmand: pcs config checkpoint diff 1 live pcs config checkpoint diff live 1 DevTestResults:
[root@r08-09-a ~]# rpm -q pcs
pcs-0.10.16-1.el8.x86_64
[root@r08-09-a ~]# pcs config checkpoint diff 1 live
Differences between checkpoint 1 (-) and live configuration (+):
Resources:
+ Resource: r1 (class=ocf provider=pacemaker type=Dummy)
+ Operations:
+ migrate_from: r1-migrate_from-interval-0s
+ interval=0s
+ timeout=20s
+ migrate_to: r1-migrate_to-interval-0s
+ interval=0s
+ timeout=20s
+ monitor: r1-monitor-interval-10s
+ interval=10s
+ timeout=20s
+ reload: r1-reload-interval-0s
+ interval=0s
+ timeout=20s
+ reload-agent: r1-reload-agent-interval-0s
+ interval=0s
+ timeout=20s
+ start: r1-start-interval-0s
+ interval=0s
+ timeout=20s
+ stop: r1-stop-interval-0s
+ interval=0s
+ timeout=20s
+ Resource: p-1 (class=ocf provider=pacemaker type=Dummy)
+ Operations:
+ monitor: p-1-monitor-interval-10s
+ interval=10s
+ timeout=20s
+ Resource: p-2 (class=ocf provider=pacemaker type=Dummy)
+ Operations:
+ monitor: p-2-monitor-interval-10s
+ interval=10s
+ timeout=20s
+ Resource: oc-1 (class=ocf provider=pacemaker type=Dummy)
+ Operations:
+ monitor: oc-1-monitor-interval-10s
+ interval=10s
+ timeout=20s
+ Resource: oc-2 (class=ocf provider=pacemaker type=Dummy)
+ Operations:
+ monitor: oc-2-monitor-interval-10s
+ interval=10s
+ timeout=20s
+ Resource: oc-set-1 (class=ocf provider=pacemaker type=Dummy)
+ Operations:
+ monitor: oc-set-1-monitor-interval-10s
+ interval=10s
+ timeout=20s
+ Resource: oc-set-2 (class=ocf provider=pacemaker type=Dummy)
+ Operations:
+ monitor: oc-set-2-monitor-interval-10s
+ interval=10s
+ timeout=20s
+ Resource: t (class=ocf provider=pacemaker type=Dummy)
+ Operations:
+ monitor: t-monitor-interval-10s
+ interval=10s
+ timeout=20s
+ Clone: s-1-clone
+ Meta Attributes: s-1-clone-meta_attributes
+ promotable=true
+ Resource: s-1 (class=ocf provider=pacemaker type=Stateful)
+ Operations:
+ monitor: s-1-monitor-interval-10s
+ interval=10s
+ timeout=20s
+ role=Master
+ monitor: s-1-monitor-interval-11s
+ interval=11s
+ timeout=20s
+ role=Slave
Stonith Devices:
+ Resource: fence-1-r8-09-a.vm (class=stonith type=fence_xvm)
+ Operations:
+ monitor: fence-1-r8-09-a.vm-monitor-interval-60s
+ interval=60s
+ Resource: fence-1-r08-09-b.vm (class=stonith type=fence_xvm)
+ Operations:
+ monitor: fence-1-r08-09-b.vm-monitor-interval-60s
+ interval=60s
+ Resource: fence-2-r8-09-a.vm (class=stonith type=fence_xvm)
+ Operations:
+ monitor: fence-2-r8-09-a.vm-monitor-interval-60s
+ interval=60s
+ Resource: fence-2-r08-09-b.vm (class=stonith type=fence_xvm)
+ Operations:
+ monitor: fence-2-r08-09-b.vm-monitor-interval-60s
+ interval=60s
+ Resource: fence-1-r08-09-a.vm (class=stonith type=fence_xvm)
+ Operations:
+ monitor: fence-1-r08-09-a.vm-monitor-interval-60s
+ interval=60s
+ Resource: fence-2-r08-09-a.vm (class=stonith type=fence_xvm)
+ Operations:
+ monitor: fence-2-r08-09-a.vm-monitor-interval-60s
+ interval=60s
Fencing Levels:
+ Target: r08-09-a.vm
+ Level 1 - fence-1-r08-09-a.vm
+ Level 2 - fence-2-r08-09-a.vm
+ Target: r08-09-b.vm
+ Level 1 - fence-1-r08-09-b.vm
+ Level 2 - fence-2-r08-09-b.vm
Location Constraints:
+ Resource: p-1
+ Enabled on:
+ Node: r08-09-a.vm (score:INFINITY) (id:location-p-1-r08-09-a.vm-INFINITY)
+ Resource: p-2
+ Disabled on:
+ Node: r08-09-a.vm (score:-INFINITY) (id:location-p-2-r08-09-a.vm--INFINITY)
+ Resource: s-1-clone
+ Constraint: location-s-1-clone
+ Rule: role=Master score=INFINITY (id:location-s-1-clone-rule)
Expression: #uname eq r08-09-a.vm (id:location-s-1-clone-rule-expr)
Ordering Constraints:
+ start oc-1 then start oc-2 (kind:Mandatory) (id:order-oc-1-oc-2-mandatory)
+ Resource Sets:
+ set oc-set-1 oc-set-2 (id:order_set_o1o2_set) (id:order_set_o1o2)
Colocation Constraints:
+ oc-2 with oc-1 (score:INFINITY) (id:colocation-oc-2-oc-1-INFINITY)
+ Resource Sets:
+ set oc-set-2 oc-set-1 (id:colocation_set_o2o1_set) setoptions score=INFINITY (id:colocation_set_o2o1)
Ticket Constraints:
+ t ticket=Ticket (id:ticket-Ticket-t)
+ Resource Sets:
+ set p-1 p-2 (id:ticket_set_p1p2_set) setoptions ticket=Ticket-set (id:ticket_set_p1p2)
Alerts:
- No alerts defined
+ Alert: Alert (path=/usr/bin/true)
+ Recipients:
+ Recipient: Alert-recipient (value=recipient-value)
Resources Defaults:
- No defaults set
+ Meta Attrs: rsc_defaults-meta_attributes
+ resource-stickiness=2
+ Meta Attrs: set-1
+ target-role=Started
Operations Defaults:
- No defaults set
+ Meta Attrs: op_defaults-meta_attributes
+ timeout=90
+ Meta Attrs: op-set-1 score=10
+ interval=30s
- Cluster Properties:
+ Cluster Properties: cib-bootstrap-options
+ cluster-infrastructure=corosync
+ cluster-name=pre1
+ dc-version=2.1.5-8.el8-a3f44794f94
+ have-watchdog=false
+ maintenance-mode=false
+ placement-strategy=minimal
Tags:
- No tags defined
+ TAG
+ p-1
+ p-2
[root@virt-500 ~]# rpm -q pcs pcs-0.10.16-1.el8.x86_64 Starting tests, logging all output to /tmp/vedder.CHERRY.STSRHTS11062.202306291158 Starting XMLRPC server.... DEBUG:STSXMLRPC_DAEMON:Debugging enabled INFO:STSXMLRPC:Server started. Logging to /tmp/vedder.CHERRY.STSRHTS11062.202306291158/stsxmlrpc.log XMLRPC server started at http://virt-482:36909/ <start name="pcs,cli,CheckpointDiff" id="pcs,cli,CheckpointDiff" pid="256418" time="Thu Jun 29 11:58:17 2023 +0200" type="cmd" /> [pcs,cli,CheckpointDiff] 2023-06-29 11:58:17 INFO: PCS_CLUSTER_DESTROY [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: cluster destroyed. [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: CHECK_CLUSTER_HEALTH [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: SETUP_PCSD_PORTS [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: CLEAR_DAEMON_CONFIGS [pcs,cli,CheckpointDiff] 2023-06-29 11:58:20 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:58:22 INFO: [pcs,cli,CheckpointDiff] 2023-06-29 11:58:22 INFO: SET_PASSWORD [pcs,cli,CheckpointDiff] 2023-06-29 11:58:23 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:58:24 INFO: ENABLE_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:58:24 INFO: PCS_HOST_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 11:58:24 INFO: running auth from virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:58:26 INFO: PCS_STATUS_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 11:58:27 INFO: PCS_CLUSTER_SETUP [pcs,cli,CheckpointDiff] 2023-06-29 11:58:27 INFO: running: pcs cluster setup STSRHTS11062 virt-500 virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:58:32 INFO: cluster created [pcs,cli,CheckpointDiff] 2023-06-29 11:58:32 INFO: PCS_CLUSTER_START [pcs,cli,CheckpointDiff] 2023-06-29 11:58:32 INFO: starting cluster from virt-500 with --all [pcs,cli,CheckpointDiff] 2023-06-29 11:58:58 INFO: cluster started [pcs,cli,CheckpointDiff] 2023-06-29 11:58:58 INFO: == SUBTEST: StonithCheckpointDiff == [pcs,cli,CheckpointDiff] 2023-06-29 11:58:58 INFO: PCS_STONITH_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 11:58:58 INFO: running: pcs stonith create fence-virt-500 fence_xvm pcmk_host_list=virt-500 on virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:58:59 INFO: Stonith device fence-virt-500 created. [pcs,cli,CheckpointDiff] 2023-06-29 11:58:59 INFO: GET_DEVICES_BY_NODE [pcs,cli,CheckpointDiff] 2023-06-29 11:58:59 INFO: PCS_STONITH_LEVEL_ADD [pcs,cli,CheckpointDiff] 2023-06-29 11:58:59 INFO: running: pcs stonith level add 1 virt-500 fence-virt-500 on virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:00 INFO: PCS_STONITH_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:00 INFO: running: pcs stonith create fence-virt-501 fence_xvm pcmk_host_list=virt-501 on virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:01 INFO: Stonith device fence-virt-501 created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:01 INFO: GET_DEVICES_BY_NODE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:01 INFO: PCS_STONITH_LEVEL_ADD [pcs,cli,CheckpointDiff] 2023-06-29 11:59:01 INFO: running: pcs stonith level add 2 virt-501 fence-virt-501 on virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:02 INFO: PCS_CONFIG_CHECKPOINT_DIFF [pcs,cli,CheckpointDiff] 2023-06-29 11:59:02 INFO: running: pcs config checkpoint diff 1 live [pcs,cli,CheckpointDiff] 2023-06-29 11:59:04 INFO: Differences between checkpoint 1 (-) and live configuration (+): [pcs,cli,CheckpointDiff] Resources: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Stonith Devices: [pcs,cli,CheckpointDiff] + Resource: fence-virt-500 (class=stonith type=fence_xvm) [pcs,cli,CheckpointDiff] + Attributes: fence-virt-500-instance_attributes [pcs,cli,CheckpointDiff] + pcmk_host_list=virt-500 [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: fence-virt-500-monitor-interval-60s [pcs,cli,CheckpointDiff] + interval=60s [pcs,cli,CheckpointDiff] + Resource: fence-virt-501 (class=stonith type=fence_xvm) [pcs,cli,CheckpointDiff] + Attributes: fence-virt-501-instance_attributes [pcs,cli,CheckpointDiff] + pcmk_host_list=virt-501 [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: fence-virt-501-monitor-interval-60s [pcs,cli,CheckpointDiff] + interval=60s [pcs,cli,CheckpointDiff] Fencing Levels: [pcs,cli,CheckpointDiff] + Target: virt-500 [pcs,cli,CheckpointDiff] + Level 1 - fence-virt-500 [pcs,cli,CheckpointDiff] + Target: virt-501 [pcs,cli,CheckpointDiff] + Level 2 - fence-virt-501 [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Location Constraints: [pcs,cli,CheckpointDiff] Ordering Constraints: [pcs,cli,CheckpointDiff] Colocation Constraints: [pcs,cli,CheckpointDiff] Ticket Constraints: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Alerts: [pcs,cli,CheckpointDiff] No alerts defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Resources Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] Operations Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] - Cluster Properties: [pcs,cli,CheckpointDiff] + Cluster Properties: cib-bootstrap-options [pcs,cli,CheckpointDiff] + cluster-infrastructure=corosync [pcs,cli,CheckpointDiff] + cluster-name=STSRHTS11062 [pcs,cli,CheckpointDiff] + dc-version=2.1.6-1.el8-6fdc9deea29 [pcs,cli,CheckpointDiff] + have-watchdog=false [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Tags: [pcs,cli,CheckpointDiff] No tags defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] 2023-06-29 11:59:05 INFO: PCS_STONITH_LEVEL_REMOVE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:05 INFO: running: pcs stonith level remove 1 target virt-500 stonith fence-virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:06 INFO: PCS_STONITH_LEVEL_REMOVE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:06 INFO: running: pcs stonith level remove 2 target virt-501 stonith fence-virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:08 INFO: PCS_CLUSTER_DESTROY [pcs,cli,CheckpointDiff] 2023-06-29 11:59:11 INFO: cluster destroyed. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:11 INFO: CHECK_CLUSTER_HEALTH [pcs,cli,CheckpointDiff] 2023-06-29 11:59:12 INFO: SETUP_PCSD_PORTS [pcs,cli,CheckpointDiff] 2023-06-29 11:59:12 INFO: CLEAR_DAEMON_CONFIGS [pcs,cli,CheckpointDiff] 2023-06-29 11:59:12 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:59:14 INFO: [pcs,cli,CheckpointDiff] 2023-06-29 11:59:14 INFO: SET_PASSWORD [pcs,cli,CheckpointDiff] 2023-06-29 11:59:14 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:59:16 INFO: ENABLE_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 11:59:16 INFO: PCS_HOST_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 11:59:16 INFO: running auth from virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:18 INFO: PCS_STATUS_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 11:59:19 INFO: PCS_CLUSTER_SETUP [pcs,cli,CheckpointDiff] 2023-06-29 11:59:19 INFO: running: pcs cluster setup STSRHTS11062 virt-500 virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:24 INFO: cluster created [pcs,cli,CheckpointDiff] 2023-06-29 11:59:24 INFO: PCS_CLUSTER_START [pcs,cli,CheckpointDiff] 2023-06-29 11:59:24 INFO: starting cluster from virt-500 with --all [pcs,cli,CheckpointDiff] 2023-06-29 11:59:51 INFO: cluster started [pcs,cli,CheckpointDiff] 2023-06-29 11:59:51 INFO: == SUBTEST: ResourceConstraintLocationCheckpointDiff == [pcs,cli,CheckpointDiff] 2023-06-29 11:59:51 INFO: PCS_RESOURCE_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:51 INFO: running: pcs resource create d-0 ocf:heartbeat:Dummy --no-default-ops [pcs,cli,CheckpointDiff] 2023-06-29 11:59:52 INFO: Resource created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:52 INFO: PCS_LOCATION_PREFERS [pcs,cli,CheckpointDiff] 2023-06-29 11:59:52 INFO: running: pcs constraint location d-0 prefers virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:53 INFO: Location constraint created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:53 INFO: PCS_RESOURCE_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:53 INFO: running: pcs resource create d-1 ocf:heartbeat:Dummy --no-default-ops [pcs,cli,CheckpointDiff] 2023-06-29 11:59:54 INFO: Resource created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:54 INFO: PCS_LOCATION_PREFERS [pcs,cli,CheckpointDiff] 2023-06-29 11:59:54 INFO: running: pcs constraint location d-1 prefers virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:55 INFO: Location constraint created. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:55 INFO: PCS_CONFIG_CHECKPOINT_DIFF [pcs,cli,CheckpointDiff] 2023-06-29 11:59:55 INFO: running: pcs config checkpoint diff 1 live [pcs,cli,CheckpointDiff] 2023-06-29 11:59:56 INFO: Differences between checkpoint 1 (-) and live configuration (+): [pcs,cli,CheckpointDiff] Resources: [pcs,cli,CheckpointDiff] + Resource: d-0 (class=ocf provider=heartbeat type=Dummy) [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: d-0-monitor-interval-10s [pcs,cli,CheckpointDiff] + interval=10s [pcs,cli,CheckpointDiff] + timeout=20s [pcs,cli,CheckpointDiff] + Resource: d-1 (class=ocf provider=heartbeat type=Dummy) [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: d-1-monitor-interval-10s [pcs,cli,CheckpointDiff] + interval=10s [pcs,cli,CheckpointDiff] + timeout=20s [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Stonith Devices: [pcs,cli,CheckpointDiff] Fencing Levels: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Location Constraints: [pcs,cli,CheckpointDiff] + Resource: d-0 [pcs,cli,CheckpointDiff] + Enabled on: [pcs,cli,CheckpointDiff] + Node: virt-500 (score:INFINITY) (id:location-d-0-virt-500-INFINITY) [pcs,cli,CheckpointDiff] + Resource: d-1 [pcs,cli,CheckpointDiff] + Enabled on: [pcs,cli,CheckpointDiff] + Node: virt-501 (score:INFINITY) (id:location-d-1-virt-501-INFINITY) [pcs,cli,CheckpointDiff] Ordering Constraints: [pcs,cli,CheckpointDiff] Colocation Constraints: [pcs,cli,CheckpointDiff] Ticket Constraints: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Alerts: [pcs,cli,CheckpointDiff] No alerts defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Resources Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] Operations Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] - Cluster Properties: [pcs,cli,CheckpointDiff] + Cluster Properties: cib-bootstrap-options [pcs,cli,CheckpointDiff] + cluster-infrastructure=corosync [pcs,cli,CheckpointDiff] + cluster-name=STSRHTS11062 [pcs,cli,CheckpointDiff] + dc-version=2.1.6-1.el8-6fdc9deea29 [pcs,cli,CheckpointDiff] + have-watchdog=false [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Tags: [pcs,cli,CheckpointDiff] No tags defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] 2023-06-29 11:59:57 INFO: PCS_RESOURCE_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:57 INFO: running: pcs resource delete d-0 [pcs,cli,CheckpointDiff] 2023-06-29 11:59:58 INFO: Resource deleted. [pcs,cli,CheckpointDiff] 2023-06-29 11:59:58 INFO: PCS_RESOURCE_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 11:59:58 INFO: running: pcs resource delete d-1 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:00 INFO: Resource deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:01 INFO: PCS_CLUSTER_DESTROY [pcs,cli,CheckpointDiff] 2023-06-29 12:00:04 INFO: cluster destroyed. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:04 INFO: CHECK_CLUSTER_HEALTH [pcs,cli,CheckpointDiff] 2023-06-29 12:00:04 INFO: SETUP_PCSD_PORTS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:04 INFO: CLEAR_DAEMON_CONFIGS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:05 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 12:00:07 INFO: [pcs,cli,CheckpointDiff] 2023-06-29 12:00:07 INFO: SET_PASSWORD [pcs,cli,CheckpointDiff] 2023-06-29 12:00:07 INFO: RESTART_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 12:00:09 INFO: ENABLE_DAEMON [pcs,cli,CheckpointDiff] 2023-06-29 12:00:09 INFO: PCS_HOST_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 12:00:09 INFO: running auth from virt-500 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:10 INFO: PCS_STATUS_AUTH [pcs,cli,CheckpointDiff] 2023-06-29 12:00:11 INFO: PCS_CLUSTER_SETUP [pcs,cli,CheckpointDiff] 2023-06-29 12:00:11 INFO: running: pcs cluster setup STSRHTS11062 virt-500 virt-501 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:16 INFO: cluster created [pcs,cli,CheckpointDiff] 2023-06-29 12:00:16 INFO: PCS_CLUSTER_START [pcs,cli,CheckpointDiff] 2023-06-29 12:00:16 INFO: starting cluster from virt-500 with --all [pcs,cli,CheckpointDiff] 2023-06-29 12:00:42 INFO: cluster started [pcs,cli,CheckpointDiff] 2023-06-29 12:00:42 INFO: == SUBTEST: TagCheckpointDiff == [pcs,cli,CheckpointDiff] 2023-06-29 12:00:42 INFO: PCS_RESOURCE_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:42 INFO: running: pcs resource create d-0 ocf:heartbeat:Dummy --no-default-ops [pcs,cli,CheckpointDiff] 2023-06-29 12:00:43 INFO: Resource created. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:43 INFO: PCS_TAG_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:43 INFO: running: pcs tag create tag-d-0 d-0 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:44 INFO: Tag created. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:44 INFO: PCS_RESOURCE_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:44 INFO: running: pcs resource create d-1 ocf:heartbeat:Dummy --no-default-ops [pcs,cli,CheckpointDiff] 2023-06-29 12:00:45 INFO: Resource created. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:45 INFO: PCS_TAG_CREATE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:45 INFO: running: pcs tag create tag-d-1 d-1 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:46 INFO: Tag created. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:46 INFO: PCS_CONFIG_CHECKPOINT_DIFF [pcs,cli,CheckpointDiff] 2023-06-29 12:00:46 INFO: running: pcs config checkpoint diff 1 live [pcs,cli,CheckpointDiff] 2023-06-29 12:00:48 INFO: Differences between checkpoint 1 (-) and live configuration (+): [pcs,cli,CheckpointDiff] Resources: [pcs,cli,CheckpointDiff] + Resource: d-0 (class=ocf provider=heartbeat type=Dummy) [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: d-0-monitor-interval-10s [pcs,cli,CheckpointDiff] + interval=10s [pcs,cli,CheckpointDiff] + timeout=20s [pcs,cli,CheckpointDiff] + Resource: d-1 (class=ocf provider=heartbeat type=Dummy) [pcs,cli,CheckpointDiff] + Operations: [pcs,cli,CheckpointDiff] + monitor: d-1-monitor-interval-10s [pcs,cli,CheckpointDiff] + interval=10s [pcs,cli,CheckpointDiff] + timeout=20s [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Stonith Devices: [pcs,cli,CheckpointDiff] Fencing Levels: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Location Constraints: [pcs,cli,CheckpointDiff] Ordering Constraints: [pcs,cli,CheckpointDiff] Colocation Constraints: [pcs,cli,CheckpointDiff] Ticket Constraints: [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Alerts: [pcs,cli,CheckpointDiff] No alerts defined [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Resources Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] Operations Defaults: [pcs,cli,CheckpointDiff] No defaults set [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] - Cluster Properties: [pcs,cli,CheckpointDiff] + Cluster Properties: cib-bootstrap-options [pcs,cli,CheckpointDiff] + cluster-infrastructure=corosync [pcs,cli,CheckpointDiff] + cluster-name=STSRHTS11062 [pcs,cli,CheckpointDiff] + dc-version=2.1.6-1.el8-6fdc9deea29 [pcs,cli,CheckpointDiff] + have-watchdog=false [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] Tags: [pcs,cli,CheckpointDiff] - No tags defined [pcs,cli,CheckpointDiff] + tag-d-0 [pcs,cli,CheckpointDiff] + d-0 [pcs,cli,CheckpointDiff] + tag-d-1 [pcs,cli,CheckpointDiff] + d-1 [pcs,cli,CheckpointDiff] [pcs,cli,CheckpointDiff] 2023-06-29 12:00:49 INFO: PCS_TAG_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:49 INFO: running: pcs tag delete tag-d-0 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:50 INFO: Tag deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:50 INFO: PCS_RESOURCE_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:50 INFO: running: pcs resource delete d-0 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:51 INFO: Resource deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:51 INFO: PCS_TAG_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:51 INFO: running: pcs tag delete tag-d-1 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:52 INFO: Tag deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:52 INFO: PCS_RESOURCE_DELETE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:52 INFO: running: pcs resource delete d-1 [pcs,cli,CheckpointDiff] 2023-06-29 12:00:53 INFO: Resource deleted. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:54 INFO: CHECK_LOGS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:54 INFO: CHECK_LOGS_FOR_CIB_REPLACE [pcs,cli,CheckpointDiff] 2023-06-29 12:00:54 INFO: PCS_CLUSTER_DESTROY [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: cluster destroyed. [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: CHECK_CLUSTER_HEALTH [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: SETUP_PCSD_PORTS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: CLEAR_DAEMON_CONFIGS [pcs,cli,CheckpointDiff] 2023-06-29 12:00:58 INFO: RESTART_DAEMON <pass name="pcs,cli,CheckpointDiff" id="pcs,cli,CheckpointDiff" pid="256418" time="Thu Jun 29 12:01:00 2023 +0200" type="cmd" duration="163" /> ------------------- Summary --------------------- Testcase Result -------- ------ pcs,cli,CheckpointDiff PASS ================================================= Total Tests Run: 1 Total PASS: 1 Total FAIL: 0 Total TIMEOUT: 0 Total KILLED: 0 Total STOPPED: 0 Test output in /tmp/vedder.CHERRY.STSRHTS11062.202306291158 Killing XMLRPC server... DEBUG:STSXMLRPC_DAEMON:Debugging enabled DEBUG:STSXMLRPC:Killing server with PID 256416 (SIGTERM) INFO:STSXMLRPC:Server terminated. > Marking as VERIFIED for pcs-0.10.16-1.el8.x86_64 |