Hide Forgot
Description of problem: Regression from RHEL6, makes it impossible to use clvmd in rhel7 when cluster nodes are offline. Version-Release number of selected component (if applicable): lvm2-cluster-2.02.99-1.el7.x86_64 How reproducible: always Steps to Reproduce: 1. start a 2 node cluster with clvmd and create a simple clustered vg/lv 2. clean shutdown one node (poweroff is fine) [root@rhel7-ha-node2 ~]# systemctl stop corosync (for example) 3. verify cluster node has left the membership (important bit) 4. try to remove the clustered lv. Actual results: [root@rhel7-ha-node1 ~]# lvremove /dev/cluster_vg/cluster_lv Do you really want to remove active clustered logical volume cluster_lv? [y/n]: y cluster request failed: Host is down Unable to deactivate logical volume "cluster_lv" cluster request failed: Host is down lv is not removed Expected results: Similar behaviour as rhel6: [root@rhel6-ha-node1 ~]# lvremove /dev/cluster_vg/cluster_lv Do you really want to remove active clustered logical volume cluster_lv? [y/n]: y Logical volume "cluster_lv" successfully removed Additional info:
on Agk request: it's not a regression in 6.5, it's a regression observed between rhel6.* and rhel7. lvm.conf is default with the only exception of locking_type set to 3.
Created attachment 799327 [details] lvremove -vvvv on rhel6 lvremove -vvvv on rhel6
Created attachment 799328 [details] lvremove -vvvv on rhel7 lvremove -vvvv on rhel7
rhel7 is with latest nightly build of lvm2 lvm2-cluster-2.02.101-0.157.el7.x86_64
Created attachment 799333 [details] clvmd debugging logs from node1 (node2 was poweroff)
lvm client side: Successful case: #locking/cluster_locking.c:502 Locking LV EQ4qhf7TgdAMeBaCOgZ0M57mqiIBTXIEhUdwleLadJmtgkYMEFu0Doqrw7k9OsAb NL (LV|NONBLOCK|CLUSTER) (0x98) Failure case: #locking/cluster_locking.c:502 Locking LV yDC7vdTMn3TGdEdEBGD3DPBcTFzHdR0tnKBwNY62WULjrIf9fUZ6vvFvcSb7gwO7 NL (LV|NONBLOCK|CLUSTER) (0x98) #locking/cluster_locking.c:161 cluster request failed: Host is down
Created attachment 799356 [details] rhel7 logs with syslog=1 loglevel debug
Created attachment 799383 [details] another attempt to capture logs
commit 431eda63cc0ebff7c62dacb313cabcffbda6573a Author: Christine Caulfield <ccaulfie> Date: Mon Sep 23 13:23:00 2013 +0100 clvmd: Fix node up/down handing in corosync module
In release 2.02.102.
As long as cluster is quorate, there are no issues removing the clustered LV. tested and verified with lvm2-2.02.103-5.el7 [root@virt-002 pacemaker]# lvremove clustered/mirror Do you really want to remove active clustered logical volume mirror? [y/n]: y Logical volume "mirror" successfully removed [root@virt-002 pacemaker]# pcs status Cluster name: STSRHTS10638 Last updated: Wed Nov 20 15:20:31 2013 Last change: Wed Nov 20 14:41:21 2013 via cibadmin on virt-002.cluster-qe.lab.eng.brq.redhat.com Stack: corosync Current DC: virt-002.cluster-qe.lab.eng.brq.redhat.com (1) - partition with quorum Version: 1.1.10-20.el7-368c726 3 Nodes configured 1 Resources configured Online: [ virt-002.cluster-qe.lab.eng.brq.redhat.com ] OFFLINE: [ virt-003.cluster-qe.lab.eng.brq.redhat.com virt-004.cluster-qe.lab.eng.brq.redhat.com ]
This request was resolved in Red Hat Enterprise Linux 7.0. Contact your manager or support representative in case you have further questions about the request.