Bug 1408110 - [GANESHA] Adding node to ganesha cluster is not assigning the correct VIP to the new node
Summary: [GANESHA] Adding node to ganesha cluster is not assigning the correct VIP to ...
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: GlusterFS
Classification: Community
Component: common-ha
Version: 3.9
Hardware: All
OS: All
urgent
high
Target Milestone: ---
Assignee: Soumya Koduri
QA Contact:
URL:
Whiteboard:
Depends On: 1406410
Blocks: 1406401
TreeView+ depends on / blocked
 
Reported: 2016-12-22 07:21 UTC by Soumya Koduri
Modified: 2017-03-08 10:23 UTC (History)
5 users (show)

Fixed In Version: glusterfs-3.9.1
Doc Type: If docs needed, set a value
Doc Text:
Clone Of: 1406410
Environment:
Last Closed: 2017-03-08 10:23:21 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Embargoed:


Attachments (Terms of Use)

Description Soumya Koduri 2016-12-22 07:21:04 UTC
+++ This bug was initially created as a clone of Bug #1406410 +++

+++ This bug was initially created as a clone of Bug #1406401 +++

Description of problem:
When a new node is being added to ganesha cluster,it should get the VIP same as mentioned in add node command.Instead the new node is being assigned with VIP of 1 of the node in existing cluster.

Version-Release number of selected component (if applicable):
# rpm -qa | grep ganesha
nfs-ganesha-2.4.1-3.el7rhgs.x86_64
glusterfs-ganesha-3.8.4-9.el7rhgs.x86_64
nfs-ganesha-gluster-2.4.1-3.el7rhgs.x86_64

How reproducible:
Consistently

Steps to Reproduce:
1.Create 4 node ganesha cluster and enable ganesha on it.
2.Add the new node to the existing ganesha cluster.
# /usr/libexec/ganesha/ganesha-ha.sh --add /var/run/gluster/shared_storage/nfs-ganesha/ dhcp47-59.lab.eng.blr.redhat.com 10.70.44.157

Node 1:
[root@dhcp46-219 ganesha]# ip addr         VIP 10.70.44.156
Node 2:
[root@dhcp47-45 ~]# ip addr                VIP 10.70.44.154
Node 3:
[root@dhcp47-3 nfs-ganesha]# ip addr       VIP 10.70.44.155
Node 4:
[root@dhcp46-241 ~]# ip addr               VIP 10.70.44.153

New Node which is being added to ganesha cluster:

[root@dhcp47-59 nfs-ganesha]# ip addr      VIP 10.70.44.154

======

[root@dhcp47-59 nfs-ganesha]# pcs status
Cluster name: ganesha-ha-360
Stack: corosync
Current DC: dhcp46-241.lab.eng.blr.redhat.com (version 1.1.15-11.el7_3.2-e174ec8) - partition with quorum
Last updated: Tue Dec 20 18:00:33 2016		Last change: Tue Dec 20 17:36:01 2016 by root via crm_attribute on dhcp47-59.lab.eng.blr.redhat.com

5 nodes and 30 resources configured

Online: [ dhcp46-219.lab.eng.blr.redhat.com dhcp46-241.lab.eng.blr.redhat.com dhcp47-3.lab.eng.blr.redhat.com dhcp47-45.lab.eng.blr.redhat.com dhcp47-59.lab.eng.blr.redhat.com ]

Full list of resources:

 Clone Set: nfs_setup-clone [nfs_setup]
     Started: [ dhcp46-219.lab.eng.blr.redhat.com dhcp46-241.lab.eng.blr.redhat.com dhcp47-3.lab.eng.blr.redhat.com dhcp47-45.lab.eng.blr.redhat.com dhcp47-59.lab.eng.blr.redhat.com ]
 Clone Set: nfs-mon-clone [nfs-mon]
     Started: [ dhcp46-219.lab.eng.blr.redhat.com dhcp46-241.lab.eng.blr.redhat.com dhcp47-3.lab.eng.blr.redhat.com dhcp47-45.lab.eng.blr.redhat.com dhcp47-59.lab.eng.blr.redhat.com ]
 Clone Set: nfs-grace-clone [nfs-grace]
     Started: [ dhcp46-219.lab.eng.blr.redhat.com dhcp46-241.lab.eng.blr.redhat.com dhcp47-3.lab.eng.blr.redhat.com dhcp47-45.lab.eng.blr.redhat.com dhcp47-59.lab.eng.blr.redhat.com ]
 Resource Group: dhcp46-219.lab.eng.blr.redhat.com-group
     dhcp46-219.lab.eng.blr.redhat.com-nfs_block	(ocf::heartbeat:portblock):	Started dhcp46-219.lab.eng.blr.redhat.com
     dhcp46-219.lab.eng.blr.redhat.com-cluster_ip-1	(ocf::heartbeat:IPaddr):	Started dhcp46-219.lab.eng.blr.redhat.com
     dhcp46-219.lab.eng.blr.redhat.com-nfs_unblock	(ocf::heartbeat:portblock):	Started dhcp46-219.lab.eng.blr.redhat.com
 Resource Group: dhcp46-241.lab.eng.blr.redhat.com-group
     dhcp46-241.lab.eng.blr.redhat.com-nfs_block	(ocf::heartbeat:portblock):	Started dhcp46-241.lab.eng.blr.redhat.com
     dhcp46-241.lab.eng.blr.redhat.com-cluster_ip-1	(ocf::heartbeat:IPaddr):	Started dhcp46-241.lab.eng.blr.redhat.com
     dhcp46-241.lab.eng.blr.redhat.com-nfs_unblock	(ocf::heartbeat:portblock):	Started dhcp46-241.lab.eng.blr.redhat.com
 Resource Group: dhcp47-3.lab.eng.blr.redhat.com-group
     dhcp47-3.lab.eng.blr.redhat.com-nfs_block	(ocf::heartbeat:portblock):	Started dhcp47-3.lab.eng.blr.redhat.com
     dhcp47-3.lab.eng.blr.redhat.com-cluster_ip-1	(ocf::heartbeat:IPaddr):	Started dhcp47-3.lab.eng.blr.redhat.com
     dhcp47-3.lab.eng.blr.redhat.com-nfs_unblock	(ocf::heartbeat:portblock):	Started dhcp47-3.lab.eng.blr.redhat.com
 Resource Group: dhcp47-45.lab.eng.blr.redhat.com-group
     dhcp47-45.lab.eng.blr.redhat.com-nfs_block	(ocf::heartbeat:portblock):	Started dhcp47-45.lab.eng.blr.redhat.com
     dhcp47-45.lab.eng.blr.redhat.com-cluster_ip-1	(ocf::heartbeat:IPaddr):	Started dhcp47-45.lab.eng.blr.redhat.com
     dhcp47-45.lab.eng.blr.redhat.com-nfs_unblock	(ocf::heartbeat:portblock):	Started dhcp47-45.lab.eng.blr.redhat.com
 Resource Group: dhcp47-59.lab.eng.blr.redhat.com-group
     dhcp47-59.lab.eng.blr.redhat.com-nfs_block	(ocf::heartbeat:portblock):	Started dhcp47-59.lab.eng.blr.redhat.com
     dhcp47-59.lab.eng.blr.redhat.com-cluster_ip-1	(ocf::heartbeat:IPaddr):	Started dhcp47-59.lab.eng.blr.redhat.com
     dhcp47-59.lab.eng.blr.redhat.com-nfs_unblock	(ocf::heartbeat:portblock):	Started dhcp47-59.lab.eng.blr.redhat.com

Failed Actions:
* dhcp47-45.lab.eng.blr.redhat.com-cluster_ip-1_monitor_15000 on dhcp47-45.lab.eng.blr.redhat.com 'not running' (7): call=123, status=complete, exitreason='none',
    last-rc-change='Tue Dec 20 17:36:01 2016', queued=0ms, exec=0ms
* dhcp46-241.lab.eng.blr.redhat.com-nfs_block_monitor_10000 on dhcp46-241.lab.eng.blr.redhat.com 'not running' (7): call=36, status=complete, exitreason='none',
    last-rc-change='Tue Dec 20 14:41:24 2016', queued=0ms, exec=0ms
* nfs-grace_monitor_5000 on dhcp47-59.lab.eng.blr.redhat.com 'not running' (7): call=69, status=complete, exitreason='none',
    last-rc-change='Tue Dec 20 17:35:56 2016', queued=0ms, exec=0ms


Daemon Status:
  corosync: active/disabled
  pacemaker: active/disabled
  pcsd: active/enabled

======

# cat ganesha-ha.conf
# Provide a unique name for the cluster.
HA_NAME="ganesha-ha-360"
# The subset of nodes of the Gluster Trusted Storage Pool that forms the ganesha
# HA cluster. Hostname should specified, IP addresses are not allowed.
# Maximum number of 16 nodes are supported.
HA_CLUSTER_NODES="dhcp46-219.lab.eng.blr.redhat.com,dhcp46-241.lab.eng.blr.redhat.com,dhcp47-3.lab.eng.blr.redhat.com,dhcp47-45.lab.eng.blr.redhat.com,dhcp47-59.lab.eng.blr.redhat.com"
# Virtual IPs of each of the nodes specified above.
VIP_dhcp46-241.lab.eng.blr.redhat.com="10.70.44.153"
VIP_dhcp47-45.lab.eng.blr.redhat.com="10.70.44.154"
VIP_dhcp47-3.lab.eng.blr.redhat.com="10.70.44.155"
VIP_dhcp46-219.lab.eng.blr.redhat.com="10.70.44.156"
VIP_dhcp47-59.lab.eng.blr.redhat.com="10.70.44.157"

Actual results:
New Node has the VIP 10.70.44.154 which is already assigned to 2nd node in existing ganesha cluster

Expected results:
New node should have VIP 10.70.44.157 which is being assigned in add node command

Additional info:

--- Additional comment from Red Hat Bugzilla Rules Engine on 2016-12-20 07:56:50 EST ---

This bug is automatically being proposed for the current release of Red Hat Gluster Storage 3 under active development, by setting the release flag 'rhgs‑3.2.0' to '?'. 

If this bug should be proposed for a different release, please manually change the proposed release flag.

--- Additional comment from Worker Ant on 2016-12-20 08:09:49 EST ---

REVIEW: http://review.gluster.org/16213 (common-ha: Correct the VIP assigned to the new node added) posted (#2) for review on master by soumya k (skoduri)

Comment 1 Worker Ant 2016-12-22 07:23:18 UTC
REVIEW: http://review.gluster.org/16269 (common-ha: Correct the VIP assigned to the new node added) posted (#1) for review on release-3.9 by soumya k (skoduri)

Comment 2 Worker Ant 2016-12-22 07:24:47 UTC
REVIEW: http://review.gluster.org/16269 (common-ha: Correct the VIP assigned to the new node added) posted (#2) for review on release-3.9 by soumya k (skoduri)

Comment 3 Worker Ant 2016-12-22 15:50:06 UTC
COMMIT: http://review.gluster.org/16269 committed in release-3.9 by Kaleb KEITHLEY (kkeithle) 
------
commit 5a65416b14da7b06aff9319f39c5c4f4e7c884fb
Author: Soumya Koduri <skoduri>
Date:   Tue Dec 20 18:22:02 2016 +0530

    common-ha: Correct the VIP assigned to the new node added
    
    There is a regression introduced with patch#16115. An incorrect
    VIP gets assigned to the new node being added to the cluster.
    This patch fixes the same.
    
    This is backport of below mainline patch:
    
    http://review.gluster.org/16213
    
    > Change-Id: I468c7d16bf7e4efa04692db83b1c5ee58fbb7d5f
    > BUG: 1406410
    > Signed-off-by: Soumya Koduri <skoduri>
    
    Change-Id: Iccac83720280d823b36c1e47194b2e17226c91db
    BUG: 1408110
    Signed-off-by: Soumya Koduri <skoduri>
    Reviewed-on: http://review.gluster.org/16269
    Reviewed-by: jiffin tony Thottan <jthottan>
    Smoke: Gluster Build System <jenkins.org>
    Reviewed-by: Kaleb KEITHLEY <kkeithle>
    NetBSD-regression: NetBSD Build System <jenkins.org>
    CentOS-regression: Gluster Build System <jenkins.org>

Comment 4 Kaushal 2017-03-08 10:23:21 UTC
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.9.1, please open a new bug report.

glusterfs-3.9.1 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution.

[1] http://lists.gluster.org/pipermail/gluster-users/2017-January/029725.html
[2] https://www.gluster.org/pipermail/gluster-users/


Note You need to log in before you can comment on or make changes to this bug.