Bug 1647074 - when peer detach is issued, throw a warning to remount volumes using other cluster IPs before proceeding
Summary: when peer detach is issued, throw a warning to remount volumes using other cl...
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: GlusterFS
Classification: Community
Component: glusterd
Version: mainline
Hardware: Unspecified
OS: Unspecified
low
urgent
Target Milestone: ---
Assignee: bugs@gluster.org
QA Contact:
URL:
Whiteboard:
Depends On: 1639568
Blocks:
TreeView+ depends on / blocked
 
Reported: 2018-11-06 16:04 UTC by Atin Mukherjee
Modified: 2019-03-25 16:31 UTC (History)
7 users (show)

Fixed In Version: glusterfs-6.0
Clone Of: 1639568
Environment:
Last Closed: 2019-03-25 16:31:51 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Gluster.org Gerrit 21572 0 None Merged cli: add a warning/confirmation message in peer detach code path 2018-11-09 07:19:25 UTC

Comment 1 Atin Mukherjee 2018-11-06 16:08:54 UTC
Description of problem:
-------------------
if a volume is mounted using a server IP, and if the server is being detached from the trusted pool, it can lead to data inconsistencies as explained in BZ#1639566

To avoid this, we must throw a warning prompt, when peer detach is issued , like below:
"first make sure that all clients mounted using the node being detached, are remounted using the ip of another server in the trusted pool, to avoid inconsistencies", if you have done that then only proceed{y/n}


Version-Release number of selected component (if applicable):
=============
mainline

Comment 2 Worker Ant 2018-11-06 16:12:06 UTC
REVIEW: https://review.gluster.org/21572 (glusterd: add a warning/confirmation message in peer detach code path) posted (#1) for review on master by Atin Mukherjee

Comment 3 Worker Ant 2018-11-09 07:19:15 UTC
REVIEW: https://review.gluster.org/21572 (cli: add a warning/confirmation message in peer detach code path) posted (#5) for review on master by Atin Mukherjee

Comment 4 Shyamsundar 2019-03-25 16:31:51 UTC
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-6.0, please open a new bug report.

glusterfs-6.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution.

[1] https://lists.gluster.org/pipermail/announce/2019-March/000120.html
[2] https://www.gluster.org/pipermail/gluster-users/


Note You need to log in before you can comment on or make changes to this bug.