Bug 1258833
Summary: | Data Tiering: Disallow attach tier on a volume where any rebalance process is in progress to avoid deadlock(like remove brick commit pending etc) | |||
---|---|---|---|---|
Product: | [Community] GlusterFS | Reporter: | Nag Pavan Chilakam <nchilaka> | |
Component: | tiering | Assignee: | Mohammed Rafi KC <rkavunga> | |
Status: | CLOSED CURRENTRELEASE | QA Contact: | bugs <bugs> | |
Severity: | urgent | Docs Contact: | ||
Priority: | urgent | |||
Version: | 3.7.5 | CC: | bugs, dlambrig, rkavunga, vagarwal | |
Target Milestone: | --- | Keywords: | Triaged | |
Target Release: | --- | |||
Hardware: | Unspecified | |||
OS: | Unspecified | |||
Whiteboard: | ||||
Fixed In Version: | glusterfs-3.7.6 | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | ||
Clone Of: | ||||
: | 1261819 1271725 (view as bug list) | Environment: | ||
Last Closed: | 2015-11-17 05:57:59 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | ||||
Bug Blocks: | 1260923, 1261819, 1271725, 1275914 |
Description
Nag Pavan Chilakam
2015-09-01 11:36:50 UTC
Workaround:
==========
>do a detach tier commit forcefully
>do a remove brick commit forcefully(though the remove brick operation doesnt show up anymore in the vol status or rebalance status
>reattach the tier
[root@nag-manual-node1 glusterfs]# gluster v detach-tier rebal commit
volume detach-tier commit: failed: Brick 10.70.46.84:/rhs/brick4/rebalhot is not decommissioned. Use start or force option
[root@nag-manual-node1 glusterfs]# gluster v detach-tier rebal commit force
volume detach-tier commit force: success
[root@nag-manual-node1 glusterfs]# gluster v info rebal
Volume Name: rebal
Type: Distribute
Volume ID: 3e272970-b319-4a35-a8cd-6845190761ee
Status: Started
Number of Bricks: 3
Transport-type: tcp
Bricks:
Brick1: 10.70.46.84:/rhs/brick1/rebal
Brick2: 10.70.46.36:/rhs/brick1/rebal
Brick3: 10.70.46.36:/rhs/brick2/rebal
Options Reconfigured:
performance.readdir-ahead: on
[root@nag-manual-node1 glusterfs]# gluster v status rebal
Status of volume: rebal
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick 10.70.46.84:/rhs/brick1/rebal 49187 0 Y 7849
Brick 10.70.46.36:/rhs/brick1/rebal 49186 0 Y 32414
Brick 10.70.46.36:/rhs/brick2/rebal 49187 0 Y 32432
NFS Server on localhost 2049 0 Y 8455
NFS Server on 10.70.46.36 2049 0 Y 402
Task Status of Volume rebal
------------------------------------------------------------------------------
There are no active volume tasks
[root@nag-manual-node1 glusterfs]# gluster v rebal rebal status
Node Rebalanced-files size scanned failures skipped status run time in secs
--------- ----------- ----------- ----------- ----------- ----------- ------------ --------------
volume rebalance: rebal: success:
[root@nag-manual-node1 glusterfs]# gluster v remove-brick rebal 10.70.46.36:/rhs/brick2/rebal commit
Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y
volume remove-brick commit: success
Check the removed bricks to ensure all files are migrated.
If files with data are found on the brick path, copy them via a gluster mount point before re-purposing the removed brick.
Nag, Thanks for catching this bug. Good work This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.7.6, please open a new bug report. glusterfs-3.7.6 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution. [1] http://www.gluster.org/pipermail/gluster-users/2015-November/024359.html [2] http://thread.gmane.org/gmane.comp.file-systems.gluster.user |