Bug 1536024 - Rebalance process is behaving differently for AFR and EC volume.
Summary: Rebalance process is behaving differently for AFR and EC volume.
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: GlusterFS
Classification: Community
Component: replicate
Version: mainline
Hardware: x86_64
OS: Linux
unspecified
high
Target Milestone: ---
Assignee: Ravishankar N
QA Contact:
URL:
Whiteboard:
Depends On: 1408418
Blocks:
TreeView+ depends on / blocked
 
Reported: 2018-01-18 13:36 UTC by Ravishankar N
Modified: 2018-06-20 17:57 UTC (History)
6 users (show)

Fixed In Version: glusterfs-v4.1.0
Doc Type: If docs needed, set a value
Doc Text:
Clone Of: 1408418
Environment:
Last Closed: 2018-06-20 17:57:42 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Embargoed:


Attachments (Terms of Use)

Description Ravishankar N 2018-01-18 13:36:48 UTC
+++ This bug was initially created as a clone of Bug #1408418 +++

Description of problem:
=======================
Rebalance process is behaving differently for AFR and EC volume that is,
for EC volume, during rebalance,it is doing heal operation BUT for AFR volume during rebalance, heal is not happening because heal is disabled, here is the bug fix which disabled the heal during rebalance for AFR volume type - 

https://bugzilla.redhat.com/show_bug.cgi?id=808977


Version-Release number of selected component (if applicable):
==============================================================
glusterfs-3.8.4-9.el6rhs.x86_64


How reproducible:
=================
Always


Steps to Reproduce:
===================
ON EC volume:

1. Have EC  volume and fuse mount it
2. Make one brick down.
3. Write enough data from fuse mount
4. Add one more sub volume to the volume
5. bring up the offline brick using volume force option
6. trigger the rebalance manually
7. Check the rebalance logs // you will see any heal related things in rebalance log


ON AFR (Dis-Rep - 2*2) volume:

1. Have AFR Dis-Rep  volume and fuse mount it
2. Make one brick down.
3. Write enough data from fuse mount
4. Add one more sub volume to the volume
5. bring up the offline brick using volume force option
6. trigger the rebalance manually
7. Check the rebalance logs  // you won't see any heal related info in rebalance log



Actual results:
===============
Rebalance process is behaving  differently  for AFR and EC volume.



Expected results:
=================
We expect same rebalance behaviour for EC and AFR volume type.
and we have to find out why heal is disabled during rebalance for AFR volume in the bug fix - 
https://bugzilla.redhat.com/show_bug.cgi?id=808977




--- Additional comment from nchilaka on 2016-12-23 07:16:02 EST ---

Also, one example of the difference in behavior:
1)in a distrep vol say 2x2 , if a brick is down and the user tries to add a new set of bricks, the add-brick fails, saying bricks are down
But the same passes on an ec-vol

-
--- Additional comment from Ravishankar N on 2017-08-28 05:25:34 EDT ---

Moving this to EC as 808977 has removed self-heal from rebalance path intentionally. I would think this needs to be done for EC too. Feel free to move back to AFR if the decision is to allow AFR also to do heals during rebalance.

--- Additional comment from Ashish Pandey on 2018-01-18 05:56:46 EST ---


I think this should not be an issue.

If something is behaving differently for AFR and EC , just that can not be a criteria for a bug until something is being messed UP.

If it needs to be fixed at all then it should be in AFR. Talked to Pranith also on this and he was saying that it should be enabled.
It was disabled because the AFR was not that stable at that time.

---
Ashish

Comment 1 Worker Ant 2018-01-18 13:38:27 UTC
REVIEW: https://review.gluster.org/19233 (glusterd: enable self-heal in daemons) posted (#1) for review on master by Ravishankar N

Comment 2 Worker Ant 2018-05-04 04:52:38 UTC
COMMIT: https://review.gluster.org/19233 committed in master by "Amar Tumballi" <amarts> with a commit message- glusterd: enable self-heal in daemons

..like rebalance, quota and tier because that seems to be the consensus
(see BZ).

Change-Id: I912336a12f4e33ea4ec55f804df403fab0dc89fc
BUG: 1536024
Signed-off-by: Ravishankar N <ravishankar>

Comment 3 Shyamsundar 2018-06-20 17:57:42 UTC
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-v4.1.0, please open a new bug report.

glusterfs-v4.1.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution.

[1] http://lists.gluster.org/pipermail/announce/2018-June/000102.html
[2] https://www.gluster.org/pipermail/gluster-users/


Note You need to log in before you can comment on or make changes to this bug.