+++ This bug was initially created as a clone of Bug #1515163 +++
Description of problem:
regression failure observed at https://build.gluster.org/job/centos6-regression/7534/console
--- Additional comment from Worker Ant on 2018-01-11 23:55:07 EST ---
REVIEW: https://review.gluster.org/19185 (tests: check volume status for shd being up) posted (#1) for review on master by Ravishankar N
--- Additional comment from Worker Ant on 2018-01-12 00:56:47 EST ---
COMMIT: https://review.gluster.org/19185 committed in master by \"Ravishankar N\" <firstname.lastname@example.org> with a commit message- tests: check volume status for shd being up
so that glusterd is also aware that shd is up and running.
While not reproducible locally, on the jenkins slaves, 'gluster vol heal patchy'
fails with "Self-heal daemon is not running. Check self-heal daemon log file.",
while infact the afr_child_up_status_in_shd() checks before that passed. In the
shd log also, I see the shd being up and connected to at least one brick before
the heal is launched.
Signed-off-by: Ravishankar N <email@example.com>
--- Additional comment from Shyamsundar on 2018-03-15 07:21:35 EDT ---
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-4.0.0, please open a new bug report.
glusterfs-4.0.0 has been announced on the Gluster mailinglists , packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist  and the update infrastructure for your distribution.
Build Used: glusterfs-3.12.2-14.el7rhgs.x86_64
1) Create 1 * 3 volume and mount
2) Disable self heal and all client side heals
3) on mount point, continously write to a file using dd ( 4GB )
4) bring down brick0
5) rename the file from another session
6) after 2 min, bring up b0
7) enable self heal daemon
8) check self heal daemon is up and running and trigger volume heal
9) wait for heal to complete
10) kill the dd if it still running
11) calculate md5sum on both the bricks
> md5sum for the file is same on all the bricks as well from mountpoint.
# md5sum /bricks/brick0/b0/file1_rename
# md5sum /bricks/brick0/b1/file1_rename
# md5sum /bricks/brick0/b2/file1_rename
# md5sum file1_rename
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.
For information on the advisory, and where to find the updated
files, follow the link below.
If the solution does not work for you, open a new bug report.