Description of problem: When one of the replica brick is down and do some file operation, gluster vol restart or brick process restart doesn't heal the files which needs to be healed. Version-Release number of selected component (if applicable): glusterfs-3.7.1-7.el6rhs.x86_64 How reproducible: 100% Steps to Reproduce: 1. Create 2*2 distribute replicate volume 2. Do fuse mount 3. create some files on mount point 4. kill one of the replica brick 5. rename the file from the mount point 6. check gluster v heal <volname> info 7. restart the volume or restart the brick process Actual results: Files are not healed Expected results: volume restart or brick process restart should heal the files which need to be healed Additional info: Volume Name: vol0 Type: Distributed-Replicate Volume ID: 53c64343-c537-428c-b7b7-a45f198c42a0 Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: 10.70.33.214:/rhs/brick1/b001 Brick2: 10.70.33.219:/rhs/brick1/b002 Brick3: 10.70.33.225:/rhs/brick1/b003 Brick4: 10.70.44.13:/rhs/brick1/b004 Options Reconfigured: performance.readdir-ahead: on features.uss: enable features.quota: on features.inode-quota: on features.quota-deem-statfs: on server.allow-insecure: on features.barrier: disable cluster.enable-shared-storage: enable
Patch posted upstream - http://review.gluster.org/11912
Downstream patch: https://code.engineering.redhat.com/gerrit/#/c/56024/
Verified as fixed in glusterfs-3.7.1-13.el7rhgs.x86_64. Heals are now happening as soon as volume is started with force.
Ravishankar, Made a few minor edits to the doc text. Could you review and sign-off?
Looks okay to me.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHSA-2015-1845.html