Description of problem:
Description of problem: We are doing a drive replacement test and have noted that the heal speed for disperse volume is very slow.
I replaced a 4TB drive with a different drive. It has been almost 26 hours since that. I see that only about 453 GB of data has been constructed back.
/dev/sdc 3.7T 3.7T 7.5G 100% /ws/disk1
/dev/sdd 3.7T 3.7T 7.5G 100% /ws/disk2
/dev/sde 3.7T 3.7T 7.6G 100% /ws/disk3
/dev/sdg 3.7T 453G 3.2T 13% /ws/disk5
Version-Release number of selected component (if applicable): glusterfs-3.8.4-18.el7rhgs.x86_64
Additional info: Volume is a distributed-disperse 2 x (4 + 2) = 12 volume, with the following options reconfigured:
Profile info and df will be uploaded to the bug shortly. Directory structure will be given
Version-Release number of selected component (if applicable):
Steps to Reproduce:
have tested replace brick for same data set for about 176GB data(on the brick being replaced) on a 2x(4+2) volume
On 3.3.1-async(184.108.40.206-8) the total time for replace brick to complete (ie heal completion) was 6hrs
on 3.4(220.127.116.11/11) was 4.5 hours
hence ie about 25% improvement.
hence moving the bug to verified
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.
For information on the advisory, and where to find the updated
files, follow the link below.
If the solution does not work for you, open a new bug report.