Description of problem: ======================== When we disable the heal deamon(to test client side heal), and if there are heal pendings, the root dir shows as "Possibly undergoing heal" always until the heal is triggered and clears the entries I see the problem only with root directory [root@dhcp35-45 ~]# [root@dhcp35-45 ~]# gluster v heal rep2 info Brick 10.70.35.45:/rhs/brick2/rep2 Status: Connected Number of entries: 0 Brick 10.70.35.130:/rhs/brick2/rep2 /zen1 /vex / - Possibly undergoing heal Status: Connected Number of entries: 3 Version-Release number of selected component (if applicable): ===== 3.8.4-28 How reproducible: always Steps to Reproduce: 1.create a 1x2 volume, disable heal deamon 2.create a zerobyte file f1 under root mount 3.kill b1 4. append data to f1 and create a new file f2 5. bring b1 up 6. check heal info Actual results: ========== root will show as possibly undergoing heal, even though there is no heal undergoing Expected results: ============ should not show wrong info
Update: ======== Build used: glusterfs-server-3.12.2-6.el7rhgs.x86_64 Verified below scenarios for both 1 * 2 and 2 * 3 1. create volume, disable heal deamon 2. create a zerobyte file f1 under root mount 3. kill b1 4. append data to f1 and create a new file f2 5. bring b1 up 6. check heal info Didn't see any "Possibly undergoing heal" messages for root dir # gluster vol heal 12 info Brick 10.70.35.61:/bricks/brick1/b0 Status: Connected Number of entries: 0 Brick 10.70.35.174:/bricks/brick1/b1 /f1 /f2 / Status: Connected Number of entries: 3 # # gluster vol heal 23 info Brick 10.70.35.61:/bricks/brick0/testvol_distributed-replicated_brick0 Status: Connected Number of entries: 0 Brick 10.70.35.174:/bricks/brick0/testvol_distributed-replicated_brick1 Status: Connected Number of entries: 0 Brick 10.70.35.17:/bricks/brick0/testvol_distributed-replicated_brick2 Status: Connected Number of entries: 0 Brick 10.70.35.163:/bricks/brick0/testvol_distributed-replicated_brick3 Status: Connected Number of entries: 0 Brick 10.70.35.136:/bricks/brick0/testvol_distributed-replicated_brick4 /f1 /f2 / Status: Connected Number of entries: 3 Brick 10.70.35.214:/bricks/brick0/testvol_distributed-replicated_brick5 /f1 /f2 / Status: Connected Number of entries: 3 # Changing status to verified
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2018:2607