Bug 1614654 - Potential fixes for tests/basic/afr/add-brick-self-heal.t failure
Summary: Potential fixes for tests/basic/afr/add-brick-self-heal.t failure
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: GlusterFS
Classification: Community
Component: tests
Version: mainline
Hardware: Unspecified
OS: Unspecified
medium
unspecified
Target Milestone: ---
Assignee: Ravishankar N
QA Contact:
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2018-08-10 07:14 UTC by Ravishankar N
Modified: 2018-10-23 15:16 UTC (History)
1 user (show)

Fixed In Version: glusterfs-5.0
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2018-10-23 15:16:52 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Embargoed:


Attachments (Terms of Use)

Description Ravishankar N 2018-08-10 07:14:01 UTC
Description of problem:

From the runs captured at https://review.gluster.org/#/c/20637/, I saw that the latest runs where this particular .t failed were at https://build.gluster.org/job/line-coverage/415 and https://build.gluster.org/job/line-coverage/421/.
In both of these runs, there are no gluster 'regression' logs available at https://build.gluster.org/job/line-coverage/<build-number>/artifact. I have raised BZ 1613721 for it.

Jenkins output is:
20:37:43 ./tests/basic/afr/add-brick-self-heal.t .. 
20:37:43 1..34
20:37:43 ok 1, LINENUM:6
20:37:43 ok 2, LINENUM:7
20:37:43 ok 3, LINENUM:8
20:37:43 ok 4, LINENUM:9
20:37:43 ok 5, LINENUM:10
20:37:43 ok 6, LINENUM:11
20:37:43 ok 7, LINENUM:12
20:37:43 ok 8, LINENUM:14
20:37:43 ok 9, LINENUM:15
20:37:43 ok 10, LINENUM:24
20:37:43 ok 11, LINENUM:27
20:37:43 ok 12, LINENUM:30
20:37:43 ok 13, LINENUM:31
20:37:43 ok 14, LINENUM:34
20:37:43 ok 15, LINENUM:35
20:37:43 ok 16, LINENUM:36
20:37:43 ok 17, LINENUM:38
20:37:43 ok 18, LINENUM:39
20:37:43 ok 19, LINENUM:40
20:37:43 ok 20, LINENUM:42
20:37:43 ok 21, LINENUM:43
20:37:43 ok 22, LINENUM:44
20:37:43 ok 23, LINENUM:45
20:37:43 ok 24, LINENUM:46
20:37:43 ok 25, LINENUM:47
20:37:43 not ok 26 Got "10" instead of "^0$", LINENUM:50
20:37:43 FAILED COMMAND: ^0$ get_pending_heal_count patchy
20:37:43 ok 27, LINENUM:53
20:37:43 ok 28, LINENUM:54
20:37:43 not ok 29 , LINENUM:57
20:37:43 FAILED COMMAND: diff /d/backends/patchy0/file1.txt /d/backends/patchy2/file1.txt
20:37:43 not ok 30 Got "" instead of "qwerty", LINENUM:60
20:37:43 FAILED COMMAND: qwerty get_text_xattr user.test /d/backends/patchy2/file5.txt
20:37:43 ok 31, LINENUM:61
20:37:43 ok 32, LINENUM:63
20:37:43 ok 33, LINENUM:64
20:37:43 ok 34, LINENUM:65
20:37:43 Failed 3/34 subtests 
20:37:43 


So self-heal has failed to heal the entries. This could be due to bricks not being up/ shd not able to connect to the bricks/ lock contention and wait for $HEAL_TIMEOUT as described in BZ 1613807.

Hence modifying the .t to check for brick up status and reducing heal timeout to 5 seconds.


Note:
After the above said runs, the test has not failed again, up to the latest revision 8 of https://review.gluster.org/#/c/2063. So sending the patch with whatever can be theoretically fixed in the .t

Comment 1 Worker Ant 2018-08-10 07:17:21 UTC
REVIEW: https://review.gluster.org/20696 (tests: potential fixes for tests/basic/afr/add-brick-self-heal.t) posted (#1) for review on master by Ravishankar N

Comment 2 Worker Ant 2018-08-13 04:56:51 UTC
COMMIT: https://review.gluster.org/20696 committed in master by "Atin Mukherjee" <amukherj> with a commit message- tests: potential fixes for tests/basic/afr/add-brick-self-heal.t

Please see bug description for details.

Change-Id: Ieb6bce6d1d5c4c31f1878dd1a1c3d007d8ff81d5
fixes: bz#1614654
Signed-off-by: Ravishankar N <ravishankar>

Comment 3 Shyamsundar 2018-10-23 15:16:52 UTC
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-5.0, please open a new bug report.

glusterfs-5.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution.

[1] https://lists.gluster.org/pipermail/announce/2018-October/000115.html
[2] https://www.gluster.org/pipermail/gluster-users/


Note You need to log in before you can comment on or make changes to this bug.