Description of problem: Customer seems to be experiencing the bug described in these bugzilla reports in their RHGS 3.5 cluster: https://github.com/gluster/glusterfs/issues/982 (BZ 1718562) https://github.com/gluster/glusterfs/issues/1046 (BZ 1776152) ...and have used the procedures in [bug:1718562] flock failure (regression) #982 to duplicate the issues reported there. Version-Release number of selected component (if applicable): Clients: glusterfs-6.0-22.el6.x86_64 glusterfs-api-6.0-22.el6.x86_64 glusterfs-client-xlators-6.0-22.el6.x86_64 glusterfs-fuse-6.0-22.el6.x86_64 glusterfs-libs-6.0-22.el6.x86_64 Servers: glusterfs-6.0-30.1.el7rhgs.x86_64 glusterfs-api-6.0-30.1.el7rhgs.x86_64 glusterfs-client-xlators-6.0-30.1.el7rhgs.x86_64 glusterfs-cli-6.0-30.1.el7rhgs.x86_64 glusterfs-events-6.0-30.1.el7rhgs.x86_64 glusterfs-fuse-6.0-30.1.el7rhgs.x86_64 glusterfs-geo-replication-6.0-30.1.el7rhgs.x86_64 glusterfs-libs-6.0-30.1.el7rhgs.x86_64 glusterfs-rdma-6.0-30.1.el7rhgs.x86_64 glusterfs-server-6.0-30.1.el7rhgs.x86_64 gluster-nagios-addons-0.2.10-2.el7rhgs.x86_64 gluster-nagios-common-0.2.4-1.el7rhgs.noarch How reproducible: On demand Steps to Reproduce: Per BZ 1718562 Additional info: Client and server sosreports on support-shell (/cases/02682288) Customer has also supplied: wb_straces.zip - straces of php-cgi processes from an affected application statedump.home.zip - statedumps taken while the wb_straces were running Customer has asked for a hotfix when available and I have set expectations that this is unlikely however it would be good to get this fixed and in an upcoming RHGS release as it is causing ongoing problems for them. Customer is very motivated and would cooperate in any way. Let me know how else I can assist.
Passing on the bug to Susant Palai who deals with the upstream bug.
Steps: 1. create all types of volume 2. mount the brick on two different node 3.prepare same script to do flock on the two clients #!/bin/bash flock_func(){ #!/bin/bash file=/bricks/brick0/test.log touch $file ( flock -xo 200 echo "client1 do something" > $file sleep 1 ) 200>$file } i=1 while [ "1" = "1" ] do flock_func ((i=i+1)) echo $i if [[ $i == 200 ]]; then break fi done 4. waited till 300 iteration ------------------ Additional info [node.example.com]#rpm -qa | grep -i glusterfs glusterfs-6.0-45.el8rhgs.x86_64 glusterfs-fuse-6.0-45.el8rhgs.x86_64 glusterfs-api-6.0-45.el8rhgs.x86_64 glusterfs-selinux-1.0-1.el8rhgs.noarch glusterfs-client-xlators-6.0-45.el8rhgs.x86_64 glusterfs-server-6.0-45.el8rhgs.x86_64 glusterfs-cli-6.0-45.el8rhgs.x86_64 glusterfs-libs-6.0-45.el8rhgs.x86_64 As i don't see any issue while running script till 300 iterations marking this bug as verified
*** Bug 1880271 has been marked as a duplicate of this bug. ***
*** Bug 1852740 has been marked as a duplicate of this bug. ***
*** Bug 1851315 has been marked as a duplicate of this bug. ***
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (glusterfs bug fix and enhancement update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2020:5603