Bug 1517260
| Summary: | Volume wrong size | ||||||
|---|---|---|---|---|---|---|---|
| Product: | [Community] GlusterFS | Reporter: | david <davidpsv17> | ||||
| Component: | glusterd | Assignee: | Amar Tumballi <atumball> | ||||
| Status: | CLOSED CURRENTRELEASE | QA Contact: | |||||
| Severity: | high | Docs Contact: | |||||
| Priority: | medium | ||||||
| Version: | 3.12 | CC: | amukherj, bugs, davidpsv17, igal.baevsky, nbalacha, rgowdapp | ||||
| Target Milestone: | --- | ||||||
| Target Release: | --- | ||||||
| Hardware: | x86_64 | ||||||
| OS: | Linux | ||||||
| Whiteboard: | |||||||
| Fixed In Version: | glusterfs-v4.1.0 | Doc Type: | If docs needed, set a value | ||||
| Doc Text: | Story Points: | --- | |||||
| Clone Of: | |||||||
| : | 1541830 1541880 (view as bug list) | Environment: | |||||
| Last Closed: | 2018-04-06 11:06:55 UTC | Type: | Bug | ||||
| Regression: | --- | Mount Type: | --- | ||||
| Documentation: | --- | CRM: | |||||
| Verified Versions: | Category: | --- | |||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||
| Embargoed: | |||||||
| Bug Depends On: | |||||||
| Bug Blocks: | 1541830, 1541880 | ||||||
| Attachments: |
|
||||||
|
Description
david
2017-11-24 12:08:18 UTC
Since the issue looks to be in aggregation of sizes of DHT subvolumes, changing the component to distribute I seem to have this bug as well. I just upgraded from 3.11.3 to 3.12.3. fuse mount from/on node1 32T 16T 15T 53% node1 brick1 20T 18T 2.1T 90% brick2 12T 90M 11T 1% node2 brick1 20T 15T 5.6T 72% brick2 12T 90M 11T 1% I removed my 2 new bricks and resized the lvm/ext4fs instead. The aggregation of total size is correct. (In reply to david from comment #0) > Hi, > I have a volume replica 3 distributed in 3 servers, the 3 severs have the > same version (3.12.3) and the same SO (Ubuntu 16.04.1), each server has > three bricks and I used this command to create the volume: > > gluster volume create volume1 replica 3 transport tcp > ubuntu1:/work/work1/test-storage1 ubuntu1:/work/work2/test-storage2 > ubuntu1:/work/work3/test-storage3 ubuntu2:/work/work1/test-storage1 > ubuntu2:/work/work2/test-storage2 ubuntu2:/work/work3/test-storage3 > ubuntu3:/work/work1/test-storage1 ubuntu3:/work/work2/test-storage2 > ubuntu3:/work/work3/test-storage3 > > Each brick has a size of 22TB,so I understand that if I mount the volume I > should have a partition of 66TB. The problem is once the partition is > mounted, if I check the size, appears that it has only 22TB.. > > I don't know what it's wrong, can anyone help me? > > Thanks!! On a different note - this is not the best way to create a replica 3 volume. All you replicas are on the same host so if any one host goes down you lose access to all the data on that replica set. A better way is: gluster volume create volume1 replica 3 transport tcp ubuntu1:/work/work1/test-storage1 ubuntu2:/work/work1/test-storage1 ubuntu3:/work/work1/test-storage1 ubuntu1:/work/work2/test-storage2 ubuntu2:/work/work2/test-storage2 ubuntu3:/work/work2/test-storage2 ubuntu1:/work/work3/test-storage3 ubuntu2:/work/work3/test-storage3 ubuntu3:/work/work3/test-storage3 Can you please mount the volume, check the size and send us the mount log? (In reply to Nithya Balachandran from comment #4) > (In reply to david from comment #0) > > Hi, > > I have a volume replica 3 distributed in 3 servers, the 3 severs have the > > same version (3.12.3) and the same SO (Ubuntu 16.04.1), each server has > > three bricks and I used this command to create the volume: > > > > gluster volume create volume1 replica 3 transport tcp > > ubuntu1:/work/work1/test-storage1 ubuntu1:/work/work2/test-storage2 > > ubuntu1:/work/work3/test-storage3 ubuntu2:/work/work1/test-storage1 > > ubuntu2:/work/work2/test-storage2 ubuntu2:/work/work3/test-storage3 > > ubuntu3:/work/work1/test-storage1 ubuntu3:/work/work2/test-storage2 > > ubuntu3:/work/work3/test-storage3 > > > > Each brick has a size of 22TB,so I understand that if I mount the volume I > > should have a partition of 66TB. The problem is once the partition is > > mounted, if I check the size, appears that it has only 22TB.. > > > > I don't know what it's wrong, can anyone help me? > > > > Thanks!! > > On a different note - this is not the best way to create a replica 3 volume. > All you replicas are on the same host so if any one host goes down you lose > access to all the data on that replica set. > > A better way is: > > gluster volume create volume1 replica 3 transport tcp > ubuntu1:/work/work1/test-storage1 ubuntu2:/work/work1/test-storage1 > ubuntu3:/work/work1/test-storage1 ubuntu1:/work/work2/test-storage2 > ubuntu2:/work/work2/test-storage2 ubuntu3:/work/work2/test-storage2 > ubuntu1:/work/work3/test-storage3 ubuntu2:/work/work3/test-storage3 > ubuntu3:/work/work3/test-storage3 Hi @Nithya! This is the same configuration that i have, isn't it? but in other order.. Thanks! (In reply to david from comment #6) > (In reply to Nithya Balachandran from comment #4) > > (In reply to david from comment #0) > > > Hi, > > > I have a volume replica 3 distributed in 3 servers, the 3 severs have the > > > same version (3.12.3) and the same SO (Ubuntu 16.04.1), each server has > > > three bricks and I used this command to create the volume: > > > > > > gluster volume create volume1 replica 3 transport tcp > > > ubuntu1:/work/work1/test-storage1 ubuntu1:/work/work2/test-storage2 > > > ubuntu1:/work/work3/test-storage3 ubuntu2:/work/work1/test-storage1 > > > ubuntu2:/work/work2/test-storage2 ubuntu2:/work/work3/test-storage3 > > > ubuntu3:/work/work1/test-storage1 ubuntu3:/work/work2/test-storage2 > > > ubuntu3:/work/work3/test-storage3 > > > > > > Each brick has a size of 22TB,so I understand that if I mount the volume I > > > should have a partition of 66TB. The problem is once the partition is > > > mounted, if I check the size, appears that it has only 22TB.. > > > > > > I don't know what it's wrong, can anyone help me? > > > > > > Thanks!! > > > > On a different note - this is not the best way to create a replica 3 volume. > > All you replicas are on the same host so if any one host goes down you lose > > access to all the data on that replica set. > > > > A better way is: > > > > gluster volume create volume1 replica 3 transport tcp > > ubuntu1:/work/work1/test-storage1 ubuntu2:/work/work1/test-storage1 > > ubuntu3:/work/work1/test-storage1 ubuntu1:/work/work2/test-storage2 > > ubuntu2:/work/work2/test-storage2 ubuntu3:/work/work2/test-storage2 > > ubuntu1:/work/work3/test-storage3 ubuntu2:/work/work3/test-storage3 > > ubuntu3:/work/work3/test-storage3 > > Hi @Nithya! > > This is the same configuration that i have, isn't it? but in other order.. > Yes, but the ordering is what determines which bricks form a replica set (contain copies of the same file). Ideally, you want to create your volume so that you have each brick of a replica set on a different node so if one node goes down the other 2 bricks on the other 2 nodes can still serve the data. In the volume create command, with a replica value of 'n', each group of n bricks forms a replica set. The first set of n bricks forms one set, then the next n and so on. So in your case, the first 3 bricks passed to the create command form a replica set, then the next 3 and so on. As the first 3 bricks are all on the same node, if ubuntu1 goes down, all the files on that set will be inaccessible. You can confirm this by checking the bricks on ubuntu1 - you should see the same files on all 3 bricks. > Thanks! Yes, sorry @Nithya. I created it how you said, I wrote it wrong.. Thanks (In reply to david from comment #8) > Yes, sorry @Nithya. I created it how you said, I wrote it wrong.. > Thanks Good to hear. :) Coming back to the original issue you reported, can you send us: 1. gluster volume info <volname> 2. mount the volume, check the size and send us the mount log Thanks, Nithya 1. output of gluster volume info: Volume Name: volume1 Type: Distributed-Replicate Volume ID: 66153ffa-dfd7-4c1f-966e-093862605b40 Status: Started Snapshot Count: 0 Number of Bricks: 3 x 3 = 9 Transport-type: tcp Bricks: Brick1: ubuntu1:/work/work1/test-storage1 Brick2: ubuntu2:/work/work1/test-storage1 Brick3: ubuntu3:/work/work1/test-storage1 Brick4: ubuntu1:/work/work2/test-storage2 Brick5: ubuntu2:/work/work2/test-storage2 Brick6: ubuntu3:/work/work2/test-storage2 Brick7: ubuntu1:/work/work3/test-storage3 Brick8: ubuntu2:/work/work3/test-storage3 Brick9: ubuntu3:/work/work3/test-storage3 Options Reconfigured: transport.address-family: inet performance.readdir-ahead: on nfs.disable: on performance.quick-read: off performance.io-thread-count: 48 performance.cache-size: 256MB performance.write-behind-window-size: 8MB performance.cache-max-file-size: 2MB performance.read-ahead: off client.event-threads: 4 server.event-threads: 4 cluster.lookup-optimize: on performance.client-io-threads: on cluster.readdir-optimize: on 2. I couldn't see a specific error after mount the volume, but in the syslog there are a lot of errors like: Dec 7 16:37:04 ubuntu1 kernel: [1308096.853903] EXT4-fs error (device sdb1): htree_dirblock_to_tree:986: inode #110821791: block 1773144535: comm glusteriotwr20: bad entry in directory: rec_len is smaller than minimal - offset=0(0), inode=3925999631, rec_len=1, name_len=0 Thanks! Hi David, Sorry for taking so long to get back. Can you please send the client mount log? (this should be in /var/log/glusterfs/<hyphenated path of mount point>.log ) after retrying the operation? Hi David, Do you still see the problem? We may have found the reason. Can you send me the contents of the /var/lib/glusterd/volume1 directory on any one of the server nodes? Regards, Nithya Hi @Nithya, ls /var/lib/glusterd/vols/volume1 bricks cksum volume1.ubuntu1.work-work1-test-storage1.vol volume1.ubuntu1.work-work2-test-storage2.vol volume1.ubuntu1.work-work3-test-storage3.vol volume1.ubuntu2.work-work1-test-storage1.vol volume1.ubuntu2.work-work2-test-storage2.vol volume1.ubuntu2.work-work3-test-storage3.vol volume1.ubuntu3.work-work1-test-storage1.vol volume1.ubuntu3.work-work2-test-storage2.vol volume1.ubuntu3.work-work3-test-storage3.vol volume1-rebalance.vol volume1.tcp-fuse.vol info node_state.info quota.cksum quota.conf rebalance run snapd.info trusted-volume1.tcp-fuse.vol Thanks! Apologies, I should have made myself clearer. I need to see the contents of the files in that directory. Specifically: volume1.ubuntu1.work-work1-test-storage1.vol volume1.ubuntu1.work-work2-test-storage2.vol volume1.ubuntu1.work-work3-test-storage3.vol volume1.ubuntu2.work-work1-test-storage1.vol volume1.ubuntu2.work-work2-test-storage2.vol volume1.ubuntu2.work-work3-test-storage3.vol volume1.ubuntu3.work-work1-test-storage1.vol volume1.ubuntu3.work-work2-test-storage2.vol volume1.ubuntu3.work-work3-test-storage3.vol Can you also please send the /var/lib/glusterd/glusterd.info file from all 3 nodes? Of course,
glusterd.info ubuntu1:
UUID=a3597de1-d634-4e4f-80c4-186180071298
operating-version=31000
glusterd.info ubuntu2:
UUID=74c61527-88ed-4c0c-8a9f-3aafb1f60c3c
operating-version=31000
glusterd.info ubuntu3:
UUID=e5b2240a-9f81-49c9-87f3-0bc4e9390661
operating-version=31000
The other files are quite big, for example this is the content of volume1.ubuntu1.work-work1-test-storage1.vol:
volume volume1-posix
type storage/posix
option shared-brick-count 3
option volume-id 66153ffa-dfd7-4c1f-966e-093862605b40
option directory /work/work1/test-storage1
end-volume
volume volume1-trash
type features/trash
option trash-internal-op off
option brick-path /work/work1/test-storage1
option trash-dir .trashcan
subvolumes volume1-posix
end-volume
volume volume1-changetimerecorder
type features/changetimerecorder
option sql-db-wal-autocheckpoint 25000
option sql-db-cachesize 12500
option ctr-record-metadata-heat off
option record-counters off
option ctr-enabled off
option record-entry on
option ctr_lookupheal_inode_timeout 300
option ctr_lookupheal_link_timeout 300
option ctr_link_consistency off
option record-exit off
option db-path /work/work1/test-storage1/.glusterfs/
option db-name test-storage1.db
option hot-brick off
option db-type sqlite3
subvolumes volume1-trash
end-volume
volume volume1-changelog
type features/changelog
option changelog-barrier-timeout 120
option changelog-dir /work/work1/test-storage1/.glusterfs/changelogs
option changelog-brick /work/work1/test-storage1
subvolumes volume1-changetimerecorder
end-volume
volume volume1-bitrot-stub
type features/bitrot-stub
option bitrot disable
option export /work/work1/test-storage1
subvolumes volume1-changelog
end-volume
volume volume1-access-control
type features/access-control
subvolumes volume1-bitrot-stub
end-volume
volume volume1-locks
type features/locks
subvolumes volume1-access-control
end-volume
volume volume1-worm
type features/worm
option worm-file-level off
option worm off
subvolumes volume1-locks
end-volume
volume volume1-read-only
type features/read-only
option read-only off
subvolumes volume1-worm
end-volume
volume volume1-leases
type features/leases
option leases off
subvolumes volume1-read-only
end-volume
volume volume1-upcall
type features/upcall
option cache-invalidation off
subvolumes volume1-leases
end-volume
volume volume1-io-threads
type performance/io-threads
option thread-count 48
subvolumes volume1-upcall
end-volume
volume volume1-selinux
type features/selinux
option selinux on
subvolumes volume1-io-threads
end-volume
volume volume1-marker
type features/marker
option inode-quota off
option quota off
option gsync-force-xtime off
option xtime off
option quota-version 0
option timestamp-file /var/lib/glusterd/vols/volume1/marker.tstamp
option volume-uuid 66153ffa-dfd7-4c1f-966e-093862605b40
subvolumes volume1-selinux
end-volume
volume volume1-barrier
type features/barrier
option barrier-timeout 120
option barrier disable
subvolumes volume1-marker
end-volume
volume volume1-index
type features/index
option xattrop-pending-watchlist trusted.afr.volume1-
option xattrop-dirty-watchlist trusted.afr.dirty
option index-base /work/work1/test-storage1/.glusterfs/indices
subvolumes volume1-barrier
end-volume
volume volume1-quota
type features/quota
option deem-statfs off
option server-quota off
option volume-uuid volume1
subvolumes volume1-index
end-volume
volume volume1-io-stats
type debug/io-stats
option count-fop-hits off
option latency-measurement off
option log-level INFO
option unique-id /work/work1/test-storage1
subvolumes volume1-quota
end-volume
volume /work/work1/test-storage1
type performance/decompounder
subvolumes volume1-io-stats
end-volume
volume volume1-server
type protocol/server
option transport.listen-backlog 10
option transport.socket.keepalive-count 9
option transport.socket.keepalive-interval 2
option transport.socket.keepalive-time 20
option transport.tcp-user-timeout 0
option event-threads 4
option transport.socket.keepalive 1
option auth.addr./work/work1/test-storage1.allow *
option auth-path /work/work1/test-storage1
option auth.login.37801a5b-f8e9-487d-9189-e8ad0f0855fa.password 4170e1ae-f028-449e-b699-06ddc61a3853
option auth.login./work/work1/test-storage1.allow 37801a5b-f8e9-487d-9189-e8ad0f0855fa
option transport.address-family inet
option transport-type tcp
subvolumes /work/work1/test-storage1
end-volume
Do you want the others?
Thanks!
Thanks David.
I wanted to see the value of the shared-brick-count. Can you confirm that it is 3 for all the volume1.ubuntu*.work-work*-test-storage*.vol files listed above?
volume volume1-posix
type storage/posix
option shared-brick-count 3 <--- *This*
option volume-id 66153ffa-dfd7-4c1f-966e-093862605b40
option directory /work/work1/test-storage1
end-volume
The shared-brick-count is a count of how many bricks share a file system and the disk space is divided accordingly. For instance, if I have a filesystem mounted at /data with 100GB space, and I create 3 subdirs inside (/data/brick1, /data/brick2, /data/brick2)to use as bricks, I really only have 100GB so the df should also only return 100GB (Patch https://review.gluster.org/#/c/17618/ introduced this change. In earlier releases, it would have returned 300GB which is incorrect).
So if all bricks of volume1 on each node are on different filesystems, this is a bug. Otherwise, it is working as expected.
Sure, grep "shared-brick-count" * volume1.ubuntu1.work-work1-test-storage1.vol: option shared-brick-count 3 volume1.ubuntu1.work-work2-test-storage2.vol: option shared-brick-count 3 volume1.ubuntu1.work-work3-test-storage3.vol: option shared-brick-count 3 volume1.ubuntu2.work-work1-test-storage1.vol: option shared-brick-count 0 volume1.ubuntu2.work-work2-test-storage2.vol: option shared-brick-count 0 volume1.ubuntu2.work-work3-test-storage3.vol: option shared-brick-count 0 volume1.ubuntu3.work-work1-test-storage1.vol: option shared-brick-count 0 volume1.ubuntu3.work-work2-test-storage2.vol: option shared-brick-count 0 volume1.ubuntu3.work-work3-test-storage3.vol: option shared-brick-count 0 Thanks Nithya Hi David, Sorry for the back and forth on this. Can you try the following on any one node? gluster v set volume1 cluster.min-free-inodes 6% And check if the df output and the *.vol files above remain the same post that? Hi David, We don't know why this is happening yet but we have a workaround until we fix this. On every node in the cluster (ubuntu1, ubuntu2 and ubuntu3): 1. Copy the shared-brick-count.sh file to /usr/lib*/glusterfs/3.12.3/filter/. (You might need to create the filter directory in this path.) 2. Give the file execute permissions. For example, on my system (the version is different): [root@rhgsserver1 dir1]# cd /usr/lib/glusterfs/3.12.5/ [root@rhgsserver1 3.12.5]# ll total 4.0K drwxr-xr-x. 2 root root 64 Feb 1 08:56 auth drwxr-xr-x. 2 root root 34 Feb 1 09:12 filter drwxr-xr-x. 2 root root 66 Feb 1 08:55 rpc-transport drwxr-xr-x. 13 root root 4.0K Feb 1 08:57 xlator [root@rhgsserver1 3.12.5]# cd filter [root@rhgsserver1 filter]# pwd /usr/lib/glusterfs/3.12.5/filter [root@rhgsserver1 filter]# ll total 4 -rwxr-xr-x. 1 root root 95 Feb 1 09:12 shared-brick-count.sh On any one node, run: gluster v set volume1 cluster.min-free-inodes 6% This should regenerate the .vol files and set the value of option shared-brick-count to 1. Please check if the .vol files to confirm this. You do not need to restart the volume. Let me know if this solves the problem. Created attachment 1389314 [details]
shared-brick-count.sh Glusterfs Filter file
See http://docs.gluster.org/en/latest/Administrator%20Guide/GlusterFS%20Filter/ for more details. Hi Nithya! this fixed my problem! Now I can see the 66T: ds2-nl2:/volume1 fuse.glusterfs 66T 9.6T 53T 16% /volume1 And this is the output of *.vol: grep "shared-brick-count" * volume1.ubuntu1.work-work1-test-storage1.vol: option shared-brick-count 1 volume1.ubuntu1.work-work2-test-storage2.vol: option shared-brick-count 1 volume1.ubuntu1.work-work3-test-storage3.vol: option shared-brick-count 1 volume1.ubuntu2.work-work1-test-storage1.vol: option shared-brick-count 1 volume1.ubuntu2.work-work2-test-storage2.vol: option shared-brick-count 1 volume1.ubuntu2.work-work3-test-storage3.vol: option shared-brick-count 1 volume1.ubuntu3.work-work1-test-storage1.vol: option shared-brick-count 1 volume1.ubuntu3.work-work2-test-storage2.vol: option shared-brick-count 1 volume1.ubuntu3.work-work3-test-storage3.vol: option shared-brick-count 1 thank you very much! David, glad that Nithya's scripts helped you to get to proper state.
I wanted to know if on your backend on ubuntu{1,2,3}, all the bricks /work/work{1,2,3} are different partitions? Can you share output of 'df -h /work/work{1,2,3} && stat /work/work{1,2,3}' ? That should help me to resolve issue faster.
Thanks,
Yes, there are different partitions: df -h | grep work /dev/sdc1 22T 3.2T 18T 16% /work/work2 /dev/sdd1 22T 3.2T 18T 16% /work/work3 /dev/sdb1 22T 3.5T 18T 17% /work/work1 stat /work/work1/ File: '/work/work1/' Size: 4096 Blocks: 8 IO Block: 4096 directory Device: 811h/2065d Inode: 2 Links: 6 Access: (0755/drwxr-xr-x) Uid: ( 0/ root) Gid: ( 0/ root) Access: 2017-07-19 08:31:20.932837957 +0200 Modify: 2018-01-31 12:00:43.291297677 +0100 Change: 2018-01-31 12:00:43.291297677 +0100 Birth: - stat /work/work2 File: '/work/work2' Size: 4096 Blocks: 8 IO Block: 4096 directory Device: 821h/2081d Inode: 2 Links: 6 Access: (0755/drwxr-xr-x) Uid: ( 0/ root) Gid: ( 0/ root) Access: 2017-07-19 08:31:55.444838468 +0200 Modify: 2018-01-30 09:14:04.692589116 +0100 Change: 2018-01-30 09:14:04.692589116 +0100 Birth: - stat /work/work3 File: '/work/work3' Size: 4096 Blocks: 8 IO Block: 4096 directory Device: 831h/2097d Inode: 2 Links: 7 Access: (0755/drwxr-xr-x) Uid: ( 0/ root) Gid: ( 0/ root) Access: 2017-07-19 08:31:51.936838416 +0200 Modify: 2018-02-02 09:59:44.940879714 +0100 Change: 2018-02-02 09:59:44.940879714 +0100 Birth: - Thanks! REVIEW: https://review.gluster.org/19464 (glusterd: shared-brick-count should consider st_dev details) posted (#1) for review on master by Amar Tumballi Hi David, One last question to validate my theory. What is the filesystem of your backend bricks ? (ie, /work/work{1,2,3})
the filesystem is ext4 for all the partitions. Regards REVIEW: https://review.gluster.org/19484 (glusterd/store: handle the case of fsid being set to 0) posted (#1) for review on master by Amar Tumballi This holds good for master too. (hence change in version). COMMIT: https://review.gluster.org/19484 committed in master by "Atin Mukherjee" <amukherj> with a commit message- glusterd/store: handle the case of fsid being set to 0 Generally this would happen when a system gets upgraded from an version which doesn't have fsid details, to a version with fsid values. Without this change, after upgrade, people would see reduced 'df ' output, causing lot of confusions. Debugging Credits: Nithya B <nbalacha> Change-Id: Id718127ddfb69553b32770b25021290bd0e7c49a BUG: 1517260 Signed-off-by: Amar Tumballi <amarts> REVIEW: https://review.gluster.org/19493 (glusterd/store: handle the case of fsid being set to 0) posted (#1) for review on release-3.12 by Amar Tumballi Steps to reproduce the problem: 1. On a version of gluster that does not include https://review.gluster.org/#/c/17618/, create a 1x3 volume with each brick on a separate filesystem partition 2. Confirm that the df -h returns the correct size for the mounted volume 3. Upgrade to a gluster version that contains the patch (3.12.3 for instance) 4. Perform a volume set operation to regenerate the volfiles (For e.g., gluster volume set <volume> cluster.min-free-disk 11%) 5. Check the value returned by df -h This should now show an incorrect value. COMMIT: https://review.gluster.org/19493 committed in release-3.12 by "jiffin tony Thottan" <jthottan> with a commit message- glusterd/store: handle the case of fsid being set to 0 Generally this would happen when a system gets upgraded from an version which doesn't have fsid details, to a version with fsid values. Without this change, after upgrade, people would see reduced 'df ' output, causing lot of confusions. Debugging Credits: Nithya B <nbalacha> Change-Id: Id718127ddfb69553b32770b25021290bd0e7c49a BUG: 1517260 Signed-off-by: Amar Tumballi <amarts> This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.12.7, please open a new bug report. glusterfs-3.12.7 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution. [1] http://lists.gluster.org/pipermail/maintainers/2018-March/004303.html [2] https://www.gluster.org/pipermail/gluster-users/ This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-v4.1.0, please open a new bug report. glusterfs-v4.1.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution. [1] http://lists.gluster.org/pipermail/announce/2018-June/000102.html [2] https://www.gluster.org/pipermail/gluster-users/ |