Considering that we want all the block files to be written without any caches we are disabling all the perf xlators. We are also enabling some of the options that proved to be good candidates in virt profile. Final list at the moment looks to be: performance.quick-read=off performance.read-ahead=off performance.io-cache=off performance.stat-prefetch=off performance.write-behind=off performance.open-behind=off performance.readdir-ahead=off network.remote-dio=enable cluster.eager-lock=enable cluster.quorum-type=auto cluster.data-self-heal-algorithm=full cluster.locking-scheme=granular cluster.shd-max-threads=8 cluster.shd-wait-qlength=10000 features.shard=on user.cifs=off server.allow-insecure=on
REVIEW: https://review.gluster.org/17403 (extras: Provide group set for gluster-block workloads) posted (#1) for review on release-3.11 by Pranith Kumar Karampuri (pkarampu)
COMMIT: https://review.gluster.org/17403 committed in release-3.11 by Shyamsundar Ranganathan (srangana) ------ commit 5b3e5f1c54fae28c0c89c1b8f769a5777e970956 Author: Pranith Kumar K <pkarampu> Date: Wed May 10 16:26:35 2017 +0530 extras: Provide group set for gluster-block workloads For gluster-block workloads I/O is always with o-direct so it doesn't benefit by any of the perf xlators so disabling all of them to save on memory. performance.quick-read=off performance.read-ahead=off performance.io-cache=off performance.stat-prefetch=off performance.write-behind=off performance.open-behind=off performance.readdir-ahead=off We want the I/O on the file to be with o-direct network.remote-dio=enable Options that are proven to give good performance with VM workloads which is very similar to gluster-block cluster.eager-lock=enable cluster.quorum-type=auto cluster.data-self-heal-algorithm=full cluster.locking-scheme=granular cluster.shd-max-threads=8 cluster.shd-wait-qlength=10000 features.shard=on It is better to turn off things we are not using user.cifs=off It is better to have allow-insecure to be on so that ports that are > 1024 in tcmu-runner are allowed. server.allow-insecure=on >Change-Id: I9a21c824fa42242f02b57569feedd03d9b6f9439 >BUG: 1450010 >Signed-off-by: Pranith Kumar K <pkarampu> >Reviewed-on: https://review.gluster.org/17254 >Smoke: Gluster Build System <jenkins.org> >Reviewed-by: Niels de Vos <ndevos> >CentOS-regression: Gluster Build System <jenkins.org> >NetBSD-regression: NetBSD Build System <jenkins.org> >Reviewed-by: Jeff Darcy <jeff.us> BUG: 1456224 Change-Id: I9a21c824fa42242f02b57569feedd03d9b6f9439 Signed-off-by: Pranith Kumar K <pkarampu> Reviewed-on: https://review.gluster.org/17403 Smoke: Gluster Build System <jenkins.org> NetBSD-regression: NetBSD Build System <jenkins.org> CentOS-regression: Gluster Build System <jenkins.org> Reviewed-by: Shyamsundar Ranganathan <srangana>
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.11.0, please open a new bug report. glusterfs-3.11.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution. [1] http://lists.gluster.org/pipermail/announce/2017-May/000073.html [2] https://www.gluster.org/pipermail/gluster-users/