REVIEW: http://review.gluster.org/14448 (cluster/afr: If possible give errno received from lower xlators) posted (#1) for review on release-3.8 by Pranith Kumar Karampuri (pkarampu)
REVIEW: http://review.gluster.org/14449 (cluster/afr: Refresh inode for inode-write fops in need) posted (#1) for review on release-3.8 by Pranith Kumar Karampuri (pkarampu)
COMMIT: http://review.gluster.org/14448 committed in release-3.8 by Niels de Vos (ndevos) ------ commit 78cfd326e23d07d199ea25cfc6c1330320ae89b0 Author: Pranith Kumar K <pkarampu> Date: Tue May 17 06:38:57 2016 +0530 cluster/afr: If possible give errno received from lower xlators In case of 3 way replication with quorum enabled with sharding, if one bricks is brought down and brought back up sometimes fops fail with EROFS because the mknod of shard file fails with two good nodes with EEXIST. So even when quorum is not met, it makes sense to unwind with the errno returned by lower xlators as much as possible. >Change-Id: Iabd91cd7c270f5dfe6cbd18c50e59c299a331552 >BUG: 1336612 >Signed-off-by: Pranith Kumar K <pkarampu> >Reviewed-on: http://review.gluster.org/14369 >Smoke: Gluster Build System <jenkins.com> >NetBSD-regression: NetBSD Build System <jenkins.org> >CentOS-regression: Gluster Build System <jenkins.com> >Reviewed-by: Ravishankar N <ravishankar> BUG: 1337822 Change-Id: Ic2450d34d3bf1fb6be754ce890aeca960fe7ad1f Signed-off-by: Pranith Kumar K <pkarampu> Reviewed-on: http://review.gluster.org/14448 Smoke: Gluster Build System <jenkins.com> NetBSD-regression: NetBSD Build System <jenkins.org> CentOS-regression: Gluster Build System <jenkins.com> Reviewed-by: Anuradha Talur <atalur> Reviewed-by: Niels de Vos <ndevos>
COMMIT: http://review.gluster.org/14449 committed in release-3.8 by Niels de Vos (ndevos) ------ commit a770c7bba13734602b11a750e037cb11e42fe706 Author: Pranith Kumar K <pkarampu> Date: Mon May 16 15:05:36 2016 +0530 cluster/afr: Refresh inode for inode-write fops in need Problem: If a named fresh-lookup is done on an loc and the fop fails on one of the bricks or not sent on one of the bricks, but by the time response comes to afr, if the brick is up, 'can_interpret' will be set to false in afr_lookup_done(), this will lead to inode-ctx for that inode to be not set, this can lead to EIO in case of a transaction as it depends on 'readable' array to be available by that point. Fix: Refresh inode for inode-write fops for the ctx to be set if it is not already done at the time of named fresh-lookup or if the file is in split-brain where we need to perform one more refresh before failing the fop to check if the file is still in split-brain or not. >BUG: 1336612 >Change-Id: I5c50b62c8de06129b8516039f7c252e5008c47a5 >Signed-off-by: Pranith Kumar K <pkarampu> >Reviewed-on: http://review.gluster.org/14368 >Smoke: Gluster Build System <jenkins.com> >NetBSD-regression: NetBSD Build System <jenkins.org> >Reviewed-by: Ravishankar N <ravishankar> >CentOS-regression: Gluster Build System <jenkins.com> BUG: 1337822 Change-Id: I0f904ebaa78b99cbb11546e08c9fc1562e9a3eef Signed-off-by: Pranith Kumar K <pkarampu> Reviewed-on: http://review.gluster.org/14449 Smoke: Gluster Build System <jenkins.com> NetBSD-regression: NetBSD Build System <jenkins.org> Reviewed-by: Krutika Dhananjay <kdhananj> Reviewed-by: Anuradha Talur <atalur> CentOS-regression: Gluster Build System <jenkins.com> Reviewed-by: Niels de Vos <ndevos>
REVIEW: http://review.gluster.org/14546 (cluster/afr: Fix warning about unused variable) posted (#1) for review on release-3.8 by Pranith Kumar Karampuri (pkarampu)
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.8.0, please open a new bug report. glusterfs-3.8.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution. [1] http://blog.gluster.org/2016/06/glusterfs-3-8-released/ [2] http://thread.gmane.org/gmane.comp.file-systems.gluster.user