Bug 1546129
Summary: | Geo-rep: glibc fix breaks geo-replication | ||
---|---|---|---|
Product: | [Community] GlusterFS | Reporter: | Kotresh HR <khiremat> |
Component: | geo-replication | Assignee: | Kotresh HR <khiremat> |
Status: | CLOSED CURRENTRELEASE | QA Contact: | |
Severity: | medium | Docs Contact: | |
Priority: | unspecified | ||
Version: | mainline | CC: | amukherj, bugs, bugzilla, csaba, dimitri.ars, fweimer, khiremat, rcyriac, rhinduja, rhs-bugs, sasundar, storage-qa-internal |
Target Milestone: | --- | ||
Target Release: | --- | ||
Hardware: | All | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | glusterfs-v4.1.0 | Doc Type: | If docs needed, set a value |
Doc Text: | Story Points: | --- | |
Clone Of: | 1542979 | Environment: | |
Last Closed: | 2018-06-20 18:00:23 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | 1542979 | ||
Bug Blocks: | 1544382 |
Description
Kotresh HR
2018-02-16 12:12:35 UTC
REVIEW: https://review.gluster.org/19544 (geo-rep: Remove lazy umount and use mount namespaces) posted (#4) for review on master by Kotresh HR COMMIT: https://review.gluster.org/19544 committed in master by "Kotresh HR" <khiremat> with a commit message- geo-rep: Remove lazy umount and use mount namespaces Lazy umounting the master volume by worker causes issues with rsync's usage of getcwd. Henc removing the lazy umount and using private mount namespace for the same. On the slave, the lazy umount is retained as we can't use private namespace in non root geo-rep setup. Change-Id: I403375c02cb3cc7d257a5f72bbdb5118b4c8779a BUG: 1546129 Signed-off-by: Kotresh HR <khiremat> Can this please be backported to 3.12 release? Yes, I will do it This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-v4.1.0, please open a new bug report. glusterfs-v4.1.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution. [1] http://lists.gluster.org/pipermail/announce/2018-June/000102.html [2] https://www.gluster.org/pipermail/gluster-users/ |