Bug 1162905

Summary: hardcoded gsyncd path causes geo-replication to fail on non-redhat systems
Product: [Community] GlusterFS Reporter: Jordan Tomkinson <jordan.tomkinson>
Component: geo-replicationAssignee: Aravinda VK <avishwan>
Status: CLOSED CURRENTRELEASE QA Contact:
Severity: medium Docs Contact:
Priority: medium    
Version: mainlineCC: avishwan, bugs, kkeithle
Target Milestone: ---Keywords: EasyFix, Triaged
Target Release: ---   
Hardware: All   
OS: Linux   
Whiteboard:
Fixed In Version: glusterfs-3.8rc2 Doc Type: Bug Fix
Doc Text:
Story Points: ---
Clone Of:
: 1165140 1165142 1304692 (view as bug list) Environment:
Last Closed: 2016-06-16 12:40:24 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 1165140, 1165142, 1304692    

Description Jordan Tomkinson 2014-11-12 00:33:27 UTC
Description of problem:
Line 16 of geo-replication/src/peer_gsec_create.in has a hard coded path of ${exec_prefix}/libexec/glusterfs/gsyncd - ignoring the value of remote_gsyncd option in gsyncd.conf

atleast in Ubuntu, there is no libexec folder anywhere on the system and the ubuntu debs provided by semiosis installs gsyncd to /usr/lib/x86_64-linux-gnu/glusterfs/

this file has not changed in 3.5 or 3.6, so affects both releases.

The only way to get distributed geo-replication working is to symlink the path to the correct location.

Version-Release number of selected component (if applicable): 3.5.1, 3.5.2, 3.6.1


How to reproduce:

follow the distributed geo-replication docs at https://github.com/gluster/glusterfs/blob/master/doc/admin-guide/en-US/markdown/admin_distributed_geo_rep.md on a non-redhat system


Steps to Reproduce:
1. setup geo-replication as per docs
2. start the geo-replication
3. status never changes from "Not Active"
4. log files show cannot execute gsyncd

Actual results:

E [glusterd-geo-rep.c:2685:glusterd_gsync_read_frm_status] 0-: Unable to read gsyncd status file
E [glusterd-geo-rep.c:2999:glusterd_read_status_file] 0-: Unable to read the statusfile for /export/sitedata brick for  ds0(master), 10.42.0.8::geo-repl(slave) session


Expected results:

geo-replication to work.

Additional info:

https://github.com/gluster/glusterfs/blob/release-3.6/geo-replication/src/peer_gsec_create.in#L16

Comment 2 Vijay Bellur 2016-01-27 09:44:09 UTC
REVIEW: http://review.gluster.org/13298 (geo-rep: Fix gsyncd prefix in gsec_create) posted (#1) for review on master by Aravinda VK (avishwan)

Comment 3 Aravinda VK 2016-01-27 10:10:33 UTC
*** Bug 1199885 has been marked as a duplicate of this bug. ***

Comment 4 Vijay Bellur 2016-01-28 09:59:41 UTC
REVIEW: http://review.gluster.org/13298 (geo-rep: Fix gsyncd prefix in gsec_create) posted (#2) for review on master by Aravinda VK (avishwan)

Comment 5 Vijay Bellur 2016-02-04 11:16:30 UTC
COMMIT: http://review.gluster.org/13298 committed in master by Kaleb KEITHLEY (kkeithle) 
------
commit 453abcb821b712f849f91a2a370a4debff89eed2
Author: Aravinda VK <avishwan>
Date:   Wed Jan 27 15:08:03 2016 +0530

    geo-rep: Fix gsyncd prefix in gsec_create
    
    gsec_create script is generated after running ./configure
    
    libexec dir was formed using $prefix/libexec, but in Debian based
    distributions libexec dir is not present, instead they use lib
    directory to store these scripts.
    
    With this patch, full libexec path is fetched during ./configure.
    
    BUG: 1162905
    Change-Id: I9f47a38e6ab0027c7df6716136fbe0635e95a593
    Signed-off-by: Aravinda VK <avishwan>
    Reviewed-on: http://review.gluster.org/13298
    Smoke: Gluster Build System <jenkins.com>
    Tested-by: Raghavendra Talur <rtalur>
    CentOS-regression: Gluster Build System <jenkins.com>
    NetBSD-regression: NetBSD Build System <jenkins.org>
    Reviewed-by: Kaleb KEITHLEY <kkeithle>

Comment 6 Niels de Vos 2016-06-16 12:40:24 UTC
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.8.0, please open a new bug report.

glusterfs-3.8.0 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution.

[1] http://blog.gluster.org/2016/06/glusterfs-3-8-released/
[2] http://thread.gmane.org/gmane.comp.file-systems.gluster.user