Bug 1240614
Summary: | Gluster nfs started running on one of the nodes of ganesha cluster, even though ganesha was running on it | ||
---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Apeksha <akhakhar> |
Component: | nfs-ganesha | Assignee: | Kaleb KEITHLEY <kkeithle> |
Status: | CLOSED ERRATA | QA Contact: | Apeksha <akhakhar> |
Severity: | urgent | Docs Contact: | |
Priority: | unspecified | ||
Version: | rhgs-3.1 | CC: | akhakhar, asrivast, jthottan, kkeithle, ndevos, nlevinki, saujain, skoduri, vagarwal |
Target Milestone: | --- | Keywords: | ZStream |
Target Release: | RHGS 3.1.1 | ||
Hardware: | x86_64 | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | glusterfs-3.7.1-12 | Doc Type: | Bug Fix |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2015-10-05 07:18:37 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | 1226817, 1251857, 1254419 | ||
Bug Blocks: | 1251815 |
Description
Apeksha
2015-07-07 11:28:16 UTC
Hi Apeksha, I just looked at your machines. You have an old instance of NFS-Ganesha running on the first machine. This process was started on the 6th of July and is no longer responsive. Ideally, you should have torn down the cluster and this should have all NFS-Ganesha services stopped. This old instance, is still showing up in service nfs-ganesha status. When you run the command, gluster nfs-ganesha enable, we internally execute "service nfs-ganesha start" and since an older instance shows up in, it says success. The real case is, NFS-Ganesha technically does not exist. I have raised a bug for this in upstream, https://bugzilla.redhat.com/show_bug.cgi?id=1119601 If you see the glusterd logs, you can clearly see that ganesha.enable on does stop the gluster NFS service, [2015-07-07 18:32:08.206306] I [MSGID: 106540] [glusterd-utils.c:4153:glusterd_nfs_pmap_deregister] 0-glusterd: De-registered NFSV3 successfully [2015-07-07 18:32:08.206495] I [MSGID: 106540] [glusterd-utils.c:4162:glusterd_nfs_pmap_deregister] 0-glusterd: De-registered NLM v4 successfully [2015-07-07 18:32:08.206676] I [MSGID: 106540] [glusterd-utils.c:4171:glusterd_nfs_pmap_deregister] 0-glusterd: De-registered NLM v1 successfully [2015-07-07 18:32:08.206869] I [MSGID: 106540] [glusterd-utils.c:4180:glusterd_nfs_pmap_deregister] 0-glusterd: De-registered ACL v3 successfully Since NFS-Ganesha is an old instance, the next time you start a volume, gluster NFS starts on that machine. On all the other nodes, NFS-Ganesha is a new and "working" instance and GLuster-NFS does not and cannot come up on those nodes. I had worked on the bug listed above and it is too intermittent to reproduce. If you can reproduce this bug consistently, you can propose this as a blocker. Reproduced the issue again: 1. Create a 6X2 volume say testvol, perform some root-squash tests Export list for localhost: /testvol (everyone) 2. Stop glusterd on server1 and start it again Stopping glusterd:[ OK ] Starting glusterd:[ OK ] 3. Now delete volume testvol 4. When we create a new volume say nfsvol1, enable ganesha on it, it gets exported as nfs volume Export list for localhost: /nfsvol1 * I followed the same steps as recorded and it didn't get reproduced. Neither did it get reproduced on my set up and the QE set up. I am not sure how to reproduce it. Please update if you hit it again and attach all the logs You had also executed refresh-config before you executed these tests. This bug is fixed as part of another fix, https://bugzilla.redhat.com/show_bug.cgi?id=1242749 ran the automated rootsquash testcases on glusterfs-3.7.1-12.el7rhgs.x86_64, did not hit this issue Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHSA-2015-1845.html |