Bug 1475755

Summary: gluster-blockd fails to start in RHGS container
Product: [Red Hat Storage] Red Hat Gluster Storage Reporter: krishnaram Karthick <kramdoss>
Component: CNS-deploymentAssignee: Michael Adam <madam>
Status: CLOSED WORKSFORME QA Contact: krishnaram Karthick <kramdoss>
Severity: high Docs Contact:
Priority: unspecified    
Version: cns-3.6CC: akhakhar, annair, hchiramm, jarrpa, kramdoss, madam, mliyazud, mzywusko, pprakash, rhs-bugs, rreddy, rtalur
Target Milestone: ---   
Target Release: ---   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2017-09-07 09:07:41 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 1445448    

Description krishnaram Karthick 2017-07-27 09:27:56 UTC
Description of problem:
Gluster-blockd service fails to start in RHGS container image as a dependency failure from rpcbind service. 

rpcbind service fails to start in a container as the port is already in use by  the rpcbind service running in the openshift node.

We'll need to reconfigure either of the ports to use a different port. This also means, we have to enable the new port number in firewall and document it. I'll raise a doc bug once we finalize what would be the port number.

Version-Release number of selected component (if applicable):
cns-deploy-5.0.0-12.el7rhgs.x86_64

How reproducible:
always

Steps to Reproduce:
1. configure cns
2. check if gluster-block service is up

Actual results:
gluster-blockd service is down

Expected results:
gluster-blockd service should be up

Additional info:

Comment 2 krishnaram Karthick 2017-07-27 09:33:11 UTC
sh-4.2# systemctl status rpcbind.socket -l
● rpcbind.socket - RPCbind Server Activation Socket
   Loaded: loaded (/usr/lib/systemd/system/rpcbind.socket; enabled; vendor preset: enabled)
   Active: failed (Result: resources)
   Listen: /var/run/rpcbind.sock (Stream)
           [::]:111 (Stream)
           0.0.0.0:111 (Stream)

Jul 27 09:06:58 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: rpcbind.socket failed to listen on sockets: Address already in use
Jul 27 09:06:58 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: Failed to listen on RPCbind Server Activation Socket.
Jul 27 09:06:58 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: Unit rpcbind.socket entered failed state.
Jul 27 09:06:58 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: Starting RPCbind Server Activation Socket.
Jul 27 09:15:46 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: rpcbind.socket failed to listen on sockets: Address already in use
Jul 27 09:15:46 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: Failed to listen on RPCbind Server Activation Socket.
Jul 27 09:15:46 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: Starting RPCbind Server Activation Socket.
Jul 27 09:15:50 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: rpcbind.socket failed to listen on sockets: Address already in use
Jul 27 09:15:50 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: Failed to listen on RPCbind Server Activation Socket.
Jul 27 09:15:50 dhcp46-203.lab.eng.blr.redhat.com systemd[1]: Starting RPCbind Server Activation Socket.
sh-4.2# exit

Comment 3 Raghavendra Talur 2017-07-27 10:03:34 UTC
Karthick,

please provide the gluster-blockd, glusterd service file from the container.
Also the netstat output from container and host
netstat -tnap | grep 111

Comment 4 Mohamed Ashiq 2017-07-27 19:32:12 UTC
Hi karthick,

Me and Talur gave it a try in our setup.

Steps we followed on host are 
1) modprobe target_core_user
2) systemctl enable rpcbind
3) systemctl start rpcbind

Then it seems to be working as expected in container in our setup's.
Can you give a try again.

Comment 5 Humble Chirammal 2017-08-02 16:26:37 UTC
Once 3 acks in place, I will move to ON_QA.

Comment 11 krishnaram Karthick 2017-09-07 09:07:41 UTC
The issue reported in this bug is no more seen in build - cns-deploy-5.0.0-32.el7rhgs

closing the bug as this issue is not seen with the latest builds.