Description of problem: Version-Release number of selected component (if applicable): # rpm -qa | grep -i vdsm vdsm-yajsonrpc-4.16.37-1.el6ev.noarch vdsm-hook-vhostmd-4.16.37-1.el6ev.noarch vdsm-reg-4.16.37-1.el6ev.noarch vdsm-cli-4.16.37-1.el6ev.noarch vdsm-jsonrpc-4.16.37-1.el6ev.noarch vdsm-python-zombiereaper-4.16.37-1.el6ev.noarch vdsm-xmlrpc-4.16.37-1.el6ev.noarch vdsm-4.16.37-1.el6ev.x86_64 vdsm-hook-ethtool-options-4.16.37-1.el6ev.noarch ovirt-node-plugin-vdsm-0.2.0-26.el6ev.noarch vdsm-python-4.16.37-1.el6ev.noarch Red Hat Enterprise Virtualization Manager Version 3.5.8-0.1.el6ev How reproducible: - Install rhev-hypervisor6-6.8-20160621.1.iso - Register and approve the node - Registration will fail Actual results: Registration failed. Expected results: Node become UP. Additional info: # service vdsmd start wdmd: unrecognized service vdsm: Start dependent wdmd [FAILED] vdsm start # vdsm-tool configure --force Checking configuration status... [FAILED] Traceback (most recent call last): File "/usr/bin/vdsm-tool", line 219, in main return tool_command[cmd]["command"](*args) File "/usr/lib/python2.6/site-packages/vdsm/tool/__init__.py", line 37, in wrapper File "/usr/lib/python2.6/site-packages/vdsm/tool/configurator.py", line 62, in configure File "/usr/lib/python2.6/site-packages/vdsm/tool/configurators/sanlock.py", line 71, in isconfigured KeyError: 'getpwnam(): name not found: sanlock'
Created attachment 1170871 [details] host-deploy log
Created attachment 1170873 [details] vdsm files
Note: That RHEV-H build is using the RHEV 3.5 package set
(In reply to Douglas Schilling Landgraf from comment #0) > # service vdsmd start > wdmd: unrecognized service > vdsm: Start dependent wdmd [FAILED] > vdsm start You must run vdsm-tool configure --force before starting vdsmd. > # vdsm-tool configure --force ... > KeyError: 'getpwnam(): name not found: sanlock' Is sanlock installed? Please attach output of "rpm -qa"
Created attachment 1171073 [details] rpm-qa
Douglass, the attached rpm-qa is from el7 system, but the bug is describing el6 system :-)
Created attachment 1171077 [details] vdsm-rpm-qa-el6.txt
(In reply to Nir Soffer from comment #7) > Douglass, the attached rpm-qa is from el7 system, but the bug is describing > el6 > system :-) Sorry Nir, I have updated a new attachment!
Thanks Douglas, Vdsm is requiring sanlock and sanlock is installed, but when host deploy is configuring sanlock, sanlock user is missing. May be vdsm installation issue, host-deploy issue or sanlock installation issue. David, do you have and idea why user sanlock was missing? Did sanlock install changed recently for el6?
Didi, can you check the host-deploy log? Do you have an idea why sanlock user is missing when host-deploy is trying to configure vdsm? Did we change anything in host-deploy for el6 recently?
Yes, sanlock was just recently rebuilt in rhel6 for bug 1344139. The problem there was a missing dependency on the useradd and groupadd commands, which the sanlock rpm needs.
Douglas, vdsm must require sanlock >= 2.8.3 to consume this fix, would you clone this bug to vdsm?
(In reply to Nir Soffer from comment #13) > Douglas, vdsm must require sanlock >= 2.8.3 to consume this fix, would you > clone > this bug to vdsm? Hi Nir, I have used edit-node and updated sanlock to 2.8.3 and registration process continued, so I can confirm it should work. Howerver, now I see: https://bugzilla.redhat.com/show_bug.cgi?id=1349218
Nir, FYI: As vdsm 3.5 won't get re-spin, I have included these changes into ovirt-node package.
(In reply to Douglas Schilling Landgraf from comment #16) > Nir, FYI: > > As vdsm 3.5 won't get re-spin, I have included these changes into ovirt-node > package. moving to node.
Based on comment 16 and comment 17, moved to ovirt-node.
(In reply to Nir Soffer from comment #18) > Based on comment 16 and comment 17, moved to ovirt-node. Hi Nir, vdsm in 3.5 will be re-spin based on the yesterday meeting. Moving back to vdsm.
Gil, I get too much spam from ack-police about missing qe ack. Would you handle this?
The vdsm was respinned
Does it happens in the latest RHEHV 6.8 ?
Test version: rhev-hypervisor6-6.8-20160707.3 vdsm-4.16.38-1.el6ev.x86_64 ovirt-node-3.2.3-34.el6.noarch RHEV-M vt 21(3.5.8-0.1.) Test steps: 1. Install rhev-hypervisor6-6.8-20160707.3 2. Register and approve the node 3. Check Node status Test result: 1. Node become UP. So the bug is fixed, change bug status to VERIFIED.