Description of problem:
vdsm and libvirt do not come up after reboot
Version-Release number of selected component (if applicable):
Steps to Reproduce:
1. install 3.5 rhevm
2. install 3.5 6.6 rhevh from above
3. configure netzworkin on rhevh, set password on rhevm page
4. add/register rhevh from rhevm side
5. put rhevh into maintenance mode
6. reboot rhevh from tui
After the reboot vdsm and libvirt are down.
host is down in RHEV-M
Host is up in RHEV-M
This does not happen on the 7.0 build.
Created attachment 958612 [details]
RHEV-H side logs
I couldn't reproduce the report. However, I see a few complains about this environment from vdsm logs you provided.
Thread-28::ERROR::2014-11-18 14:21:35,201::API::1692::vds::(_getHaInfo) failed to retrieve Hosted Engine HA score '[Errno 2] No such file or directory: '/etc/ovirt-hosted-engine/hosted-engine.conf''Is the Hosted Engine setup finished?
Thread-30::INFO::2014-11-18 14:21:36,214::logUtils::44::dispatcher::(wrapper) Run and protect: getSpmStatus(spUUID=u'00000002-0002-0002-0002-0000000002bb', options=None)
Thread-30::ERROR::2014-11-18 14:21:36,214::task::866::Storage.TaskManager.Task::(_setError) Task=`52ef1f68-60c7-429f-89b5-ab8f4486e641`::Unexpected error
Traceback (most recent call last):
File "/usr/share/vdsm/storage/task.py", line 873, in _run
File "/usr/share/vdsm/logUtils.py", line 45, in wrapper
File "/usr/share/vdsm/storage/hsm.py", line 609, in getSpmStatus
File "/usr/share/vdsm/storage/hsm.py", line 325, in getPool
StoragePoolUnknown: Unknown pool id, pool not connected: (u'00000002-0002-0002-0002-0000000002bb',)
Have you re-used your rhev-h instance or have you configured it as hosted-engine/storage anytime? Adding Sandro and Dan in CC too.
It might have been a single time. I also could not reproduce it anymore.
I might have also been to unpatient. It took a while for the nodes to come up.
(That's where I wanted to leave these lines, but 1152916 comment 8).
I've met this bug again.
vdsm does not come up because libvirt des not come up, libvirt does not come up because it requires a configured network.
A workaround is to
1. run dhclient to aquire an ip
2. start libvirtd manually
3. start vdsmd manually
Now and on all subsequent reboots vdsm will come up correct (as far as I can tell).
Created attachment 980216 [details]
logs from right after the registration
This attachement shows the logs and configuration right after the registration when vdsmd is up.
Created attachment 980217 [details]
logs after the first boot
This attachement shows the logs and configuration right after the first reboot after registration
Please note that the filenames of the attachements are wrong, the files are xz compressed, not bz2.
It didn't reproduce with these version & flow:
Tested with rhevm 3.5 vt13.6, rhev-hypervisor6-6.6-20140114.0
(rhev-h was vdsm-upgraded from rhev-hypervisor6-6.6-20141218.0)
1. Add rhev-h to rhevm
2. Put host in maintenance.
3. restart rhev-h via TUI
4. Activate rhev-h => became active.
Reducing the priority because the reproducability seems to be low.
Will this bug be fixed on 3.5.0? Then we should clear the doctext flag.
ok, RHEV Hypervisor - 6.6 - 20150128.0.el6ev
Fixed in rhev-3.5.0 https://gerrit.eng.lab.tlv.redhat.com/15482 hence does not require a release note.
Thanks for the confirmation - dropped the known issue explanation from the advisory to reflect the '-' here.
bugs were moved by ERRATA to RELEASE PENDING bug not closed probably due to errata error.
closing as 3.5.0 is released.