Bug 1406820 - hosted-engine --deploy never completes
Summary: hosted-engine --deploy never completes
Keywords:
Status: CLOSED DUPLICATE of bug 1326798
Alias: None
Product: Red Hat Enterprise Virtualization Manager
Classification: Red Hat
Component: ovirt-hosted-engine-setup
Version: 4.0.0
Hardware: x86_64
OS: Linux
unspecified
high
Target Milestone: ovirt-4.1.1
: ---
Assignee: Simone Tiraboschi
QA Contact: meital avital
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2016-12-21 14:49 UTC by Anthony Green
Modified: 2023-09-14 03:36 UTC (History)
5 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2017-01-05 09:03:58 UTC
oVirt Team: Integration
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)
installer log file (42.87 KB, application/x-gzip)
2016-12-21 14:51 UTC, Anthony Green
no flags Details
/var/log/messages (47.84 KB, application/x-gzip)
2016-12-21 14:52 UTC, Anthony Green
no flags Details
hosted-engine --deploy answers file (3.14 KB, text/plain)
2016-12-21 14:53 UTC, Anthony Green
no flags Details

Description Anthony Green 2016-12-21 14:49:56 UTC
Description of problem:
It stalls at "Running engine-setup on the appliance".  No VM appears to be running on the host.  I'm running the latest RHV 4 from RHN as of today, but I'm not sure what minor version that would be called.

[root@rhv1 ~]# rpm -qa | grep ovirt
ovirt-engine-sdk-python-3.6.9.1-1.el7ev.noarch
ovirt-imageio-common-0.3.0-0.el7ev.noarch
ovirt-host-deploy-1.5.3-1.el7ev.noarch
ovirt-vmconsole-1.0.4-1.el7ev.noarch
ovirt-hosted-engine-setup-2.0.3-2.el7ev.noarch
ovirt-setup-lib-1.0.2-1.el7ev.noarch
ovirt-imageio-daemon-0.4.0-0.el7ev.noarch
ovirt-hosted-engine-ha-2.0.4-1.el7ev.noarch
ovirt-vmconsole-host-1.0.4-1.el7ev.noarch

I'm running on ravello, with nested virtualization enabled on my host.

Version-Release number of selected component (if applicable):


How reproducible:


Steps to Reproduce:
1.
2.
3.

Actual results:


Expected results:


Additional info:

Comment 1 Anthony Green 2016-12-21 14:51:17 UTC
Created attachment 1234442 [details]
installer log file

Comment 2 Anthony Green 2016-12-21 14:52:18 UTC
Created attachment 1234443 [details]
/var/log/messages

Comment 3 Anthony Green 2016-12-21 14:53:24 UTC
Created attachment 1234444 [details]
hosted-engine --deploy answers file

Comment 4 Yaniv Lavi 2017-01-03 08:24:13 UTC
What is your use case?
Why are you trying to run RHV on Ravello?

Comment 5 Anthony Green 2017-01-03 12:37:56 UTC
(In reply to Yaniv Dary from comment #4)
> What is your use case?
> Why are you trying to run RHV on Ravello?

I'm trying to test configurations for which I don't have enough hardware to test on my own (even using nested virt locally).   Thanks for looking into this.

Comment 6 Simone Tiraboschi 2017-01-03 14:33:33 UTC
The engine VM died pretty early:

Dec 21 09:29:32 rhv1 kvm: 1 guest now active
Dec 21 09:29:33 rhv1 journal: libvirt version: 2.0.0, package: 10.el7_3.2 (Red Hat, Inc. <http://bugzilla.redhat.com/bugzilla>, 2016-11-10-04:43:57, x86-034.build.eng.bos.redhat.com)
Dec 21 09:29:33 rhv1 journal: hostname: rhv1.atgreen.org
Dec 21 09:29:33 rhv1 journal: Failed to acquire lock: File exists
Dec 21 09:30:01 rhv1 systemd: Started Session 4 of user root.
Dec 21 09:30:01 rhv1 systemd: Starting Session 4 of user root.
Dec 21 09:31:32 rhv1 systemd: Starting Cleanup of Temporary Directories...
Dec 21 09:31:32 rhv1 systemd: Started Cleanup of Temporary Directories.
Dec 21 09:31:41 rhv1 systemd-logind: New session 5 of user root.
Dec 21 09:31:41 rhv1 systemd: Started Session 5 of user root.
Dec 21 09:31:41 rhv1 systemd: Starting Session 5 of user root.
Dec 21 09:31:41 rhv1 dbus-daemon: dbus[565]: [system] Activating service name='org.freedesktop.problems' (using servicehelper)
Dec 21 09:31:41 rhv1 dbus[565]: [system] Activating service name='org.freedesktop.problems' (using servicehelper)
Dec 21 09:31:41 rhv1 dbus[565]: [system] Successfully activated service 'org.freedesktop.problems'
Dec 21 09:31:41 rhv1 dbus-daemon: dbus[565]: [system] Successfully activated service 'org.freedesktop.problems'
Dec 21 09:31:51 rhv1 journal: internal error: character device console0 is not using a PTY
Dec 21 09:32:10 rhv1 kernel: ovirtmgmt: port 2(vnet0) entered disabled state
Dec 21 09:32:10 rhv1 kernel: device vnet0 left promiscuous mode
Dec 21 09:32:10 rhv1 kernel: ovirtmgmt: port 2(vnet0) entered disabled state
Dec 21 09:32:10 rhv1 NetworkManager[662]: <info>  [1482330730.8935] device (vnet0): state change: disconnected -> unmanaged (reason 'unmanaged') [30 10 3]
Dec 21 09:32:10 rhv1 journal: internal error: End of file from monitor
Dec 21 09:32:10 rhv1 kvm: 0 guests now active

Antony, could you please attach VDSM and libvirt logs?
Could you please try running it after stopping and masking NetworkManager?

systemctl stop NetworkManager
systemctl mask NetworkManager

Comment 7 Anthony Green 2017-01-05 03:14:47 UTC
This suddenly started working whey I masked NetworkManager.  Although I haven't tested enough to know if that was the fix, or if it's some other transient ravello related problem.

Now I've hit the problem of vdsm not recognizing the CPU for the cluster, or something like that. 

What would be great is if somebody already had RHV4 running successfully on Ravello, because it looks like some hacks to libvirt's cpu_map.xml file are required.

Comment 8 Simone Tiraboschi 2017-01-05 09:03:58 UTC
(In reply to Anthony Green from comment #7)
> This suddenly started working whey I masked NetworkManager.  Although I
> haven't tested enough to know if that was the fix, or if it's some other
> transient ravello related problem.

Unfortunately now we constantly see that issue when NetworkManager is active; please see https://bugzilla.redhat.com/1326798

*** This bug has been marked as a duplicate of bug 1326798 ***

Comment 9 Yaniv Lavi 2017-01-15 14:58:21 UTC
Can you confirm it work with NM being down?

Comment 10 Red Hat Bugzilla 2023-09-14 03:36:36 UTC
The needinfo request[s] on this closed bug have been removed as they have been unresolved for 1000 days


Note You need to log in before you can comment on or make changes to this bug.