Bug 991087
| Summary: | network is lost after installing all in one on fedora 19 | ||
|---|---|---|---|
| Product: | [Retired] oVirt | Reporter: | Ohad Basan <obasan> |
| Component: | ovirt-engine-installer | Assignee: | Yedidyah Bar David <didi> |
| Status: | CLOSED DUPLICATE | QA Contact: | |
| Severity: | unspecified | Docs Contact: | |
| Priority: | urgent | ||
| Version: | 3.3 | CC: | acathrow, danken, eedri, iheim, mgoldboi, obasan, oschreib, yeylon |
| Target Milestone: | --- | ||
| Target Release: | --- | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | network | ||
| Fixed In Version: | Doc Type: | Bug Fix | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2013-08-27 15:27:32 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Ohad Basan
2013-08-01 15:02:50 UTC
I really hope the Ofer understands what this bug is all about, since I do not. To track this properly, please report the versions of the relevant packages (including vdsm), the content of the relevant ifcfg files before and after installation, and relevant logs. I tried reproducing that on a clean machine
using ovirt-engine-3.3.0-0.2.master.20130813190148.git4d3992c.fc19.noarch
and now the bridge is not being created at all and the all in one plugin does not seem to be running
2013-08-14 11:05:42 DEBUG otopi.context context._executeMethod:132 method exception
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/otopi/context.py", line 122, in _executeMethod
method['method']()
File "/usr/share/ovirt-engine/setup/bin/../plugins/ovirt-engine-setup/all-in-one/vdsm.py", line 236, in _closeup
self.environment[osetupcons.AIOEnv.LOCAL_CLUSTER]
File "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/brokers.py", line 1331, in get
headers={}
File "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/proxy.py", line 52, in get
return self.request(method='GET', url=url, headers=headers)
File "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/proxy.py", line 112, in request
persistent_auth=self._persistent_auth)
File "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/proxy.py", line 134, in __doRequest
persistent_auth=persistent_auth
File "/usr/lib/python2.7/site-packages/ovirtsdk/web/connection.py", line 133, in doRequest
raise RequestError, response
RequestError:
status: 503
reason: Service Unavailable
it's a clean f19 machine
Did not manage to reproduce. Started from a clean f19 VM. Before setup, I had: [root@ovirte1 network-scripts]# cat ifcfg-eth0 HWADDR=52:54:00:1C:7B:48 TYPE=Ethernet BOOTPROTO=dhcp DEFROUTE=yes PEERDNS=yes PEERROUTES=yes IPV4_FAILURE_FATAL=no IPV6INIT=yes IPV6_AUTOCONF=yes IPV6_DEFROUTE=yes IPV6_PEERDNS=yes IPV6_PEERROUTES=yes IPV6_FAILURE_FATAL=no NAME=eth0 UUID=d41795c2-637e-4f48-b348-122317a98883 ONBOOT=yes After setup with allinone, I had: [root@ovirte1 network-scripts]# cat ifcfg-eth0 # Generated by VDSM version 4.12.0-78.git8f7eebf.fc19 DEVICE=eth0 ONBOOT=yes HWADDR=52:54:00:1c:7b:48 BRIDGE=ovirtmgmt MTU=1500 NM_CONTROLLED=no STP=no [root@ovirte1 network-scripts]# cat ifcfg-ovirtmgmt # Generated by VDSM version 4.12.0-78.git8f7eebf.fc19 DEVICE=ovirtmgmt ONBOOT=yes TYPE=Bridge DELAY=0 BOOTPROTO=dhcp DEFROUTE=yes NM_CONTROLLED=no STP=no Naturally, network configuration changes might temporarily disconnect the network, but this didn't happen for me this time. We still might want to add a recommendation to run under 'screen' just in case, as we do in hosted-engine: http://gerrit.ovirt.org/17258 Versions: ovirt-engine-3.4.0-0.2.master.20130822162248.gite4782ef.fc19.noarch vdsm-4.12.0-78.git8f7eebf.fc19.x86_64 Did the all in one installation finished for you succeessfully with host in "up" state? this is odd. =======================================================================
[ INFO ] Stage: Closing up
--== SUMMARY ==--
[WARNING] Warning: Not enough memory is available on the host. Minimum requirement is 4096MB, and 16384MB is recommended.
SSH fingerprint: 68:55:f3:1b:d8:77:27:7a:48:9e:b0:25:00:88:81:4d
Internal CA SHA1 Fingerprint=80:D5:19:2E:DB:F7:4F:51:B2:0B:98:E7:96:75:C6:D2:53:1B:0A:90
Web access is enabled at:
http://ovirte1.home.local:80/ovirt-engine
https://ovirte1.home.local:443/ovirt-engine
Please use the user "admin" and password specified in order to login into oVirt Engine
--== END OF SUMMARY ==--
[ INFO ] Starting engine service
[ INFO ] Restarting httpd
[ INFO ] Waiting for VDSM host to become operational. This may take several minutes...
[ INFO ] Still waiting for VDSM host to become operational...
[ INFO ] The VDSM Host is now operational
[ INFO ] Generating answer file '/var/lib/ovirt-engine/setup/answers/20130827165149-setup.conf'
[ INFO ] Stage: Clean up
Log file is located at /var/log/ovirt-engine/setup/ovirt-engine-setup-20130827163658.log
[ INFO ] Stage: Pre-termination
[ INFO ] Stage: Termination
=======================================================================
You had in your log:
=======================================================================
2013-08-14 11:05:40 DEBUG otopi.plugins.ovirt_engine_setup.all-in-one.vdsm vdsm._closeup:184 Connecting to the Engine
2013-08-14 11:05:41 DEBUG otopi.plugins.ovirt_engine_setup.all-in-one.vdsm vdsm._closeup:202 Creating the local data center
2013-08-14 11:05:41 DEBUG otopi.plugins.ovirt_engine_setup.all-in-one.vdsm vdsm._closeup:212 Creating the local cluster into the local data center
2013-08-14 11:05:42 DEBUG otopi.plugins.ovirt_engine_setup.all-in-one.vdsm vdsm._closeup:227 Adding the local host to the local cluster
2013-08-14 11:05:42 DEBUG otopi.context context._executeMethod:132 method exception
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/otopi/context.py", line 122, in _executeMethod
method['method']()
File "/usr/share/ovirt-engine/setup/bin/../plugins/ovirt-engine-setup/all-in-one/vdsm.py", line 236, in _closeup
self.environment[osetupcons.AIOEnv.LOCAL_CLUSTER]
File "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/brokers.py", line 1331, in get
headers={}
File "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/proxy.py", line 52, in get
return self.request(method='GET', url=url, headers=headers)
File "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/proxy.py", line 112, in request
persistent_auth=self._persistent_auth)
File "/usr/lib/python2.7/site-packages/ovirtsdk/infrastructure/proxy.py", line 134, in __doRequest
persistent_auth=persistent_auth
File "/usr/lib/python2.7/site-packages/ovirtsdk/web/connection.py", line 133, in doRequest
raise RequestError, response
RequestError:
status: 503
reason: Service Unavailable
=======================================================================
meaning it did manage to connect to the engine and do some stuff, but failed
during 'Adding the local host to the local cluster'. This can be a failure of the engine, host-deploy or vdsm. If you still have this machine, please attach all of the relevant logs including host-deploy/vdsm/engine.log. If you don't, please try to reproduce using current versions.
Also, perhaps it was fixed by: http://gerrit.ovirt.org/18262
No, it's not related to http://gerrit.ovirt.org/18262 . Your log shows that you first had: 2013-08-14 11:05:37 INFO otopi.plugins.ovirt_engine_setup.core.misc misc._closeup:85 Starting engine service and then: 2013-08-14 11:05:37 INFO otopi.plugins.ovirt_engine_common.system.apache apache._closeup:76 Restarting httpd and only later 503. Verified again, just to make sure.
ran cleanup, removed some packages, restored ifcfg-* from backup, rebooted.
Had:
========================================================================
[root@ovirte1 ~]# nm-tool
NetworkManager Tool
State: connected (global)
- Device: eth0 [eth0] ---------------------------------------------------------
Type: Wired
Driver: virtio_net
State: connected
Default: yes
HW Address: 52:54:00:1C:7B:48
Capabilities:
Carrier Detect: yes
Wired Properties
Carrier: on
IPv4 Settings:
Address: 192.168.122.12
Prefix: 24 (255.255.255.0)
Gateway: 192.168.122.1
DNS: 192.168.122.1
========================================================================
Then installed again packages, ran engine-setup which finished successfully, then had:
========================================================================
[root@ovirte1 ~]# nm-tool
NetworkManager Tool
State: disconnected
- Device: bond0 ----------------------------------------------------------------
Driver: bonding
State: disconnected
Default: no
Capabilities:
Carrier Detect: yes
- Device: eth0 -----------------------------------------------------------------
Type: Wired
Driver: virtio_net
State: unmanaged
Default: no
HW Address: 52:54:00:1C:7B:48
Capabilities:
Carrier Detect: yes
Wired Properties
Carrier: on
========================================================================
So eth0 did change from "connected" to "unmanaged" by NM.
Ohad, seems like this issue has been resolved already. Please re-open if it happens again. *** This bug has been marked as a duplicate of bug 1001186 *** |