Bug 1325938
Summary: | VM stay in 'powering down' after stopping VM | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Product: | [oVirt] ovirt-engine | Reporter: | Israel Pinto <ipinto> | ||||||||||
Component: | BLL.Virt | Assignee: | Arik <ahadas> | ||||||||||
Status: | CLOSED CURRENTRELEASE | QA Contact: | Israel Pinto <ipinto> | ||||||||||
Severity: | high | Docs Contact: | |||||||||||
Priority: | high | ||||||||||||
Version: | 3.6.5 | CC: | bugs, mgoldboi, michal.skrivanek, sbonazzo | ||||||||||
Target Milestone: | ovirt-4.0.0-beta | Flags: | rule-engine:
ovirt-4.0.0+
rule-engine: planning_ack+ rule-engine: devel_ack+ mavital: testing_ack+ |
||||||||||
Target Release: | 4.0.0 | ||||||||||||
Hardware: | Unspecified | ||||||||||||
OS: | Unspecified | ||||||||||||
Whiteboard: | |||||||||||||
Fixed In Version: | Doc Type: | Bug Fix | |||||||||||
Doc Text: | Story Points: | --- | |||||||||||
Clone Of: | Environment: | ||||||||||||
Last Closed: | 2016-08-01 12:26:26 UTC | Type: | Bug | ||||||||||
Regression: | --- | Mount Type: | --- | ||||||||||
Documentation: | --- | CRM: | |||||||||||
Verified Versions: | Category: | --- | |||||||||||
oVirt Team: | Virt | RHEL 7.3 requirements from Atomic Host: | |||||||||||
Cloudforms Team: | --- | Target Upstream Version: | |||||||||||
Embargoed: | |||||||||||||
Attachments: |
|
Description
Israel Pinto
2016-04-11 12:56:13 UTC
Created attachment 1145991 [details]
engine_log
Created attachment 1145992 [details]
host_1_logs
Created attachment 1145993 [details]
host_2_logs
Created attachment 1145994 [details]
vm_tab_screenshot
The major problem here is the fact that the host was switched to maintenance while the engine thinks a VM is running on that host. The engine puts the host in powering-down state and expect it to be down or not to be reported anymore - but the fact that the host switched to maintenance means that the host is not monitored - and thus the VM will be stuck in this state until the host is activated. The engine switches the host to maintenance because it thinks no VM is running on the host - and is determined according to the number of VMs that are reported by VDSM. VDSM reports that no VM is running on the host, which is true. So we should probably change the test to check the number of VMs running on the host according to the *database* instead of the report from VDSM. sounds as a race, rare enough to wait for 4.0. might got exposed more often due to VM status events introduced in 3.6 Moving from 4.0 alpha to 4.0 beta since 4.0 alpha has been already released and bug is not ON_QA. Verify with: RHEVM Version: 4.0.2-0.2.rc1.el7ev Hosts: OS Version:RHEL - 7.2 - 9.el7_2.1 Kernel Version:3.10.0 - 327.22.2.el7.x86_64 KVM Version:2.3.0 - 31.el7_2.16 LIBVIRT Version:libvirt-1.2.17-13.el7_2.5 VDSM Version:vdsm-4.18.5.1-1.el7ev SPICE Version:0.12.4 - 15.el7_2.1 Steps with automation test(Check that migration is done to source host again): One VM, 2 Hosts in Cluster Loop 3 times: 1. Migrate VM 2. Put Host into maintenance (no delay between steps) Test - PASS |