Bug 1025183
| Summary: | VmPoolMonitor throws a NullPointerException while starting a guest that in turn remains down with its images locked. | ||
|---|---|---|---|
| Product: | Red Hat Enterprise Virtualization Manager | Reporter: | rhev-integ |
| Component: | ovirt-engine | Assignee: | Tomas Jelinek <tjelinek> |
| Status: | CLOSED ERRATA | QA Contact: | Pavel Novotny <pnovotny> |
| Severity: | urgent | Docs Contact: | |
| Priority: | urgent | ||
| Version: | 3.1.5 | CC: | acathrow, ahadas, ahoness, byount, cboyle, iheim, lpeer, lyarwood, mavital, michal.skrivanek, pspacek, rgolan, Rhev-m-bugs, rnori, srevivo, tjelinek, yeylon, yzaslavs |
| Target Milestone: | --- | Keywords: | ZStream |
| Target Release: | 3.2.5 | ||
| Hardware: | All | ||
| OS: | Linux | ||
| Whiteboard: | virt | ||
| Fixed In Version: | Doc Type: | Bug Fix | |
| Doc Text: |
A problem existed with VmPoolMonitor throwing a NullPointerException while starting a guest that in turn remains down with its images locked. There was a chance that the prestarted pools' VMs ended up in an "image locked" state if turned off and the engine decided to turn them on immediately. This was caused by a race condition in async task management
The race condition is now fixed using proper locking of the shared resources. This means the prestarted VMs can be freely turned off and they will not end up in "image locked" state.
|
Story Points: | --- |
| Clone Of: | 1015638 | Environment: | |
| Last Closed: | 2013-12-18 14:10:02 UTC | Type: | --- |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | 1015638 | ||
| Bug Blocks: | |||
|
Description
rhev-integ
2013-10-31 08:03:37 UTC
Steps to Reproduce by Tomas: Have a pool with bunch of VMs (e.g. 20) all of them prestarted. Have the VmPoolMonitorIntervalInMinutes small (e.g. 1) and the VmPoolMonitorBatchSize big (20) (both from vdc_options). Now wait until all the VMs start. Now, select half of them and stop them. If it does not fail, wait until they start again and then stop half of them again. Repeat until it fails. This bug is currently attached to errata RHBA-2013:16431. If this change is not to be documented in the text for this errata please either remove it from the errata, set the requires_doc_text flag to minus (-), or leave a "Doc Text" value of "--no tech note required" if you do not have permission to alter the flag. Otherwise to aid in the development of relevant and accurate release documentation, please fill out the "Doc Text" field above with these four (4) pieces of information: * Cause: What actions or circumstances cause this bug to present. * Consequence: What happens when the bug presents. * Fix: What was done to fix the bug. * Result: What now happens when the actions or circumstances above occur. (NB: this is not the same as 'the bug doesn't present anymore') Once filled out, please set the "Doc Type" field to the appropriate value for the type of change made and submit your edits to the bug. For further details on the Cause, Consequence, Fix, Result format please refer to: https://bugzilla.redhat.com/page.cgi?id=fields.html#cf_release_notes Thanks in advance. Verified in rhevm-3.2.5-0.48.el6ev.noarch (sf22). Verified according to steps in comment 1: 1. Set VmPoolMonitorBatchSize=20 and VmPoolMonitorIntervalInMinutes=1. 2a. Create 2 prestarted pools, 10 VMs in each. 2b. Create 1 prestarted pool with 20 VMs. 3. Wait untill all VMs are up and then power off/shutdown approx. half of them at once. 4. Repeat step 3 with different VMs and slightly different number of them. Results: All stopped VMs were automatically started up again. No errors during the whole process were observed, no NPE in engine.log/server.log, no errors or warnings in events. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. http://rhn.redhat.com/errata/RHBA-2013-1831.html |