Bug 1478848 - Use migration profile in HE maintenance migration
Summary: Use migration profile in HE maintenance migration
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: ovirt-engine
Classification: oVirt
Component: BLL.HostedEngine
Version: 4.1.5.1
Hardware: x86_64
OS: Linux
low
medium
Target Milestone: ovirt-4.2.0
: ---
Assignee: Andrej Krejcir
QA Contact: Nikolai Sednev
URL:
Whiteboard:
Depends On: 1467063 1512534
Blocks: 1458745
TreeView+ depends on / blocked
 
Reported: 2017-08-07 10:10 UTC by Nikolai Sednev
Modified: 2017-12-20 10:49 UTC (History)
7 users (show)

Fixed In Version:
Doc Type: No Doc Update
Doc Text:
undefined
Clone Of: 1467063
Environment:
Last Closed: 2017-12-20 10:49:15 UTC
oVirt Team: SLA
Embargoed:
rule-engine: ovirt-4.2+
rule-engine: planning_ack+
msivak: devel_ack+
mavital: testing_ack+


Attachments (Terms of Use)
sosreport from the engine (9.63 MB, application/x-xz)
2017-08-08 10:51 UTC, Nikolai Sednev
no flags Details
puma18 (11.03 MB, application/x-xz)
2017-08-08 10:52 UTC, Nikolai Sednev
no flags Details
puma19 (10.58 MB, application/x-xz)
2017-08-08 10:53 UTC, Nikolai Sednev
no flags Details


Links
System ID Private Priority Status Summary Last Updated
oVirt gerrit 84437 0 master MERGED agent: Fix parameter name in RPC call for VM migration 2017-11-22 08:55:52 UTC
oVirt gerrit 84438 0 master MERGED agent: Use migration policy "Suspend workload if needed" 2017-11-22 08:55:55 UTC

Comment 3 Doron Fediuck 2017-08-08 08:32:28 UTC
We're unable to reproduce in a clean environment.
Please provide a reproducer with its relevant log files in a clean environment.
Please state the load (memory and cpu) and the destination machine.

Comment 4 Nikolai Sednev 2017-08-08 08:44:58 UTC
(In reply to Doron Fediuck from comment #3)
> We're unable to reproduce in a clean environment.
> Please provide a reproducer with its relevant log files in a clean
> environment.
> Please state the load (memory and cpu) and the destination machine.

Bug #1467063 contains everything you've asked, logs and everything.
Reproduction steps also mentioned within the original bug, e.g.

1.Clean deployment of SHE over NFS+addition of one or two NFS data storage domains on pair of ha-hosts.
2.Ensure you're running SHE VM on SPM ha-host.
3.Migrate SHE-VM to none-SPM.
4.Migrate SHE-VM back to SPM ha-host and run monitor hosted-engine --vm-status from CLI.
5.During migration of SHE-VM to SPM you expected to catch the score penalty being dropped to by 50 points due to migration retry.

Comment 5 Yaniv Kaul 2017-08-08 08:52:26 UTC
Re-setting needinfo per comment 3

Comment 6 Nikolai Sednev 2017-08-08 10:51:27 UTC
Created attachment 1310516 [details]
sosreport from the engine

Comment 7 Nikolai Sednev 2017-08-08 10:52:45 UTC
Created attachment 1310519 [details]
puma18

Comment 8 Nikolai Sednev 2017-08-08 10:53:29 UTC
Created attachment 1310520 [details]
puma19

Comment 9 Nikolai Sednev 2017-08-08 10:55:48 UTC
Puma 18 was the SPM host.

Comment 10 Doron Fediuck 2017-08-09 13:15:39 UTC
According to the engine logs the HE VM maanaged to migrate to the SPM host:
2017-07-02 07:03:15,175-04 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand] (DefaultQuartzScheduler5) [601534c9] START, FullListVDSCommand(HostName = puma18.scl.lab.tlv.redhat.com, FullListVDSCommandParameters:{runAsync='true', hostId='4962ca54-b194-40c5-99fd-01a402cc9ecd', vmIds='[5d543bb5-edfe-4414-a507-c06ea2d25368]'}), log id: 75ba1ffc

Comment 11 Nikolai Sednev 2017-08-09 13:48:11 UTC
(In reply to Doron Fediuck from comment #10)
> According to the engine logs the HE VM maanaged to migrate to the SPM host:
> 2017-07-02 07:03:15,175-04 INFO 
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler5) [601534c9] START, FullListVDSCommand(HostName =
> puma18.scl.lab.tlv.redhat.com,
> FullListVDSCommandParameters:{runAsync='true',
> hostId='4962ca54-b194-40c5-99fd-01a402cc9ecd',
> vmIds='[5d543bb5-edfe-4414-a507-c06ea2d25368]'}), log id: 75ba1ffc

And it also failed in the middle of migration, so there was a re-try and thus due to that re-try score was dropped by 50 points, then it was raised back to 3400, once re-try succeeded.
This bug is about re-try and what caused it.

Comment 15 Nikolai Sednev 2017-12-12 11:44:07 UTC
Tested on cleanly deployed ovirt-hosted-engine-setup-2.2.1-0.0.master.20171206172737.gitd3001c8.el7.centos.noarch with ovirt-engine-appliance-4.2-20171210.1.el7.centos.noarch, over Gluster storage and with one NFS datat storage domain.
Original issue was not reproduced, hence moving to verified.
See also screencast attached.

Comment 17 Sandro Bonazzola 2017-12-20 10:49:15 UTC
This bugzilla is included in oVirt 4.2.0 release, published on Dec 20th 2017.

Since the problem described in this bug report should be
resolved in oVirt 4.2.0 release, published on Dec 20th 2017, it has been closed with a resolution of CURRENT RELEASE.

If the solution does not work for you, please open a new bug report.


Note You need to log in before you can comment on or make changes to this bug.