Created attachment 822537 [details] agent and vdsm logs Description of problem: Migration from host wit high cpu to low cpu failed with libvirt error. Version-Release number of selected component (if applicable): ovirt-hosted-engine-ha-0.1.0-0.5.1.beta1.el6ev.noarch How reproducible: Always Steps to Reproduce: 1. install hosted-engine 2. run hosted-engine --deploy on first host and finish process 3. run hosted-engine --deploy on second host with same path to storage to add host to HA rhevm 4. Try to migrate from host with high_cpu to host with low_cpu Actual results: Migration failed with libvirt error: libvirtError: operation failed: migration job: unexpectedly failed Expected results: Vm migration success Additional info: cluster cpu level all time on intel conroe family host with low_cpu - Intel(R) Xeon(R) CPU 5130 @ 2.00GHz host with high cpu - Intel(R) Xeon(R) CPU E5649 @ 2.53GHz Migration from low to high works fine
Note, also see bug 1015721 for the scenario from which this originated.
Created attachment 823957 [details] logs Added new logs include libvirt log
Also checked on regular rhevm is23, the same problem exist: define: problematic_host - migration on this host failed host_1 host_2 migration from problematic_host to host_1 or host_2 work fine migration from host_1 to host_2 or from host_2 to host_1 work fine migration from host_1 or host_2 to problematic_host, failed on all three hosts vdsm, libvirt and selinux versions the same
After the investigation the problem seems to be more general, and it also appear in rhevm and not just in hosted engine About reproducing, I was success to reproduce this bug just for migration to host cyan-vdsf.qa.lab.tlv.redhat.com, so if will need to do some extra investigation on this host, just tell me.
this looks exactly as bug 1013617. Would you please check the details there? It's VERIFIED
It's look as the same bug, but I have a new version of vdsm on all hosts, also after little investigation I found that problematic host was have some additional vdsm packets: vdsm-tests-4.13.0-0.9.beta1.el6ev.noarch vdsm-hook-qemucmdline-4.13.0-0.9.beta1.el6ev.noarch vdsm-hook-vhostmd-4.13.0-0.9.beta1.el6ev.noarch vdsm-api-4.13.0-0.9.beta1.el6ev.noarch vdsm-debug-plugin-4.13.0-0.9.beta1.el6ev.noarch vdsm-debuginfo-4.13.0-0.9.beta1.el6ev.x86_64 vdsm-gluster-4.13.0-0.9.beta1.el6ev.noarch vdsm-reg-4.13.0-0.9.beta1.el6ev.noarch So after: yum erase vdsm* yum install vdsm Migration started to work, so one of packages was responsible for this error. So now I am not sure if it bug or just some test package.
might be good to see which one is it…but it also might have been something messed up with environment which was corrected by vdsm reinstall (like restorecon) If you don't want to invest time into nailing this down I'd just closed the bug….
ok close bug, if I will encounter the same problem in future, will investigate more deeply Thanks