Bug 1763815

Summary: NFV live migration fails with dpdk "--iova-mode va": Failed to load virtio-net:virtio
Product: Red Hat Enterprise Linux Fast Datapath Reporter: Neil Horman <nhorman>
Component: openvswitchAssignee: Adrián Moreno <amorenoz>
openvswitch sub component: ovs-dpdk QA Contact: Pei Zhang <pezhang>
Status: CLOSED ERRATA Docs Contact:
Severity: high    
Priority: high CC: aadam, amorenoz, atragler, chayang, ctrautma, dmarchan, fiezzi, jhsiao, jinzhao, juzhang, kanderso, kfida, ktraynor, maxime.coquelin, ovs-qe, pezhang, qding, tredaelli
Version: FDB 18.04Keywords: Regression
Target Milestone: ---   
Target Release: ---   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: openvswitch-2.9.0-125.el7fdp Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: 1738751
: 1792399 (view as bug list) Environment:
Last Closed: 2020-03-10 09:35:13 UTC Type: ---
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 1738751, 1764000    
Bug Blocks:    

Comment 1 Neil Horman 2019-10-21 16:37:06 UTC
Note: Given that this issue occurs in standalone dpdk it should likely be fixed in ovs-dpdk as well

Comment 4 Neil Horman 2019-12-13 15:13:15 UTC
I'm afraid I don't have an OVS setup to test on, do you happen to have one I can use?

Comment 5 Adrián Moreno 2019-12-16 11:21:31 UTC
Not really, I would just set up an OVS with two vhost interfaces, connect a qemu instance to each of them and perform the live migration while the guest is transmitting over the vhost interface. I can attach some scripts I used to reproduce a similar issue.

Thanks!

Comment 10 Pei Zhang 2020-02-05 12:23:29 UTC
Steps: Following steps of Description. We tested with Live migration with ovs+vhost-user+dpdk scenarios.

Reproduced with openvswitch-2.9.0-124.el7fdp.x86_64.

# /bin/virsh migrate --verbose --persistent --live rhel7.8 qemu+ssh://10.73.72.196/system
Migration: [100 %]error: internal error: qemu unexpectedly closed the monitor: 2020-02-05T05:48:14.685282Z qemu-kvm: -chardev socket,id=charnet1,path=/tmp/vhostuser0.sock,server: info: QEMU waiting for connection on: disconnected:unix:/tmp/vhostuser0.sock,server
2020-02-05T05:48:15.121134Z qemu-kvm: -chardev socket,id=charnet2,path=/tmp/vhostuser1.sock,server: info: QEMU waiting for connection on: disconnected:unix:/tmp/vhostuser1.sock,server
2020-02-05T05:48:30.260841Z qemu-kvm: VQ 0 size 0x400 < last_avail_idx 0x38ad - used_idx 0x4cb6
2020-02-05T05:48:30.261044Z qemu-kvm: Failed to load virtio-net:virtio
2020-02-05T05:48:30.261061Z qemu-kvm: error while loading state for instance 0x0 of device '0000:00:06.0:00.0/virtio-net'
2020-02-05T05:48:30.263631Z qemu-kvm: load of migration failed: Operation not permitted


Verified with openvswitch-2.9.0-125.el7fdp.x86_64:

# /bin/virsh migrate --verbose --persistent --live rhel7.8 qemu+ssh://10.73.72.196/system
Migration: [100 %]


So this bug has been fixed very well.


Note: We hit other migration issue which below bug is tracking.

Bug 1791267 - Guest vhost-user ports stop receiving MoonGen packets after migration

Comment 12 errata-xmlrpc 2020-03-10 09:35:13 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2020:0742

Comment 13 Red Hat Bugzilla 2023-09-14 05:44:48 UTC
The needinfo request[s] on this closed bug have been removed as they have been unresolved for 1000 days