| Summary: | KVM migration works once, fails second time | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Product: | [Community] GlusterFS | Reporter: | Stephan Ellis <stephan.ellis> | ||||||||
| Component: | fuse | Assignee: | shishir gowda <sgowda> | ||||||||
| Status: | CLOSED WORKSFORME | QA Contact: | |||||||||
| Severity: | medium | Docs Contact: | |||||||||
| Priority: | medium | ||||||||||
| Version: | 3.3-beta | CC: | amarts, gluster-bugs, nsathyan | ||||||||
| Target Milestone: | --- | Keywords: | Triaged | ||||||||
| Target Release: | --- | ||||||||||
| Hardware: | x86_64 | ||||||||||
| OS: | Linux | ||||||||||
| Whiteboard: | |||||||||||
| Fixed In Version: | glusterfs-3.4.0qa6 | Doc Type: | Bug Fix | ||||||||
| Doc Text: | Story Points: | --- | |||||||||
| Clone Of: | |||||||||||
| : | 848343 (view as bug list) | Environment: | |||||||||
| Last Closed: | 2012-12-21 05:50:49 UTC | Type: | --- | ||||||||
| Regression: | --- | Mount Type: | --- | ||||||||
| Documentation: | --- | CRM: | |||||||||
| Verified Versions: | Category: | --- | |||||||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||||||
| Bug Depends On: | |||||||||||
| Bug Blocks: | 848343 | ||||||||||
| Attachments: |
|
||||||||||
|
Description
Stephan Ellis
2011-12-07 13:56:10 UTC
Created attachment 724 Created attachment 725 Using the same volume mounted as nfs, migration works flawlessly. I have to libvirt based KVM hosts. I've set up a gluster replicate volume between them, XFS as the actual filesystem for the bricks. Gluster volume is mounted on both hosts using the native fuse client. When migrating a VM from hyper1 to hyper2, it works. Migration from hyper2 to hyper1 fails. Same results if the first migration is hyper2 -> hyper1, second migration hyper1 -> hyper2 fails. Relevant log lines from hyper1 when migrating a second time: Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.063: info : qemudDispatchServer:1398 : Turn off polkit auth for privileged client 18310 Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.094: info : qemuSecurityDACSetOwnership:40 : Setting DAC user and group on '/mnt/gstor/ar-lab.img' to '107:107' Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.114: info : qemudDispatchSignalEvent:397 : Received unexpected signal 17 Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.220: info : qemudDispatchSignalEvent:397 : Received unexpected signal 17 Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.225: info : brProbeVnetHdr:449 : Enabling IFF_VNET_HDR Dec 7 10:53:00 hyper1 kernel: device vnet1 entered promiscuous mode Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.229: info : brProbeVnetHdr:449 : Enabling IFF_VNET_HDR Dec 7 10:53:00 hyper1 kernel: vsCore: port 2(vnet1) entering forwarding state Dec 7 10:53:00 hyper1 kernel: device vnet2 entered promiscuous mode Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.233: info : brProbeVnetHdr:449 : Enabling IFF_VNET_HDR Dec 7 10:53:00 hyper1 kernel: vsPrivate: port 4(vnet2) entering forwarding state Dec 7 10:53:00 hyper1 kernel: device vnet5 entered promiscuous mode Dec 7 10:53:00 hyper1 kernel: vsCluster: port 2(vnet5) entering forwarding state Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.254: info : qemudDispatchSignalEvent:397 : Received unexpected signal 17 Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.310: info : udevGetDeviceProperty:116 : udev reports device 'vnet2' does not have property 'DRIVER' Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.310: info : udevGetDeviceProperty:116 : udev reports device 'vnet2' does not have property 'PCI_CLASS' Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.317: info : udevGetDeviceProperty:116 : udev reports device 'vnet1' does not have property 'DRIVER' Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.317: info : udevGetDeviceProperty:116 : udev reports device 'vnet1' does not have property 'PCI_CLASS' Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.322: info : udevGetDeviceProperty:116 : udev reports device 'vnet5' does not have property 'DRIVER' Dec 7 10:53:00 hyper1 libvirtd: 10:53:00.322: info : udevGetDeviceProperty:116 : udev reports device 'vnet5' does not have property 'PCI_CLASS' Dec 7 10:53:02 hyper1 libvirtd: 10:53:02.091: info : qemuSecurityDACRestoreSecurityFileLabel:80 : Restoring DAC user and group on '/mnt/gstor/ar-lab.img' Dec 7 10:53:02 hyper1 libvirtd: 10:53:02.091: info : qemuSecurityDACSetOwnership:40 : Setting DAC user and group on '/mnt/gstor/ar-lab.img' to '0:0' Dec 7 10:53:02 hyper1 kernel: vsCore: port 2(vnet1) entering disabled state Dec 7 10:53:02 hyper1 kernel: device vnet1 left promiscuous mode Dec 7 10:53:02 hyper1 kernel: vsCore: port 2(vnet1) entering disabled state Dec 7 10:53:02 hyper1 kernel: vsPrivate: port 4(vnet2) entering disabled state Dec 7 10:53:02 hyper1 kernel: device vnet2 left promiscuous mode Dec 7 10:53:02 hyper1 kernel: vsPrivate: port 4(vnet2) entering disabled state Dec 7 10:53:02 hyper1 kernel: vsCluster: port 2(vnet5) entering disabled state Dec 7 10:53:02 hyper1 kernel: device vnet5 left promiscuous mode Dec 7 10:53:02 hyper1 kernel: vsCluster: port 2(vnet5) entering disabled state Dec 7 10:53:02 hyper1 libvirtd: 10:53:02.346: error : qemudDomainMigrateFinish2:11763 : internal error guest unexpectedly quit Suspect it to be an issue with O_DIRECT in open(). can we get 'strace -f -v' output? Created attachment 727 command was: strace -f -v -o migtrace.txt virsh migrate --live ar-lab qemu+ssh://hyper1/system Hi Stephan, There has been some work done for hosting of VM images on top of GlusterFS now... Can you please confirm if its fixed in latest qa releases? (as of today its glusterfs-3.4.0qa6) For us this is working fine. |