Bug 1005677

Summary: qemu dies when replugging virtserialport too quickly
Product: [Fedora] Fedora Reporter: Lukáš Doktor <ldoktor>
Component: qemuAssignee: Amit Shah <amit.shah>
Status: CLOSED EOL QA Contact: Fedora Extras Quality Assurance <extras-qa>
Severity: unspecified Docs Contact:
Priority: unspecified    
Version: 20CC: amit.shah, berrange, cfergeau, crobinso, dwmw2, itamar, pbonzini, rjones, scottt.tw, virt-maint
Target Milestone: ---   
Target Release: ---   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: Bug Fix
Doc Text:
Story Points: ---
Clone Of: 1005150 Environment:
Last Closed: 2015-06-29 12:21:12 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 1005150    
Bug Blocks:    
Attachments:
Description Flags
Strace of qemu just before death.
none
Guest script which repeatedly writes some data into vs1 serial port. none

Description Lukáš Doktor 2013-09-09 07:13:01 UTC
Created attachment 795508 [details]
Strace of qemu just before death.

+++ This bug was initially created as a clone of Bug #1005150 +++

Description of problem:
While testing fixed https://bugzilla.redhat.com/show_bug.cgi?id=796048 issue with port hot-plugging I discovered another problem.

When replugging the virtserialport very quickly while transferring data through it using the simple reproducer the qemu dies (ret=1) without any message. Please find the strace log. (last few cycles)

Version-Release number of selected component (if applicable):
Host (F19):
qemu-kvm-1.4.2-7.fc19.x86_64
(upstream qemu 1.5.+ - git)
Guest (F19):
kernel-3.10.10-200.fc19.x86_64

How reproducible:
Always (after few seconds)

Steps to Reproduce (simple reproducer):
1. execute qemu with one virtserialport :
qemu-system-x86_64 -device virtio-serial-pci -chardev socket,id=devvs1,path=/tmp/vs1,server,nowait -device virtserialport,chardev=devvs1,name=vs1,id=vs1 -chardev socket,id=hmp,path=/tmp/hmp,nowait,server -mon chardev=hmp -chardev socket,id=ser,path=/tmp/ser,nowait,server -device isa-serial,chardev=ser -hda f19-64.qcow2 -enable-kvm -m 512

2. in guest execute simple_reproducer.py

3. on host read the vs1
    while :; do sudo socat /tmp/vs1 -; done

4. on host start re-plugging the port
    while :; do echo -e "device_del vs1\ndevice_add virtserialport,id=vs1,chardev=devvs1,name=vs1" | sudo socat /tmp/hmp -; done

Actual results:
The qemu dies (ret=1) without any messages, no log in serial console, no coredump. When putting a sleep > 0.1 between port replugs, qemu survives (>1 minute).

Expected results:
Some data should be received in host (or at least the guest should survive)

Comment 1 Lukáš Doktor 2013-09-09 15:30:39 UTC
Created attachment 795667 [details]
Guest script which repeatedly writes some data into vs1 serial port.

Comment 2 Lukáš Doktor 2013-09-20 10:32:00 UTC
Hi guys, I was asked in another similar bug https://bugzilla.redhat.com/show_bug.cgi?id=1005150 to retest this.

Host (F19):
upstream qemu-1.6.50 (2d1fe1873a984d1c2c89ffa3d12949cafc718551)
Guest (F19):
kernel-3.11.1-200.fc19.x86_64

Following the same steps results in the same results. Only now I saw a kernel message:
[  209.822469] virtio_console virtio0: output:id 102 is not a head!

in the serial console log. After this message there were few plug/unplug iterations and than exit with ret = 1.

Comment 3 Lukáš Doktor 2013-09-20 10:35:02 UTC
The output of serial console (-serial stdio), where I executed the simple_reproducer.py:

[root@localhost tmp]# python repr.py 


[Errno 19] No such device
[Errno 19] No such device
[Errno 19] No such device
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'

# CUT

[Errno 19] No such device
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 19] No such device
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 19] No such device
[Errno 19] No such device
[Errno 19] No such device
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[  209.822469] virtio_console virtio0: output:id 102 is not a head!
[Errno 19] No such device
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 19] No such device
[Errno 19] No such device
[Errno 19] No such device
[Errno 19] No such device
[Errno 6] No such device or address: '/dev/virtio-ports/vs1'
[Errno 6] No such device or address: '/dev/virtio-ports/vs1'
[Errno 19] No such device
[Errno 19] No such device

# CUT

[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'
[Errno 2] No such file or directory: '/dev/virtio-ports/vs1'[medic@localhost ~ 1]$

Comment 4 Amit Shah 2013-09-20 11:16:58 UTC
So now qemu doesn't quit?

That means this the bug you were seeing was related to the guest kernel failing.  With the new kernel, you don't see anything similar?

Comment 5 Lukáš Doktor 2013-09-22 07:00:09 UTC
Hi Amit, unfortunately not. Qemu is the process, which dies with ret = 1. No messages in the qemu log, serial, nor in ssh.

Comment 6 Fedora End Of Life 2015-01-09 19:46:49 UTC
This message is a notice that Fedora 19 is now at end of life. Fedora 
has stopped maintaining and issuing updates for Fedora 19. It is 
Fedora's policy to close all bug reports from releases that are no 
longer maintained. Approximately 4 (four) weeks from now this bug will
be closed as EOL if it remains open with a Fedora 'version' of '19'.

Package Maintainer: If you wish for this bug to remain open because you
plan to fix it in a currently maintained version, simply change the 'version' 
to a later Fedora version.

Thank you for reporting this issue and we are sorry that we were not 
able to fix it before Fedora 19 is end of life. If you would still like 
to see this bug fixed and are able to reproduce it against a later version 
of Fedora, you are encouraged  change the 'version' to a later Fedora 
version prior this bug is closed as described in the policy above.

Although we aim to fix as many bugs as possible during every release's 
lifetime, sometimes those efforts are overtaken by events. Often a 
more recent Fedora release includes newer upstream software that fixes 
bugs or makes them obsolete.

Comment 7 Lukáš Doktor 2015-01-21 16:01:44 UTC
Retested on:

host: kernel-3.17.2-200.fc20.x86_64
qemu: upstream tag v2.2.0-rc5
guest: kernel-3.17.8-200.fc20

with the same results. It dies in less then 1 second.

Comment 8 Fedora End Of Life 2015-05-29 09:22:49 UTC
This message is a reminder that Fedora 20 is nearing its end of life.
Approximately 4 (four) weeks from now Fedora will stop maintaining
and issuing updates for Fedora 20. It is Fedora's policy to close all
bug reports from releases that are no longer maintained. At that time
this bug will be closed as EOL if it remains open with a Fedora  'version'
of '20'.

Package Maintainer: If you wish for this bug to remain open because you
plan to fix it in a currently maintained version, simply change the 'version' 
to a later Fedora version.

Thank you for reporting this issue and we are sorry that we were not 
able to fix it before Fedora 20 is end of life. If you would still like 
to see this bug fixed and are able to reproduce it against a later version 
of Fedora, you are encouraged  change the 'version' to a later Fedora 
version prior this bug is closed as described in the policy above.

Although we aim to fix as many bugs as possible during every release's 
lifetime, sometimes those efforts are overtaken by events. Often a 
more recent Fedora release includes newer upstream software that fixes 
bugs or makes them obsolete.

Comment 9 Lukáš Doktor 2015-06-26 11:14:13 UTC
OK I give up. I had been retesting this every couple of months for 2 years now without any response. Feel free to test it yourself, the simple reproducer is in the report.

Comment 10 Cole Robinson 2015-06-26 18:17:00 UTC
Sorry this has lagged for so long, but the fedora qemu tracker only gets so much attention. The best way to get this realistically fixed IMO is to reproduce it against a RHEL version and file the bug + reproducer there, and QE will probably run with it

Comment 11 Amit Shah 2015-06-29 08:48:06 UTC
One way of ensuring someone's looking at it is to assign it to the relevant person.  In this case, that's me, and I could (should!) have assigned this bug to myself when I saw it.  Fixing that now.

Comment 12 Fedora End Of Life 2015-06-29 12:21:12 UTC
Fedora 20 changed to end-of-life (EOL) status on 2015-06-23. Fedora 20 is
no longer maintained, which means that it will not receive any further
security or bug fix updates. As a result we are closing this bug.

If you can reproduce this bug against a currently maintained version of
Fedora please feel free to reopen this bug against that version. If you
are unable to reopen this bug, please file a new report against the
current release. If you experience problems, please add a comment to this
bug.

Thank you for reporting this bug and we are sorry it could not be fixed.