Bug 2051870 - SIGABRT when blockcopy from rbd
Summary: SIGABRT when blockcopy from rbd
Keywords:
Status: CLOSED DUPLICATE of bug 2037135
Alias: None
Product: Red Hat Enterprise Linux Advanced Virtualization
Classification: Red Hat
Component: qemu-kvm
Version: 8.6
Hardware: Unspecified
OS: Unspecified
unspecified
unspecified
Target Milestone: rc
: 8.5
Assignee: Virtualization Maintenance
QA Contact: Tingting Mao
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2022-02-08 09:03 UTC by Han Han
Modified: 2022-02-23 01:37 UTC (History)
10 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2022-02-08 12:01:56 UTC
Type: Bug
Target Upstream Version:
Embargoed:
hhan: needinfo-


Attachments (Terms of Use)
The XMLs and logs (43.17 KB, application/gzip)
2022-02-08 09:03 UTC, Han Han
no flags Details


Links
System ID Private Priority Status Summary Last Updated
Red Hat Issue Tracker RHELPLAN-111439 0 None None None 2022-02-08 09:13:52 UTC

Description Han Han 2022-02-08 09:03:18 UTC
Created attachment 1859724 [details]
The XMLs and logs

Description of problem:
As subject

Version-Release number of selected component (if applicable):
libvirt-8.0.0-3.module+el8.6.0+14098+5bee65f4.x86_64
qemu-kvm-6.2.0-5.module+el8.6.0+14025+ca131e0a.x86_64

How reproducible:
100%

Steps to Reproduce:
1. Prepare a running VM with rbd source (How to prepare ceph secret https://libvirt.org/formatsecret.html#CephUsageType)
<disk type="network" device="disk">
  <driver name="qemu" type="raw" cache="none" />
  <source protocol="rbd" name="rbd/rbd.img" index="1">
    <host name="10.66.146.31" port="6789" />
    <auth username="admin">
      <secret type="ceph" usage="ceph_example" />
    </auth>
  </source>
  <target dev="vda" bus="virtio" />
  <alias name="virtio-disk0" />
  <address type="pci" domain="0x0000" bus="0x00" slot="0x07"
  function="0x0" />
</disk>

2. Copy vda to local file
➜  ~ virsh blockcopy avocado-vt-vm1 vda /var/lib/libvirt/images/clone.qcow2 --transient-job --verbose --wait                                                                                 
Block Copy: [ 99 %]error: failed to query job for disk vda
error: Unable to read from monitor: Connection reset by peer

3. Check the coredump backtrace
Program terminated with signal SIGABRT, Aborted.
#0  __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:50
50        return ret;
[Current thread is 1 (Thread 0x7efeec390e40 (LWP 11164))]
(gdb) bt
#0  __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:50
#1  0x00007efee8c14db5 in __GI_abort () at abort.c:79
#2  0x00007efee8c14c89 in __assert_fail_base (fmt=0x7efee8d7d858 "%s%s%s:%u: %s%sAssertion `%s' failed.\n%n", assertion=0x7efedb167bc1 "req.bytes <= bytes",                                 
    file=0x7efedb167b68 "../block/rbd.c", line=1355, function=<optimized out>) at assert.c:92
#3  0x00007efee8c3a3a6 in __GI___assert_fail (assertion=assertion@entry=0x7efedb167bc1 "req.bytes <= bytes", file=file@entry=0x7efedb167b68 "../block/rbd.c", line=line@entry=1355,          
    function=function@entry=0x7efedb167fe0 <__PRETTY_FUNCTION__.32898> "qemu_rbd_co_block_status") at assert.c:101                                                                           
#4  0x00007efedb16591f in qemu_rbd_co_block_status (bs=<optimized out>, want_zero=<optimized out>, offset=2149580800, bytes=65536, pnum=0x7efe0f3f7f28, map=<optimized out>,                 
    file=0x7efe0f3f7bf8) at ../block/rbd.c:1355
#5  0x000055da9c8df70c in bdrv_co_block_status (bs=0x55da9d681de0, want_zero=want_zero@entry=true, offset=2149580800, bytes=65536, pnum=pnum@entry=0x7efe0f3f7f28,                           
    map=map@entry=0x7efe0f3f7ca0, file=0x7efe0f3f7ca8) at ../block/io.c:2489
#6  0x000055da9c8df7e5 in bdrv_co_block_status (bs=bs@entry=0x55da9d846bb0, want_zero=want_zero@entry=true, offset=offset@entry=2149580800, bytes=bytes@entry=65536,                         
    pnum=pnum@entry=0x7efe0f3f7f28, map=map@entry=0x0, file=0x0) at ../block/io.c:2557
#7  0x000055da9c8e2451 in bdrv_co_common_block_status_above (bs=bs@entry=0x55da9d846bb0, base=base@entry=0x0, include_base=include_base@entry=false, want_zero=want_zero@entry=true,         
    offset=offset@entry=2149580800, bytes=bytes@entry=65536, pnum=0x7efe0f3f7f28, map=0x0, file=0x0, depth=0x7efe0f3f7d54) at ../block/io.c:2651                                             
#8  0x000055da9c8b5852 in bdrv_common_block_status_above (bs=bs@entry=0x55da9d846bb0, base=base@entry=0x0, include_base=include_base@entry=false, want_zero=want_zero@entry=true,            
    offset=offset@entry=2149580800, bytes=65536, pnum=0x7efe0f3f7f28, map=0x0, file=0x0, depth=0x0) at block/block-gen.c:444                                                                 
#9  0x000055da9c8e26d4 in bdrv_block_status_above (bs=bs@entry=0x55da9d846bb0, base=base@entry=0x0, offset=offset@entry=2149580800, bytes=<optimized out>, pnum=pnum@entry=0x7efe0f3f7f28,   
    map=map@entry=0x0, file=0x0) at ../block/io.c:2727
#10 0x000055da9c8e9c08 in mirror_iteration (s=0x55da9dce7700) at ../block/mirror.c:556
#11 mirror_run (job=0x55da9dce7700, errp=<optimized out>) at ../block/mirror.c:1033
#12 0x000055da9c8c51c3 in job_co_entry (opaque=0x55da9dce7700) at ../job.c:955
#13 0x000055da9c9bbd93 in coroutine_trampoline (i0=<optimized out>, i1=<optimized out>) at ../util/coroutine-ucontext.c:173                                                                  
#14 0x00007efee8c17060 in ?? () at ../sysdeps/unix/sysv/linux/x86_64/__start_context.S:91 from /lib64/libc.so.6                                                                              
#15 0x00007ffce89ded90 in ?? ()
#16 0x0000000000000000 in ?? ()


Actual results:
The SIGABRT coredump

Expected results:
No SIGABRT coredump

Additional info:
The XMLs and logs in the attachment:
bt.txt: the full threads backtrace
libvirtd.log: The log of libvirtd and qmp
qemu.log: The log of qemu cmdline
vm.xml: The xml of VM

Comment 3 CongLi 2022-02-08 11:42:27 UTC
Hi Stefano,

Can you help confirm if it's the same root cause as BZ2037135?

Thanks.


Note You need to log in before you can comment on or make changes to this bug.