RHEL Engineering is moving the tracking of its product development work on RHEL 6 through RHEL 9 to Red Hat Jira (issues.redhat.com). If you're a Red Hat customer, please continue to file support cases via the Red Hat customer portal. If you're not, please head to the "RHEL project" in Red Hat Jira and file new tickets here. Individual Bugzilla bugs in the statuses "NEW", "ASSIGNED", and "POST" are being migrated throughout September 2023. Bugs of Red Hat partners with an assigned Engineering Partner Manager (EPM) are migrated in late September as per pre-agreed dates. Bugs against components "kernel", "kernel-rt", and "kpatch" are only migrated if still in "NEW" or "ASSIGNED". If you cannot log in to RH Jira, please consult article #7032570. That failing, please send an e-mail to the RH Jira admins at rh-issues@redhat.com to troubleshoot your issue as a user management inquiry. The email creates a ServiceNow ticket with Red Hat. Individual Bugzilla bugs that are migrated will be moved to status "CLOSED", resolution "MIGRATED", and set with "MigratedToJIRA" in "Keywords". The link to the successor Jira issue will be found under "Links", have a little "two-footprint" icon next to it, and direct you to the "RHEL project" in Red Hat Jira (issue links are of type "https://issues.redhat.com/browse/RHEL-XXXX", where "X" is a digit). This same link will be available in a blue banner at the top of the page informing you that that bug has been migrated.
Bug 2157060 - 100g ice card: the guest start failed with Setting CPU affinity failed: Invalid argument when running sriov dpdk pvp case on rhel9.2
Summary: 100g ice card: the guest start failed with Setting CPU affinity failed: Inval...
Keywords:
Status: CLOSED DUPLICATE of bug 2154750
Alias: None
Product: Red Hat Enterprise Linux 9
Classification: Red Hat
Component: libvirt
Version: 9.2
Hardware: Unspecified
OS: Unspecified
unspecified
unspecified
Target Milestone: rc
: ---
Assignee: Virtualization Maintenance
QA Contact: yalzhang@redhat.com
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2022-12-30 11:28 UTC by liting
Modified: 2023-01-30 04:55 UTC (History)
5 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2023-01-06 06:10:45 UTC
Type: Bug
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Issue Tracker RHELPLAN-143403 0 None None None 2022-12-30 11:39:58 UTC

Description liting 2022-12-30 11:28:27 UTC
Description of problem:
100g ice card: the guest start failed with Setting CPU affinity failed: Invalid argument when running sriov dpdk pvp case on rhel9.2

Version-Release number of selected component (if applicable):
[root@wsfd-advnetlab153 ~]# uname -r
5.14.0-226.el9.x86_64
[root@wsfd-advnetlab153 perf]# rpm -qa|grep libvi
libvirt-libs-8.10.0-2.el9.x86_64
libvirt-client-8.10.0-2.el9.x86_64
libvirt-daemon-8.10.0-2.el9.x86_64
libvirt-daemon-driver-qemu-8.10.0-2.el9.x86_64
libvirt-daemon-driver-secret-8.10.0-2.el9.x86_64
libvirt-daemon-driver-storage-core-8.10.0-2.el9.x86_64
libvirt-daemon-driver-network-8.10.0-2.el9.x86_64
libvirt-daemon-config-network-8.10.0-2.el9.x86_64
libvirt-daemon-driver-nwfilter-8.10.0-2.el9.x86_64
libvirt-daemon-config-nwfilter-8.10.0-2.el9.x86_64
libvirt-daemon-driver-storage-iscsi-8.10.0-2.el9.x86_64
libvirt-daemon-driver-nodedev-8.10.0-2.el9.x86_64
libvirt-daemon-driver-storage-scsi-8.10.0-2.el9.x86_64
libvirt-daemon-driver-storage-mpath-8.10.0-2.el9.x86_64
libvirt-daemon-driver-storage-logical-8.10.0-2.el9.x86_64
libvirt-daemon-driver-storage-disk-8.10.0-2.el9.x86_64
libvirt-daemon-driver-interface-8.10.0-2.el9.x86_64
libvirt-daemon-driver-storage-rbd-8.10.0-2.el9.x86_64
libvirt-daemon-driver-storage-8.10.0-2.el9.x86_64
libvirt-8.10.0-2.el9.x86_64
python3-libvirt-8.10.0-1.el9.x86_64

How reproducible:


Steps to Reproduce:
1. create one vf for 2 pfs
2. bind the vf to dpdk vfio-pci driver
3. start guest with following xml
<domain type='kvm'>
  <name>g1</name>
  <memory unit='KiB'>8388608</memory>
  <currentMemory unit='KiB'>8388608</currentMemory>
  <memoryBacking>
    <hugepages>
      <page size='1048576' unit='KiB'/>
    </hugepages>
    <locked/>
    <access mode='shared'/>
  </memoryBacking>
  <vcpu placement='static'>3</vcpu>
  <cputune>
    <vcpupin vcpu='0' cpuset='5'/>
    <vcpupin vcpu='1' cpuset='59'/>
    <vcpupin vcpu='2' cpuset='3'/>
    <emulatorpin cpuset='1,57'/>
  </cputune>
  <numatune>
    <memory mode='strict' nodeset='0'/>
  </numatune>
  <resource>
    <partition>/machine</partition>
  </resource>
  <os>
    <type arch='x86_64' machine='q35'>hvm</type>
    <boot dev='hd'/>
  </os>
  <features>
    <acpi/>
    <apic/>
    <pmu state='off'/>
    <vmport state='off'/>
    <ioapic driver='qemu'/>
  </features>
  <cpu mode='host-passthrough' check='none'>
    <feature policy='require' name='tsc-deadline'/>
    <numa>
      <cell id='0' cpus='0-2' memory='8388608' unit='KiB' memAccess='shared'/>
    </numa>
  </cpu>
  <clock offset='utc'>
    <timer name='rtc' tickpolicy='catchup'/>
    <timer name='pit' tickpolicy='delay'/>
    <timer name='hpet' present='no'/>
  </clock>
  <on_poweroff>destroy</on_poweroff>
  <on_reboot>restart</on_reboot>
  <on_crash>restart</on_crash>
  <pm>
    <suspend-to-mem enabled='no'/>
    <suspend-to-disk enabled='no'/>
  </pm>
  <devices>
    <emulator>/usr/libexec/qemu-kvm</emulator>
    <disk type='file' device='disk'>
      <driver name='qemu' type='qcow2'/>
      <source file='/var/lib/libvirt/images/g1.qcow2'/>
      <backingStore/>
      <target dev='vda' bus='virtio'/>
      <alias name='virtio-disk0'/>
      <address type='pci' domain='0x0000' bus='0x01' slot='0x00' function='0x0'/>
    </disk>
    <controller type='usb' index='0' model='none'>
      <alias name='usb'/>
    </controller>
    <controller type='pci' index='0' model='pcie-root'>
      <alias name='pcie.0'/>
    </controller>
    <controller type='pci' index='1' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='1' port='0x10'/>
      <alias name='pci.1'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0'/>
    </controller>
    <controller type='pci' index='2' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='2' port='0x11'/>
      <alias name='pci.2'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x0'/>
    </controller>
    <controller type='pci' index='3' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='3' port='0x8'/>
      <alias name='pci.3'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x04' function='0x0'/>
    </controller>
    <controller type='pci' index='4' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='4' port='0x9'/>
      <alias name='pci.4'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x05' function='0x0'/>
    </controller>
    <controller type='pci' index='5' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='5' port='0xa'/>
      <alias name='pci.5'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x06' function='0x0'/>
    </controller>
    <controller type='pci' index='6' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='6' port='0xb'/>
      <alias name='pci.6'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0'/>
    </controller>
    <controller type='sata' index='0'>
      <alias name='ide'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x1f' function='0x2'/>
    </controller>
    <interface type='bridge'>
      <mac address='52:54:00:01:02:03'/>
      <source bridge='virbr0'/>
      <model type='virtio'/>
    </interface>
<hostdev mode='subsystem' type='pci' managed='yes'>
  <source>
    <address type='pci' domain='0x0000' bus='0xca' slot='0x01' function='0x0'/>
  </source>
  <mac address='00:de:ad:01:01:01'/>
</hostdev>
<hostdev mode='subsystem' type='pci' managed='yes'>
  <source>
    <address type='pci' domain='0x0000' bus='0xcd' slot='0x01' function='0x0'/>
  </source>
  <mac address='00:de:ad:02:02:02'/>
</hostdev>
    <serial type='pty'>
      <source path='/dev/pts/1'/>
      <target type='isa-serial' port='0'>
        <model name='isa-serial'/>
      </target>
      <alias name='serial0'/>
    </serial>
    <console type='pty' tty='/dev/pts/1'>
      <source path='/dev/pts/1'/>
      <target type='serial' port='0'/>
      <alias name='serial0'/>
    </console>
    <input type='mouse' bus='ps2'>
      <alias name='input0'/>
    </input>
    <input type='keyboard' bus='ps2'>
      <alias name='input1'/>
    </input>
    <graphics type='vnc' port='5900' autoport='yes' listen='0.0.0.0'>
      <listen type='address' address='0.0.0.0'/>
    </graphics>
    <video>
      <model type='cirrus' vram='16384' heads='1' primary='yes'/>
      <alias name='video0'/>
      <address type='pci' domain='0x0000' bus='0x05' slot='0x00' function='0x0'/>
    </video>
    <memballoon model='virtio'>
      <alias name='balloon0'/>
      <address type='pci' domain='0x0000' bus='0x06' slot='0x00' function='0x0'/>
    </memballoon>
    <iommu model='intel'>
      <driver intremap='on' caching_mode='on' iotlb='on'/>
    </iommu>
  </devices>
  <seclabel type='dynamic' model='selinux' relabel='yes'/>
</domain>

Actual results:
virsh start guest failed.
The issue start from RHEL-9.2.0-20221226.0. The RHEL-9.2.0-20221227.0 and RHEL-9.2.0-20221228.0 also has this issue.
[root@wsfd-advnetlab153 perf]# virsh start g1
error: Failed to start domain 'g1'
error: internal error: process exited while connecting to monitor: 2022-12-29T09:04:05.260067Z qemu-kvm: Setting CPU affinity failed: Invalid argument

RHEL-9.2.0-20221226.0 job:
https://beaker.engineering.redhat.com/jobs/7381987
RHEL-9.2.0-20221227.0 job:
https://beaker.engineering.redhat.com/jobs/7381105
RHEL-9.2.0-20221228.0 job:
https://beaker.engineering.redhat.com/jobs/7381027

The issue is not exist with RHEL-9.2.0-20221220.0.
https://beaker.engineering.redhat.com/jobs/7380993

Expected results:
virsh start guest successfully.

Additional info:

Comment 1 yalzhang@redhat.com 2023-01-05 02:18:10 UTC
It should be the same issue with bug 2154750.

Comment 2 liting 2023-01-06 03:30:02 UTC
RHEL-9.2.0-20221224.0 also has this issue
https://beaker.engineering.redhat.com/jobs/7382376

Comment 3 yalzhang@redhat.com 2023-01-06 06:10:45 UTC
Actually, this issue occurs since qemu-kvm-7.2.0-1.el9.x86_64

1. test on qemu-kvm-7.1.0:
# rpm -q qemu-kvm libvirt
qemu-kvm-7.1.0-7.el9.x86_64
libvirt-8.10.0-2.el9.x86_64

Set vm with xml includes:
<cputune>
    <emulatorpin cpuset='1'/>
  </cputune>
  <numatune>
    <memory mode='strict' nodeset='0'/>
  </numatune>
(host cpu 1 belongs to host node 1, not host node 0)

the related qemu cmdline:
-object '{"qom-type":"memory-backend-ram","id":"pc.ram","size":2147483648,"host-nodes":[0],"policy":"bind"}' \

2. Update to qemu-kvm-7.2.0-1.el9.x86_64, vm start failed with the same setting
# rpm -q libvirt  qemu-kvm
libvirt-8.10.0-2.el9.x86_64
qemu-kvm-7.2.0-1.el9.x86_64

# virsh start rhel 
error: Failed to start domain 'rhel'
error: internal error: process exited while connecting to monitor: 2023-01-06T06:05:25.568613Z qemu-kvm: Setting CPU affinity failed: Invalid argument

3. Update the setting to be:
<cputune>
    <emulatorpin cpuset='2'/>
  </cputune>
  <numatune>
    <memory mode='strict' nodeset='0'/>
  </numatune>
(host cpu 2 belongs to host node 0)

# virsh start rhel 
Domain 'rhel' started

related qemu cmdline:
-object '{"qom-type":"thread-context","id":"tc-pc.ram","node-affinity":[0]}' \
-object '{"qom-type":"memory-backend-ram","id":"pc.ram","size":2147483648,"host-nodes":[0],"policy":"bind","prealloc-context":"tc-pc.ram"}' \

*** This bug has been marked as a duplicate of bug 2154750 ***

Comment 4 liting 2023-01-30 04:55:05 UTC
rhel9.2 ctc2 still has this issue
https://beaker.engineering.redhat.com/jobs/7476912


Note You need to log in before you can comment on or make changes to this bug.