Note: This bug is displayed in read-only format because
the product is no longer active in Red Hat Bugzilla.
RHEL Engineering is moving the tracking of its product development work on RHEL 6 through RHEL 9 to Red Hat Jira (issues.redhat.com). If you're a Red Hat customer, please continue to file support cases via the Red Hat customer portal. If you're not, please head to the "RHEL project" in Red Hat Jira and file new tickets here. Individual Bugzilla bugs in the statuses "NEW", "ASSIGNED", and "POST" are being migrated throughout September 2023. Bugs of Red Hat partners with an assigned Engineering Partner Manager (EPM) are migrated in late September as per pre-agreed dates. Bugs against components "kernel", "kernel-rt", and "kpatch" are only migrated if still in "NEW" or "ASSIGNED". If you cannot log in to RH Jira, please consult article #7032570. That failing, please send an e-mail to the RH Jira admins at rh-issues@redhat.com to troubleshoot your issue as a user management inquiry. The email creates a ServiceNow ticket with Red Hat. Individual Bugzilla bugs that are migrated will be moved to status "CLOSED", resolution "MIGRATED", and set with "MigratedToJIRA" in "Keywords". The link to the successor Jira issue will be found under "Links", have a little "two-footprint" icon next to it, and direct you to the "RHEL project" in Red Hat Jira (issue links are of type "https://issues.redhat.com/browse/RHEL-XXXX", where "X" is a digit). This same link will be available in a blue banner at the top of the page informing you that that bug has been migrated.
Description of problem:
guest dirty-rate calculated is not accurate when loading stress using stress cmd
Version-Release number of selected component (if applicable):
qemu-kvm-6.0.0-19.module+el8.5.0+11385+6e7d542e.x86_64
How reproducible:
100%
Steps to Reproduce:
[In guest]# stress --vm 1 --vm-bytes 40M
[On host]# virsh domdirtyrate-calc avocado-vt-vm1
Start to calculate domain's memory dirty rate successfully.
# virsh domstats avocado-vt-vm1 --dirtyrate
Domain: 'avocado-vt-vm1'
dirtyrate.calc_status=2
dirtyrate.calc_start_time=348414
dirtyrate.calc_period=1
dirtyrate.megabytes_per_second=4
(The dirtyrate is not more than 10MiB/s)
Expected results:
The dirty rate calculated should be around 40MiB/s
Additional info:
For the other info, please check:
https://bugzilla.redhat.com/show_bug.cgi?id=1812723#c20
When I use stress tool to calculate dirty rate before and during migration, get strange results as followings(qemu-kvm-6.0.0-19.module+el8.5.0+11385+6e7d542e.x86_64):
1.in vm:
# stress --vm 1 --vm-bytes 40M
2.before migration, query dirty rate many times, it's around 6MBps:
{"execute":"calc-dirty-rate", "arguments": {"calc-time": 1}}
{"return": {}}
{"execute":"query-dirty-rate"}
{"return": {"status": "measured", "dirty-rate": 6, "start-time": 1035316, "calc-time": 1}}
3.during migration is active, check dirty rate many times, it's mostly more than 100MBps:
(qemu) info migrate
globals:
store-global-state: on
only-migratable: off
send-configuration: on
send-section-footer: on
decompress-error-check: on
clear-bitmap-shift: 18
Migration status: active
total time: 427250 ms
expected downtime: 413 ms
setup: 2 ms
transferred ram: 49035815 kbytes
throughput: 945.20 mbps
remaining ram: 38792 kbytes
total ram: 4211528 kbytes
duplicate: 1369381 pages
skipped: 0 pages
normal: 12232039 pages
normal bytes: 48928156 kbytes
dirty sync count: 3228
page size: 4 kbytes
multifd bytes: 0 kbytes
pages-per-second: 28980
dirty pages rate: 29275 pages ----> Query migrate data many times, always around 30000 pages
So seems the STRESS tool is not ideal for migration to load stress when we test dirty-rate feature.
As we know, stressapptest tool is better suited for migration than the stress tool when we test dirty-rate feature, and there's no other issues when we use stressapptest testing as below comment.
https://bugzilla.redhat.com/show_bug.cgi?id=1812723#c17
Hi Dave,
Could you help check Comment 1 & Comment 2 from me? Shall we update calculating dirty-rate cases to add stress tool as a test tool besides stressapptest?
Leo - between this bug and bug 1919863, are there commonalities? Any chance either one can be worked on / resolved for RHEL 9.2? I'm trying to stay ahead of problems that will end up on the "aging" close wontfix list.
Comment 9RHEL Program Management
2023-01-03 07:27:55 UTC
After evaluating this issue, there are no plans to address it further or fix it in an upcoming release. Therefore, it is being closed. If plans change such that this issue will be fixed in an upcoming release, then the bug can be reopened.