Bug 1289285
Summary: | Live Migration dynamic cpu throttling for auto-convergence (qemu-kvm-rhev) | |||
---|---|---|---|---|
Product: | Red Hat Enterprise Linux 7 | Reporter: | Hai Huang <hhuang> | |
Component: | qemu-kvm-rhev | Assignee: | Juan Quintela <quintela> | |
Status: | CLOSED ERRATA | QA Contact: | Qianqian Zhu <qizhu> | |
Severity: | unspecified | Docs Contact: | ||
Priority: | unspecified | |||
Version: | 7.3 | CC: | chayang, dmoessne, hhuang, huding, jcall, jen, juzhang, pezhang, qizhu, sherold, virt-maint, xfu | |
Target Milestone: | rc | Keywords: | FutureFeature | |
Target Release: | --- | |||
Hardware: | Unspecified | |||
OS: | Unspecified | |||
Whiteboard: | ||||
Fixed In Version: | qemu-kvm-rhev-2.5.0-1.el7 | Doc Type: | Enhancement | |
Doc Text: | Story Points: | --- | ||
Clone Of: | ||||
: | 1289288 1289290 1289291 (view as bug list) | Environment: | ||
Last Closed: | 2016-11-07 21:42:26 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | ||||
Bug Blocks: | 1252426, 1288337, 1289288, 1289290, 1289291, 1305606, 1313485, 1358141 |
Description
Hai Huang
2015-12-07 20:15:06 UTC
Test with: host: hp-dl585g7-05.lab.eng.pek2.redhat.com hp-dl585g7-04.lab.eng.pek2.redhat.com NIC Speed: 1000Mb/s Packages: qemu-kvm-rhev-2.6.0-19.el7.x86_64 kernel-3.10.0-478.el7.x86_64 Test matrix: migrate_set_speed 100M migrate_set_downtime 0.5 Mem stress: stressapptest -M {30, 50, 60, 100} migrate_set_parameter cpu-throttle-initial {20, 30} migrate_set_parameter cpu-throttle-increment {30, 5} Steps: 1. Launch guest on both src and guest: /usr/libexec/qemu-kvm -name linux -cpu Opteron_G5 -m 2048 -realtime mlock=off -smp 2,sockets=2,cores=1,threads=1 -device virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x6 -drive file=/nfsmnt/RHEL-Server-7.3-64-virtio-scsi.raw,if=none,id=scsi0,format=raw -device virtio-scsi-pci,id=scsi0 -device scsi-disk,drive=scsi0,scsi-id=0,lun=0 -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x8 -spice port=5901,disable-ticketing -vga qxl -global qxl-vga.revision=3 -netdev tap,id=hostnet0,vhost=on -device virtio-net-pci,netdev=hostnet0,id=net0,mac=3C:D9:2B:09:AB:44,bus=pci.0,addr=0x4 -monitor unix:/tmp/hmpmonitor,server,nowait 2. Set parameters for auto converge and check it on: migrate_set_capability auto-converge on migrate_set_parameter cpu-throttle-initial $INITIAL migrate_set_parameter cpu-throttle-increment $INCREMENT 3. Stress guest with: stressapptest -M $stress_mem -s 10000 4. Set speed and downtime: migrate_set_speed 100M migrate_set_downtime 0.5 5. Start migration: migrate -d tcp:$DEST_HOST_IP:$DEST_HOST_PORT Result: First, the cpu throttling percentage will start from $INITIAL and increased by $INCREMENT until migration could finish. With 30M memory stress, normal migration could finish by itself, and auto converge wont start anyway. With 50M memory stress, normal migration could finish by itself, but with auto converge on, migration time will be reduced from 51665 to 17419. And the cpu throttling percentage will finally up to to 40. With 60M memory stress, normal migration can't finish, with auto converge on the migration will finish when cpu throttling percentage up to 60-80. With 100M memory stress, normal migration can't finish, with auto converge on the migration will finish when cpu throttling percentage up to 90. Detailed data: Stress Autoconvege cpu throttle initial cpu throttle increment cpu throttle percentage final total time transferred ram speed downtime dirty sync counts Guest's CPU usage Avg 30M off - - - 11822 1135558 93.8MB/s 905 18 97.58% on 20 10 0 10697 1030037 94MB/s 776 11 96.71% 50M off - - - 51665 5254388 99.3MB/s 505 278 95.89% on 20 10 40 17419 1762094 98.8MB/s 395 57 100% 60M off - - - Unable to finish - - - - 97.73% on 20 10 80 34539 3509805 99.2MB/s 382 82 98.45% on 30 5 60 36549 3712110 99.2MB/s 398 74 100% 100M off - - - Unable to finish - - - - 100% on 20 10 90 39724 4028026 99MB/s 568 35 98.91% Hi Hai, Do you think we can verify this bug based on the above results? Thanks, Qianqian yes, the test results (with 50M memory stress,, 60M, and 100M) look good. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHBA-2016-2673.html |