Bug 1648659
Summary: | VM crash with error "Lost connection with qemu process" after disk resize + fast IO - Core dump to iothread | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Product: | [oVirt] ovirt-engine | Reporter: | Yosi Ben Shimon <ybenshim> | ||||||||
Component: | BLL.Storage | Assignee: | Nobody <nobody> | ||||||||
Status: | CLOSED DEFERRED | QA Contact: | Avihai <aefrat> | ||||||||
Severity: | medium | Docs Contact: | |||||||||
Priority: | high | ||||||||||
Version: | 4.2.7 | CC: | aefrat, bugs, tnisan | ||||||||
Target Milestone: | --- | Keywords: | Automation | ||||||||
Target Release: | --- | ||||||||||
Hardware: | x86_64 | ||||||||||
OS: | Unspecified | ||||||||||
Whiteboard: | |||||||||||
Fixed In Version: | Doc Type: | If docs needed, set a value | |||||||||
Doc Text: | Story Points: | --- | |||||||||
Clone Of: | Environment: | ||||||||||
Last Closed: | 2021-09-29 11:33:06 UTC | Type: | Bug | ||||||||
Regression: | --- | Mount Type: | --- | ||||||||
Documentation: | --- | CRM: | |||||||||
Verified Versions: | Category: | --- | |||||||||
oVirt Team: | Storage | RHEL 7.3 requirements from Atomic Host: | |||||||||
Cloudforms Team: | --- | Target Upstream Version: | |||||||||
Embargoed: | |||||||||||
Bug Depends On: | 1671173, 1716349, 1809434 | ||||||||||
Bug Blocks: | |||||||||||
Attachments: |
|
This bug has not been marked as blocker for oVirt 4.3.0. Since we are releasing it tomorrow, January 29th, this bug has been re-targeted to 4.3.1. The issue happened again on ovirt-engine 4.3.2.1-0.1.el7 in automation Tier1 runs on the same TestCase5063. Try to reproduce locally did not succeed. Attached logs. Created attachment 1546616 [details]
engine and vdsm logs of the issue on ovirt-engine 4.3.2.1-0.1.el7
From VDSM log it looks like the extend did not really sinked in as allocated size: 1074659328 . Can you(Dev) please check it out ? Libvirt logs of that time are missing. 2019-03-20 17:57:08,780+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') Requesting extension for volume ca783c33-0ce1-4ee4-b6b3-8a462ed21fef on domain d3fc8698-a66f-4f00-a869-88557d5b1555 (apparent: 1073741824, capacity: 2147483648, allocated: 1074659328, physical: 1073741824 threshold_state: exceeded) (vm:1262) VDSM log: 2019-03-20 17:57:07,638+0200 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call Host.getAllVmStats succeeded in 0.00 seconds (__init__:312) 2019-03-20 17:57:07,807+0200 DEBUG (mpathlistener) [storage.udev] Received udev event (action=change, device=Device(u'/sys/devices/virtual/block/dm-13')) (udev:218) 2019-03-20 17:57:08,729+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') abnormal vm stop device ua-4f688f32-db98-4862-9c12-392d3dca88a9 error enospc (vm:5042) 2019-03-20 17:57:08,729+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') CPU stopped: onIOError (vm:6020) 2019-03-20 17:57:08,747+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') Requesting extension for volume ca783c33-0ce1-4ee4-b6b3-8a462ed21fef on domain d3fc8698-a66f-4f00-a869-88557d5b1555 (apparent: 1073741824, capacity: 2147483648, allocated: 1074659328, physical: 1073741824 threshold_state: exceeded) (vm:1262) 2019-03-20 17:57:08,748+0200 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='2ab6b5e8-dfde-41d3-8be3-bac017a3f0b5') moving from state finished -> state preparing (task:602) 2019-03-20 17:57:08,748+0200 INFO (libvirt/events) [vdsm.api] START sendExtendMsg(spUUID='f7507337-e82f-464a-be2e-dfadf68d1349', volDict={'newSize': 2147483648, 'domainID': 'd3fc8698-a66f-4f00-a869-88557d5b1555', 'name': 'vdb', 'poolID': 'f7507337-e82f-464a-be2e-dfadf68d1349', 'clock': <Clock(total=0.01*, extend-volume=0.01*)>, 'internal': False, 'volumeID': 'ca783c33-0ce1-4ee4-b6b3-8a462ed21fef', 'imageID': '4f688f32-db98-4862-9c12-392d3dca88a9'}, newSize=2147483648, callbackFunc=<bound method Vm.__afterVolumeExtension of <vdsm.virt.vm.Vm object at 0x7fabd035aa10>>) from=internal, task_id=2ab6b5e8-dfde-41d3-8be3-bac017a3f0b5 (api:48) 2019-03-20 17:57:08,749+0200 DEBUG (libvirt/events) [storage.SPM.Messages.Extend] new extend msg created: domain: d3fc8698-a66f-4f00-a869-88557d5b1555, volume: ca783c33-0ce1-4ee4-b6b3-8a462ed21fef (mailbox:128) 2019-03-20 17:57:08,749+0200 INFO (libvirt/events) [vdsm.api] FINISH sendExtendMsg return=None from=internal, task_id=2ab6b5e8-dfde-41d3-8be3-bac017a3f0b5 (api:54) 2019-03-20 17:57:08,749+0200 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='2ab6b5e8-dfde-41d3-8be3-bac017a3f0b5') finished: None (task:1201) 2019-03-20 17:57:08,749+0200 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='2ab6b5e8-dfde-41d3-8be3-bac017a3f0b5') moving from state finished -> state finished (task:602) 2019-03-20 17:57:08,749+0200 DEBUG (libvirt/events) [storage.ResourceManager.Owner] Owner.releaseAll requests {} resources {} (resourceManager:911) 2019-03-20 17:57:08,749+0200 DEBUG (libvirt/events) [storage.ResourceManager.Owner] Owner.cancelAll requests {} (resourceManager:948) 2019-03-20 17:57:08,749+0200 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='2ab6b5e8-dfde-41d3-8be3-bac017a3f0b5') ref 0 aborting False (task:1002) 2019-03-20 17:57:08,758+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') CPU stopped: onSuspend (vm:6020) 2019-03-20 17:57:08,761+0200 WARN (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') device vdb out of space (vm:3961) 2019-03-20 17:57:08,762+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') abnormal vm stop device ua-4f688f32-db98-4862-9c12-392d3dca88a9 error enospc (vm:5042) 2019-03-20 17:57:08,763+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') CPU stopped: onIOError (vm:6020) 2019-03-20 17:57:08,767+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') Requesting extension for volume ca783c33-0ce1-4ee4-b6b3-8a462ed21fef on domain d3fc8698-a66f-4f00-a869-88557d5b1555 (apparent: 1073741824, capacity: 2147483648, allocated: 1074659328, physical: 1073741824 threshold_state: exceeded) (vm:1262) 2019-03-20 17:57:08,767+0200 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='6e08c1d8-f1a4-46ae-a0e3-a026b8f8e171') moving from state finished -> state preparing (task:602) 2019-03-20 17:57:08,767+0200 INFO (libvirt/events) [vdsm.api] START sendExtendMsg(spUUID='f7507337-e82f-464a-be2e-dfadf68d1349', volDict={'newSize': 2147483648, 'domainID': 'd3fc8698-a66f-4f00-a869-88557d5b1555', 'name': 'vdb', 'poolID': 'f7507337-e82f-464a-be2e-dfadf68d1349', 'clock': <Clock(total=0.00*, extend-volume=0.00*)>, 'internal': False, 'volumeID': 'ca783c33-0ce1-4ee4-b6b3-8a462ed21fef', 'imageID': '4f688f32-db98-4862-9c12-392d3dca88a9'}, newSize=2147483648, callbackFunc=<bound method Vm.__afterVolumeExtension of <vdsm.virt.vm.Vm object at 0x7fabd035aa10>>) from=internal, task_id=6e08c1d8-f1a4-46ae-a0e3-a026b8f8e171 (api:48) 2019-03-20 17:57:08,768+0200 DEBUG (libvirt/events) [storage.SPM.Messages.Extend] new extend msg created: domain: d3fc8698-a66f-4f00-a869-88557d5b1555, volume: ca783c33-0ce1-4ee4-b6b3-8a462ed21fef (mailbox:128) 2019-03-20 17:57:08,768+0200 INFO (libvirt/events) [vdsm.api] FINISH sendExtendMsg return=None from=internal, task_id=6e08c1d8-f1a4-46ae-a0e3-a026b8f8e171 (api:54) 2019-03-20 17:57:08,768+0200 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='6e08c1d8-f1a4-46ae-a0e3-a026b8f8e171') finished: None (task:1201) 2019-03-20 17:57:08,768+0200 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='6e08c1d8-f1a4-46ae-a0e3-a026b8f8e171') moving from state finished -> state finished (task:602) 2019-03-20 17:57:08,768+0200 DEBUG (libvirt/events) [storage.ResourceManager.Owner] Owner.releaseAll requests {} resources {} (resourceManager:911) 2019-03-20 17:57:08,768+0200 DEBUG (libvirt/events) [storage.ResourceManager.Owner] Owner.cancelAll requests {} (resourceManager:948) 2019-03-20 17:57:08,768+0200 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='6e08c1d8-f1a4-46ae-a0e3-a026b8f8e171') ref 0 aborting False (task:1002) 2019-03-20 17:57:08,776+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') abnormal vm stop device ua-4f688f32-db98-4862-9c12-392d3dca88a9 error enospc (vm:5042) 2019-03-20 17:57:08,776+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') CPU stopped: onIOError (vm:6020) 2019-03-20 17:57:08,780+0200 INFO (libvirt/events) [virt.vm] (vmId='d8b804a1-3337-4e10-b480-5a73f4171b6c') Requesting extension for volume ca783c33-0ce1-4ee4-b6b3-8a462ed21fef on domain d3fc8698-a66f-4f00-a869-88557d5b1555 (apparent: 1073741824, capacity: 2147483648, allocated: 1074659328, physical: 1073741824 threshold_state: exceeded) (vm:1262) 2019-03-20 17:57:08,780+0200 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='02c8d42d-ca83-4b82-8c0d-fa5bd4d7c8a0') moving from state finished -> state preparing (task:602) 2019-03-20 17:57:08,781+0200 INFO (libvirt/events) [vdsm.api] START sendExtendMsg(spUUID='f7507337-e82f-464a-be2e-dfadf68d1349', volDict={'newSize': 2147483648, 'domainID': 'd3fc8698-a66f-4f00-a869-88557d5b1555', 'name': 'vdb', 'poolID': 'f7507337-e82f-464a-be2e-dfadf68d1349', 'clock': <Clock(total=0.00*, extend-volume=0.00*)>, 'internal': False, 'volumeID': 'ca783c33-0ce1-4ee4-b6b3-8a462ed21fef', 'imageID': '4f688f32-db98-4862-9c12-392d3dca88a9'}, newSize=2147483648, callbackFunc=<bound method Vm.__afterVolumeExtension of <vdsm.virt.vm.Vm object at 0x7fabd035aa10>>) from=internal, task_id=02c8d42d-ca83-4b82-8c0d-fa5bd4d7c8a0 (api:48) 2019-03-20 17:57:08,781+0200 DEBUG (libvirt/events) [storage.SPM.Messages.Extend] new extend msg created: domain: d3fc8698-a66f-4f00-a869-88557d5b1555, volume: ca783c33-0ce1-4ee4-b6b3-8a462ed21fef (mailbox:128) 2019-03-20 17:57:08,781+0200 INFO (libvirt/events) [vdsm.api] FINISH sendExtendMsg return=None from=internal, task_id=02c8d42d-ca83-4b82-8c0d-fa5bd4d7c8a0 (api:54) Once again same issue is encountered in automation TC5063 at a specific environment (infra env which is probably much faster than our(storage) envs). Also Core dump to |/usr/libexec/abrt-hook-ccpp 6 18446744073709551615 114797 107 107 1553965254 IO iothread1 is seen in journalctl log. See ALL logs(including libvirt and qemu) are attached. See details below. Issue occures multiple in latest 4.3 releases , Please assign and check it out. Build info: ovirt-engine ovirt-engine-4.3.3.1-0.1.el7.noarch ovirt-imageio-proxy ovirt-imageio-proxy-1.5.1-0.el7ev.noarch redhat_release Red Hat Enterprise Linux Server 7.6 (Maipo) build_info_product rhv build_info_version 4.3 build_info_name rhv-4.3.3-1 uptime 12021 kernel_release 3.10.0-957.el7.x86_64 nics_number 2 Host Name lynx18.lab.eng.tlv2.redhat.com vdsm vdsm-4.30.12-1.el7ev.x86_64 vdsm-cli null libvirt libvirt-4.5.0-10.el7_6.6.x86_64 qemu-img-rhev qemu-img-rhev-2.12.0-21.el7.x86_64 glusterfs glusterfs-3.12.2-47.el7.x86_64 vdsm-python vdsm-python-4.30.12-1.el7ev.noarch sanlock sanlock-3.6.0-1.el7.x86_64 ovirt-imageio-daemon ovirt-imageio-daemon-1.5.1-0.el7ev.noarch redhat_release Red Hat Enterprise Linux Server 7.6 (Maipo) uuid 4c4c4544-005a-3010-8056-b7c04f563232 build_info_product rhv build_info_version 4.2 build_info_name rhv-4.2.8-9 uptime 1642 kernel_release 3.10.0-957.10.1.el7.x86_64 nics_number 9 Details : Scenario: 16:54:43 2019-03-30 19:54:42,968 INFO Test Setup 1: Creating VM vm_TestCase5063_3019544296 16:54:43 2019-03-30 19:54:43,151 INFO Test Setup 2: [class] Description: clone vm from a pre-defined template golden_template_default_rhel8.0 with {'vol_sparse': True, 'name': 'vm_TestCase5063_3019544296', 'clone': False, 'cluster': 'golden_env_mixed_1', 'storagedomain': 'iscsi_1', 'display_type': 'spice', 'timeout': 900, 'virtio_scsi': True, 'vol_format': 'cow', 'os_type': 'RHEL6x64', 'type': 'desktop', 'wait': True} 16:55:05 2019-03-30 19:55:05,816 INFO Test Setup 3: Starting VM vm_TestCase5063_3019544296 16:55:05 2019-03-30 19:55:05,819 INFO Test Setup 4: [class] Start VM vm_TestCase5063_3019544296 with {'wait_for_ip': False, 'pause': False, 'use_cloud_init': False, 'timeout': 600, 'wait_for_status': 'up'} 16:56:17 2019-03-30 19:56:17,613 INFO Test Setup 5: Creating disk disk_TestCase5063_3019561761 16:56:17 2019-03-30 19:56:17,616 INFO Test Setup 6: [class] Add a disk with {'description': '', 'provisioned_size': 1073741824, 'format': 'cow', 'bootable': False, 'wipe_after_delete': False, 'alias': 'disk_TestCase5063_3019561761', 'storagedomain': 'iscsi_1', 'shareable': False, 'sparse': True, 'active': True, 'interface': 'virtio'} 16:56:38 2019-03-30 19:56:38,353 INFO Test Setup 7: Attach disk disk_TestCase5063_3019561761 to VM vm_TestCase5063_3019544296 16:56:38 2019-03-30 19:56:38,356 INFO Test Setup 8: [class] Attach disk to VM with {'read_only': False, 'vm_name': 'vm_TestCase5063_3019544296', 'alias': 'disk_TestCase5063_3019561761', 'bootable': False, 'interface': 'virtio', 'active': True} 16:56:39 2019-03-30 19:56:39,577 INFO 001: storage/rhevmtests.storage.storage_virtual_disk_resize.test_virtual_disk_resize.TestCase5063.test_thin_block_resize[iscsi] 16:56:39 2019-03-30 19:56:39,577 INFO - VM with thin disk and OS 16:56:39 2019-03-30 19:56:39,578 INFO - Resize the VM disk to 2G total 16:56:39 2019-03-30 19:56:39,578 INFO - Send IOs to disk 16:56:39 2019-03-30 19:56:39,578 INFO - Check LV size on VDSM and disk size on guest 16:56:39 2019-03-30 19:56:39,578 INFO STORAGE: ISCSI 16:56:39 2019-03-30 19:56:39,578 INFO Test Step 9: Resizing disk disk_TestCase5063_3019561761 16:56:39 2019-03-30 19:56:39,579 INFO Test Step 10: Resizing disk disk_TestCase5063_3019561761 16:56:39 2019-03-30 19:56:39,581 INFO Test Step 11: Extend an already existing vm vm_TestCase5063_3019544296 disk disk_TestCase5063_3019561761 with {'provisioned_size': 2147483648} => Step 12 (Send IOs to disk): On the re-sized disk we create a partition,file-system + mount and at about 20:00:46(17:00:46 local time) we start dd(Send IOs to disk) of 1358MB to disk disk to make the disk extend 2019-03-30 20:00:46,034 - MainThread - RemoteExecutor - DEBUG - [root.17.139/qum5net] Executing: /bin/dd bs=1M count=1358 if=/dev/vda of=/mount-point_3020004311/written_test_storage status=none -> RIGHT AFTER THAT THE ISSUE(iothread1 pip failed + VM pause due to no space ) OCCURS: journalctl log: Mar 30 20:00:54 lynx18.lab.eng.tlv2.redhat.com kernel: Core dump to |/usr/libexec/abrt-hook-ccpp 6 18446744073709551615 114797 107 107 1553965254 IO iothread1 114804 pipe failed Mar 30 20:00:54 lynx18.lab.eng.tlv2.redhat.com kvm[116765]: 0 guests now active Mar 30 20:00:54 lynx18.lab.eng.tlv2.redhat.com kernel: ovirtmgmt: port 2(vnet0) entered disabled state Mar 30 20:00:54 lynx18.lab.eng.tlv2.redhat.com kernel: device vnet0 left promiscuous mode Mar 30 20:00:54 lynx18.lab.eng.tlv2.redhat.com kernel: ovirtmgmt: port 2(vnet0) entered disabled state Mar 30 20:00:54 lynx18.lab.eng.tlv2.redhat.com NetworkManager[6547]: <info> [1553965254.4229] device (vnet0): state change: disconnected -> unmanaged (reason 'unmanaged', sys-iface-state: 'removed') Mar 30 20:00:54 lynx18.lab.eng.tlv2.redhat.com NetworkManager[6547]: <info> [1553965254.4233] device (vnet0): released from master device ovirtmgmt Mar 30 20:00:54 lynx18.lab.eng.tlv2.redhat.com lldpad[85234]: recvfrom(Event interface): No buffer space available Mar 30 20:00:54 lynx18.lab.eng.tlv2.redhat.com systemd-machined[114798]: Machine qemu-1-vmTestCase50633019544296 terminated. Engine: 2019-03-30 20:00:52,403+03 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ForkJoinPool-1-worker-7) [] EVENT_ID: VM_PAUSED_ENOSPC(138), VM vm_TestCase5063_3019544296 has been paused due to no Storage space error. 2019-03-30 20:00:56,432+03 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] (EE-ManagedThreadFactory-engineScheduled-Thread-95) [] VM '2e0eed87-5b81-4c2f-a439-c9fdd1b036c0'(vm_TestCase5063_3019544296) moved from 'Paused' --> 'Up' 2019-03-30 20:00:56,472+03 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (EE-ManagedThreadFactory-engineScheduled-Thread-95) [] EVENT_ID: VM_RECOVERED_FROM_PAUSE_ERROR(196), VM vm_TestCase5063_3019544296 has recovered from paused back to up. VDSM: 2019-03-30 20:00:51,826+0300 DEBUG (mpathlistener) [storage.udev] Received udev event (action=change, device=Device(u'/sys/devices/virtual/block/dm-13')) (udev:218) 2019-03-30 20:00:52,287+0300 INFO (libvirt/events) [virt.vm] (vmId='2e0eed87-5b81-4c2f-a439-c9fdd1b036c0') abnormal vm stop device ua-12d0b157-1409-4c8b-987b-06cc22c964b6 error enospc (vm:5051) 2019-03-30 20:00:52,288+0300 INFO (libvirt/events) [virt.vm] (vmId='2e0eed87-5b81-4c2f-a439-c9fdd1b036c0') CPU stopped: onIOError (vm:6029) 2019-03-30 20:00:52,304+0300 INFO (libvirt/events) [virt.vm] (vmId='2e0eed87-5b81-4c2f-a439-c9fdd1b036c0') Requesting extension for volume c8741dc2-72a9-4941-a4b5-3be664b05519 on domain 0d8d55a2-0e4e-43a2-b179 -6530b1571ad0 (apparent: 1073741824, capacity: 2147483648, allocated: 1075118080, physical: 1073741824 threshold_state: exceeded) (vm:1262) qum5net 2019-03-30 20:00:52,320+0300 INFO (libvirt/events) [virt.vm] (vmId='2e0eed87-5b81-4c2f-a439-c9fdd1b036c0') CPU stopped: onSuspend (vm:6029) 2019-03-30 20:00:52,323+0300 WARN (libvirt/events) [virt.vm] (vmId='2e0eed87-5b81-4c2f-a439-c9fdd1b036c0') device vdb out of space (vm:3970) 2019-03-30 20:00:52,324+0300 INFO (libvirt/events) [virt.vm] (vmId='2e0eed87-5b81-4c2f-a439-c9fdd1b036c0') abnormal vm stop device ua-12d0b157-1409-4c8b-987b-06cc22c964b6 error enospc (vm:5051) 2019-03-30 20:00:52,324+0300 INFO (libvirt/events) [virt.vm] (vmId='2e0eed87-5b81-4c2f-a439-c9fdd1b036c0') CPU stopped: onIOError (vm:6029) 2019-03-30 20:00:52,328+0300 INFO (libvirt/events) [virt.vm] (vmId='2e0eed87-5b81-4c2f-a439-c9fdd1b036c0') Requesting extension for volume c8741dc2-72a9-4941-a4b5-3be664b05519 on domain 0d8d55a2-0e4e-43a2-b179-6530b1571ad0 (apparent: 1073741824, capacity: 2147483648, allocated: 1075118080, physical: 1073741824 threshold_state: exceeded) (vm:1262) 2019-03-30 20:00:52,329+0300 DEBUG (libvirt/events) [storage.TaskManager.Task] (Task='26d09cc3-c9b3-419c-a48a-54bb44d3f1c4') moving from state finished -> state preparing (task:602) 2019-03-30 20:00:52,329+0300 INFO (libvirt/events) [vdsm.api] START sendExtendMsg(spUUID='85ee87fb-fc8c-4e1e-8c53-9b26857c8e03', volDict={'newSize': 2147483648, 'domainID': '0d8d55a2-0e4e-43a2-b179-6530b1571ad0', 'name': 'vdb', 'poolID': '85ee87fb-fc8c-4e1e-8c53-9b26857c8e03', 'clock': <Clock(total=0.00*, extend-volume=0.00*)>, 'internal': False, 'volumeID': 'c8741dc2-72a9-4941-a4b5-3be664b05519', 'imageID': '12d0b157-1409-4c8b-987b-06cc22c964b6'}, newSize=2147483648, callbackFunc=<bound method Vm.__afterVolumeExtension of <vdsm.virt.vm.Vm object at 0x7f67d40cd550>>) from=internal, task_id=26d09cc3-c9b3-419c-a48a-54bb44d3f1c4 (api:48) Created attachment 1549893 [details]
4.3.3.1 logs
This bug/RFE is more than 2 years old and it didn't get enough attention so far, and is now flagged as pending close. Please review if it is still relevant and provide additional details/justification/patches if you believe it should get more attention for the next oVirt release. This bug didn't get any attention in a long time, and it's not planned in foreseeable future. oVirt development team has no plans to work on it. Please feel free to reopen if you have a plan how to contribute this feature/bug fix. |
Created attachment 1504383 [details] logs Description of problem: The automation test case failed while trying to perform dd command on the VM. From the art log: 2018-10-31 00:40:59,411 - MainThread - rhevmtests.storage.helpers - INFO - Performing command '/bin/dd bs=1M count=1358 if=/dev/vda of=/mount-point_3100405716/written_test_storage status=none' 2018-10-31 00:40:59,542 - MainThread - RemoteExecutor - DEBUG - [root.17.129/qum5net] Executing: /bin/dd bs=1M count=1358 if=/dev/vda of=/mount-point_3100405716/written_test_storage status=none 2018-10-31 00:46:59,854 - MainThread - RemoteExecutor - DEBUG - [root.17.129/qum5net] Results of command: /bin/dd bs=1M count=1358 if=/dev/vda of=/mount-point_3100405716/written_test_storage status=none 2018-10-31 00:46:59,858 - MainThread - RemoteExecutor - DEBUG - [root.17.129/qum5net] OUT: None 2018-10-31 00:46:59,858 - MainThread - RemoteExecutor - DEBUG - [root.17.129/qum5net] ERR: None 2018-10-31 00:46:59,858 - MainThread - RemoteExecutor - DEBUG - [root.17.129/qum5net] RC: None 2018-10-31 00:47:00,198 - MainThread - art.logging - ERROR - Status: failed *** Note that OUT & ERR are None. From engine log: 2018-10-31 00:41:03,980+02 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] (ForkJoinPool-1-worker-9) [] VM '37749548-6574-40ac-986b-ec89b85877f8'(vm_TestCase5063_3100372732) moved from 'Up' --> 'Paused' 2018-10-31 00:41:04,063+02 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ForkJoinPool-1-worker-9) [] EVENT_ID: VM_PAUSED(1,025), VM vm_TestCase5063_3100372732 has been paused. 2018-10-31 00:41:04,079+02 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ForkJoinPool-1-worker-9) [] EVENT_ID: VM_PAUSED_ENOSPC(138), VM vm_TestCase5063_3100372732 has been paused due to no Storage space error. 2018-10-31 00:41:18,655+02 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] (EE-ManagedThreadFactory-engineScheduled-Thread-44) [] VM '37749548-6574-40ac-986b-ec89b85877f8' was reported as Down on VDS '34ef902a-a22f-4aa3-83ec-82b2ebe534d4'(host_mixed_2) 2018-10-31 00:41:18,657+02 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] (EE-ManagedThreadFactory-engineScheduled-Thread-44) [] START, DestroyVDSCommand(HostName = host_mixed_2, DestroyVmVDSCommandParameters:{hostId='34ef902a-a22f-4aa3-83ec-82b2ebe534d4', vmId='37749548-6574-40ac-986b-ec89b85877f8', secondsToWait='0', gracefully='false', reason='', ignoreNoVm='true'}), log id: 34b65d10 2018-10-31 00:41:18,663+02 INFO [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] (EE-ManagedThreadFactory-engineScheduled-Thread-44) [] FINISH, DestroyVDSCommand, log id: 34b65d10 2018-10-31 00:41:18,663+02 INFO [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] (EE-ManagedThreadFactory-engineScheduled-Thread-44) [] VM '37749548-6574-40ac-986b-ec89b85877f8'(vm_TestCase5063_3100372732) moved from 'Paused' --> 'Down' 2018-10-31 00:41:18,745+02 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (EE-ManagedThreadFactory-engineScheduled-Thread-44) [] EVENT_ID: VM_DOWN_ERROR(119), VM vm_TestCase5063_3100372732 is down with error. Exit message: Lost connection with qemu process. From qemu log: 2018-10-30 22:41:06.745+0000: shutting down, reason=crashed Version-Release number of selected component (if applicable): ovirt-engine-4.2.7.4-0.1.el7ev.noarch vdsm-4.20.43-1.el7ev.x86_64 libvirt-4.5.0-10.el7_6.2.x86_64 qemu-img-rhev-2.12.0-18.el7_6.1.x86_64 How reproducible: Unknown. Didn't manage to reproduce it. Happened once during the analysis. Steps to Reproduce: 1. 2. 3. Actual results: The dd command failed Expected results: The operation should succeed. Additional info: Attached art, engine, vdsm and relevant qemu logs.