Bug 1684266
| Summary: | Exporting OVA timed out leaving orphan volume | ||
|---|---|---|---|
| Product: | Red Hat Enterprise Virtualization Manager | Reporter: | Javier Coscia <jcoscia> |
| Component: | vdsm | Assignee: | Shmuel Melamud <smelamud> |
| Status: | CLOSED ERRATA | QA Contact: | Nisim Simsolo <nsimsolo> |
| Severity: | medium | Docs Contact: | |
| Priority: | unspecified | ||
| Version: | 4.2.7 | CC: | aefrat, ahadas, frolland, lsurette, lsvaty, michal.skrivanek, nsimsolo, ovirt, rdlugyhe, smelamud, srevivo, swachira, tnisan, ycui |
| Target Milestone: | ovirt-4.4.0 | Flags: | lsvaty:
testing_plan_complete-
|
| Target Release: | --- | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | rhv-4.4.0-29 | Doc Type: | Bug Fix |
| Doc Text: |
When a large disk is converted as part of VM export to OVA, it takes a long time. Previously, the SSH channel the export script timed out and closed due to the long period of inactivity, leaving an orphan volume. The current release fixes this issue: Now, the export script adds some traffic to the SSH channel during disk conversion to prevent the SSH channel from being closed.
|
Story Points: | --- |
| Clone Of: | Environment: | ||
| Last Closed: | 2020-08-04 13:26:25 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | Virt | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | 1825638 | ||
| Bug Blocks: | |||
|
Description
Javier Coscia
2019-02-28 20:24:28 UTC
Arik, can you please have a look? Javier, do we know if the VM was running? It's not clear from the logs (doesn't start early enough) So, we can increase the timeout here, but it's not a guarantee that the guest disks are unlocked. Tal - since the VM is down, we'll need help from storage to investigate why the LVs were locked. Any ideas? (In reply to Ryan Barry from comment #8) > So, we can increase the timeout here, but it's not a guarantee that the > guest disks are unlocked. Tal - since the VM is down, we'll need help from > storage to investigate why the LVs were locked. Any ideas? I suppose that unlocking the LV failed because qemu-img process was still running and using the disk. Only the SSH connection was closed by timeout. I thought this also, but comment#6 looks like it's down (I'm waiting to get the rest of the engine log to see if something else brought it back up) (In reply to Ryan Barry from comment #10) > I thought this also, but comment#6 looks like it's down (I'm waiting to get > the rest of the engine log to see if something else brought it back up) 2019-02-08 - it is a week before the issue. Why we need to take this into account? The engine.log in attachment 1539639 [details] starts from 2019-02-09 and the Engine is running at that moment. That's why you're on the bug ;) Because I didn't find entries in the current engine log to indicate whether the VM which was being exported was up or down. If it was down a week beforehand, the qemu-img process shouldn't still be running and using the disk unless I've missed something in the logs (In reply to Ryan Barry from comment #12) > Because I didn't find entries in the current engine log to indicate whether > the VM which was being exported was up or down. If it was down a week > beforehand, the qemu-img process shouldn't still be running and using the > disk unless I've missed something in the logs The qemu-img process is not related to the VM running. During the export, pack_ova.py script executes qemu-img to convert disks. Shmuel, what are the next steps? I am currently looking for an easy way to populate SSH channel with some traffic during the conversion process and verifying that it does not break the functionality. I'll post the patch soon. Moving to Virt since you are looking into the solution Since we have a patch, can you please target this bug? 4.4? 4.3.5? sync2jira sync2jira WARN: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops
WARN: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops
WARN: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops
WARN: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops
WARN: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops: Bug status (ON_QA) wasn't changed but the folowing should be fixed:
[Found non-acked flags: '{}', ]
For more info please contact: rhv-devops
Verified: ovirt-engine-4.4.1.1-0.5.el8ev.noarch vdsm-4.40.17-1.el8ev.x86_64 libvirt-daemon-6.0.0-22.module+el8.2.1+6815+1c792dc8.x86_64 qemu-kvm-4.2.0-22.module+el8.2.1+6758+cb8d64c2.x86_64 Verification scenario: 1. Create VM with 200GB preallocated NFS disk. Install latest RHEL8 OS on it and verify OS is running properly. 2. Export OVA to host NFS mount (use NFS in order to make export time to take longer). Verify OVA exported successfully and took more than 30 minutes (it took 43:51 minutes to export it). 3. Import exported OVA (Import time was 2:26 minutes) 4. Run imported VM. Verify OS is running properly and disk size is 200GB with thin provision allocation policy. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (RHV RHEL Host (ovirt-host) 4.4), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2020:3246 |