Description of problem: When a snapshot is crated on top of a thin provisioned disk and the disk is then extended the live storage migration fails when syncing the images between the storage domains. Version-Release number of selected component (if applicable): vdsm-4.19.36-1.el7ev vdsm-4.17.43-1.el7ev How reproducible: 100% Steps to Reproduce: 1. Create a VM with thin provisioned disk on block storage domain. 2. Create a snapshot. 3. Live migrate the disk to another block storage domain. Actual results: 46524dd8-8079-4b8e-9dc4-377414e86480::ERROR::2017-12-04 01:24:15,005::image::464::Storage.Image::(_interImagesCopy) Copy image error: image=eb2241b7-a2d3-4047-b865-570176de709c, src domain=b30263e4-49d8-468b-9c71-d68d5f8d79e2, dst domain=061d3f67-c803-4a02-8a4b-06d6bc1802be Traceback (most recent call last): File "/usr/share/vdsm/storage/image.py", line 455, in _interImagesCopy backingFormat=backingFormat) File "/usr/lib/python2.7/site-packages/vdsm/qemuimg.py", line 207, in convert raise QImgError(rc, out, err) QImgError: ecode=1, stdout=[], stderr=['qemu-img: error while writing sector 276866432: No space left on device', 'qemu-img: Failed to flush the refcount block cache: No space left on device'], message=None Expected results: The clone operation works Additional info: The command triggering this issue is: /usr/bin/qemu-img convert ... -f qcow2 /rhev/data-center/mnt/blockSD/b30263e4-49d8-468b-9c71-d68d5f8d79e2/images/eb2241b7-a2d3-4047-b865-570176de709c/0a795045-6a48-41f9-8a3f-b6fa25682f98 -O qcow2 -o compat=0.10,backing_file=9dd0fb65-1881-4fea-bc7c-498ef367f1f8,backing_fmt=qcow2 /rhev/data-center/mnt/blockSD/061d3f67-c803-4a02-8a4b-06d6bc1802be/images/eb2241b7-a2d3-4047-b865-570176de709c/0a795045-6a48-41f9-8a3f-b6fa25682f98 (cwd /rhev/data-center/mnt/blockSD/b30263e4-49d8-468b-9c71-d68d5f8d79e2/images/eb2241b7-a2d3-4047-b865-570176de709c) The volume and its parent is 0a795045-6a48-41f9-8a3f-b6fa25682f98 b30263e4-49d8-468b-9c71-d68d5f8d79e2 106.00g IU_eb2241b7-a2d3-4047-b865-570176de709c,MD_5,PU_9dd0fb65-1881-4fea-bc7c-498ef367f1f8 9dd0fb65-1881-4fea-bc7c-498ef367f1f8 b30263e4-49d8-468b-9c71-d68d5f8d79e2 65.00g IU_eb2241b7-a2d3-4047-b865-570176de709c,MD_4,PU_00000000-0000-0000-0000-000000000000 The volumes on the destination are created inthe following way: ... lvcreate ... --size 1024m --addtag OVIRT_VOL_INITIALIZING --name 9dd0fb65-1881-4fea-bc7c-498ef367f1f8 061d3f67-c803-4a02-8a4b-06d6bc1802be (cwd None) .. lvcreate ... --size 1024m --addtag OVIRT_VOL_INITIALIZING --name 0a795045-6a48-41f9-8a3f-b6fa25682f98 061d3f67-c803-4a02-8a4b-06d6bc1802be (cwd None) ... ... lvextend ... --autobackup n --size 66560m 061d3f67-c803-4a02-8a4b-06d6bc1802be/9dd0fb65-1881-4fea-bc7c-498ef367f1f8 (cwd None) ... lvextend ... --autobackup n --size 108544m 061d3f67-c803-4a02-8a4b-06d6bc1802be/0a795045-6a48-41f9-8a3f-b6fa25682f98 (cwd None) The sizes are matching. The qcow2 structure is created in this way: ... /usr/bin/qemu-img create -f qcow2 -o compat=0.10 /rhev/data-center/c36e24b4-16e5-45ca-838d-dd054563401e/061d3f67-c803-4a02-8a4b-06d6bc1802be/images/eb2241b7-a2d3-4047-b865-570176de709c/9dd0fb65-1881-4fea-bc7c-498ef367f1f8 68719476736 (cwd None) ... /usr/bin/qemu-img create -f qcow2 -o compat=0.10 -b 9dd0fb65-1881-4fea-bc7c-498ef367f1f8 -F qcow2 /rhev/data-center/c36e24b4-16e5-45ca-838d-dd054563401e/061d3f67-c803-4a02-8a4b-06d6bc1802be/images/eb2241b7-a2d3-4047-b865-570176de709c/0a795045-6a48-41f9-8a3f-b6fa25682f98 (cwd /rhev/data-center/c36e24b4-16e5-45ca-838d-dd054563401e/061d3f67-c803-4a02-8a4b-06d6bc1802be/images/eb2241b7-a2d3-4047-b865-570176de709c) Ten we convert the parent (successful) ... /usr/bin/qemu-img convert -t none -T none -f qcow2 /rhev/data-center/mnt/blockSD/b30263e4-49d8-468b-9c71-d68d5f8d79e2/images/eb2241b7-a2d3-4047-b865-570176de709c/9dd0fb65-1881-4fea-bc7c-498ef367f1f8 -O qcow2 -o compat=0.10 /rhev/data-center/mnt/blockSD/061d3f67-c803-4a02-8a4b-06d6bc1802be/images/eb2241b7-a2d3-4047-b865-570176de709c/9dd0fb65-1881-4fea-bc7c-498ef367f1f8 (cwd None) and then its leaf fails with the mentioned error: ... /usr/bin/qemu-img convert ... -f qcow2 /rhev/data-center/mnt/blockSD/b30263e4-49d8-468b-9c71-d68d5f8d79e2/images/eb2241b7-a2d3-4047-b865-570176de709c/0a795045-6a48-41f9-8a3f-b6fa25682f98 -O qcow2 -o compat=0.10,backing_file=9dd0fb65-1881-4fea-bc7c-498ef367f1f8,backing_fmt=qcow2 /rhev/data-center/mnt/blockSD/061d3f67-c803-4a02-8a4b-06d6bc1802be/images/eb2241b7-a2d3-4047-b865-570176de709c/0a795045-6a48-41f9-8a3f-b6fa25682f98 (cwd /rhev/data-center/mnt/blockSD/b30263e4-49d8-468b-9c71-d68d5f8d79e2/images/eb2241b7-a2d3-4047-b865-570176de709c)
Roman, can you share the qemu-kvm-rhev versions you're using?
(In reply to Allon Mureinik from comment #3) > Roman, can you share the qemu-kvm-rhev versions you're using? qemu-img-rhev, of course.
One more, with ELS: vdsm-4.17.35-1.el7ev.noarch qemu-kvm-rhev-2.3.0-31.el7_2.21.x86_64
Just to clarify our action plan: There is a real gap in qemu-img[-[rh]ev]]. Fixing it is tracked by bug 1527898. From RHV's side, we need two BZs: 1. A dependency bump to consume the fix for bug 1527898 once its ready 2. An actual handling in RHV which will recognize a situation where we know the copying will fail and block it with a validation message that instructs the user which snapshot(s) he should merge in order to allow the copying to succeed.
Benny - the included patch clearly handles live and cold move flows. Does it also cover importing [without collapse] to a V3 block storage domain? (if it doesn't, please open another BZ on it and link back here). Regardless, this is a non-obvious patch - please add some doctext to explain it.
As discussed offline, this issue does not apply to the import scenario as the LV created large enough
INFO: Bug status (ON_QA) wasn't changed but the folowing should be fixed: [Project 'ovirt-engine'/Component 'vdsm' mismatch] For more info please contact: rhv-devops
(In reply to RHV Bugzilla Automation and Verification Bot from comment #11) > INFO: Bug status (ON_QA) wasn't changed but the folowing should be fixed: > > [Project 'ovirt-engine'/Component 'vdsm' mismatch] > > For more info please contact: rhv-devops Moving to enigne.
Verified with the following code: -------------------------------------- ovirt-engine-4.2.1.2-0.1.el7.noarch vdsm-4.20.14-1.el7ev.x86_64 Verified with the following scenario: -------------------------------------- Steps to Reproduce: 1. Create a VM with thin provisioned disk on block storage domain. 2. Create a snapshot. 3. Live migrate the disk to another block storage domain. Actual results: Live migration is successful. No errors reported Moving to VERIFIED
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHEA-2018:1488
BZ<2>Jira Resync
sync2jira