RHEL Engineering is moving the tracking of its product development work on RHEL 6 through RHEL 9 to Red Hat Jira (issues.redhat.com). If you're a Red Hat customer, please continue to file support cases via the Red Hat customer portal. If you're not, please head to the "RHEL project" in Red Hat Jira and file new tickets here. Individual Bugzilla bugs in the statuses "NEW", "ASSIGNED", and "POST" are being migrated throughout September 2023. Bugs of Red Hat partners with an assigned Engineering Partner Manager (EPM) are migrated in late September as per pre-agreed dates. Bugs against components "kernel", "kernel-rt", and "kpatch" are only migrated if still in "NEW" or "ASSIGNED". If you cannot log in to RH Jira, please consult article #7032570. That failing, please send an e-mail to the RH Jira admins at rh-issues@redhat.com to troubleshoot your issue as a user management inquiry. The email creates a ServiceNow ticket with Red Hat. Individual Bugzilla bugs that are migrated will be moved to status "CLOSED", resolution "MIGRATED", and set with "MigratedToJIRA" in "Keywords". The link to the successor Jira issue will be found under "Links", have a little "two-footprint" icon next to it, and direct you to the "RHEL project" in Red Hat Jira (issue links are of type "https://issues.redhat.com/browse/RHEL-XXXX", where "X" is a digit). This same link will be available in a blue banner at the top of the page informing you that that bug has been migrated.
Bug 896507 - 3.1 - engine: live snapshot fails due to race on multiple move of disks (live storage migration)
Summary: 3.1 - engine: live snapshot fails due to race on multiple move of disks (live...
Keywords:
Status: CLOSED DUPLICATE of bug 903248
Alias: None
Product: Red Hat Enterprise Linux 6
Classification: Red Hat
Component: vdsm
Version: 6.3
Hardware: x86_64
OS: Linux
high
high
Target Milestone: rc
: 6.4
Assignee: Eduardo Warszawski
QA Contact: Dafna Ron
URL:
Whiteboard: storage,
Depends On: 876558
Blocks:
TreeView+ depends on / blocked
 
Reported: 2013-01-17 13:21 UTC by Chris Pelland
Modified: 2013-02-28 07:01 UTC (History)
18 users (show)

Fixed In Version: vdsm-4.10.2-1.2.el6
Doc Type: Bug Fix
Doc Text:
Previously, live snapshotting failed because of a race condition that existed when you tried to move virtual machine disks between storage domains. This error occurred when a host interrupted the block volume creation process between the lvcreate step and the lvchange step. A patch has been introduced which adds an init tag in lvcreate so that other hosts can identify logical volumes engaged in the creation process as "partial". When identified as "partial", hosts ignore these logical volumes. This eliminates the race condition that caused live snapshotting to fail.
Clone Of: 876558
Environment:
Last Closed: 2013-02-07 15:09:14 UTC
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)
logs (1.53 MB, application/x-gzip)
2013-01-29 14:01 UTC, Dafna Ron
no flags Details


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHBA-2013:0555 0 normal SHIPPED_LIVE rhev 3.1.z - vdsm bug fix update 2013-02-26 00:06:22 UTC

Comment 4 Dafna Ron 2013-01-29 14:00:26 UTC
tested on vdsm-4.10.2-1.2.el6.x86_64

we are failing to create snapshot with: 

Thread-4533::ERROR::2013-01-29 13:22:10,610::libvirtvm::2197::vm.Vm::(diskReplicateFinish) vmId=`cc5c2485-e7ac-4094-9d30-b2c819b8430b`::Unable to stop the replication for the drive: vda
Traceback (most recent call last):
  File "/usr/share/vdsm/libvirtvm.py", line 2194, in diskReplicateFinish
    self._dom.blockJobAbort(srcDrive.name, blockJobFlags)
  File "/usr/share/vdsm/libvirtvm.py", line 515, in f
    ret = attr(*args, **kwargs)
  File "/usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py", line 83, in wrapper
    ret = f(*args, **kwargs)
  File "/usr/lib64/python2.6/site-packages/libvirt.py", line 512, in blockJobAbort
    if ret == -1: raise libvirtError ('virDomainBlockJobAbort() failed', dom=self)
libvirtError: internal error unable to execute QEMU command '__com.redhat_drive-reopen': Could not open '/rhev/data-center/f1c4c67b-0647-40c7-975a-837777656129/a68a5097-3032-4e99-bc43-fdebf4f46df2/images/83dd02e0-2e19-4179-827d-7d49cded
ccb5/3b6d4f67-594a-4736-87e0-aed140a8ef5c': Operation not permitted


full logs will be attached

Comment 5 Dafna Ron 2013-01-29 14:01:02 UTC
Created attachment 689846 [details]
logs

Comment 6 Dafna Ron 2013-01-29 14:03:28 UTC
I reproduced on by creating 30 vm's (using pool and detaching the vms). 
the vm's were on 2 iscsi storage domains and I live migrated all the disks (selecting each disk -> move) to a 3ed domain. 

moving back to devel with all logs

Comment 7 Ayal Baron 2013-02-07 12:41:08 UTC
(In reply to comment #4)
> tested on vdsm-4.10.2-1.2.el6.x86_64
> 
> we are failing to create snapshot with: 
> 
> Thread-4533::ERROR::2013-01-29
> 13:22:10,610::libvirtvm::2197::vm.Vm::(diskReplicateFinish)
> vmId=`cc5c2485-e7ac-4094-9d30-b2c819b8430b`::Unable to stop the replication
> for the drive: vda
> Traceback (most recent call last):
>   File "/usr/share/vdsm/libvirtvm.py", line 2194, in diskReplicateFinish
>     self._dom.blockJobAbort(srcDrive.name, blockJobFlags)
>   File "/usr/share/vdsm/libvirtvm.py", line 515, in f
>     ret = attr(*args, **kwargs)
>   File "/usr/lib64/python2.6/site-packages/vdsm/libvirtconnection.py", line
> 83, in wrapper
>     ret = f(*args, **kwargs)
>   File "/usr/lib64/python2.6/site-packages/libvirt.py", line 512, in
> blockJobAbort
>     if ret == -1: raise libvirtError ('virDomainBlockJobAbort() failed',
> dom=self)
> libvirtError: internal error unable to execute QEMU command
> '__com.redhat_drive-reopen': Could not open
> '/rhev/data-center/f1c4c67b-0647-40c7-975a-837777656129/a68a5097-3032-4e99-
> bc43-fdebf4f46df2/images/83dd02e0-2e19-4179-827d-7d49cded
> ccb5/3b6d4f67-594a-4736-87e0-aed140a8ef5c': Operation not permitted
> 
> 
> full logs will be attached

Fede, isn't this issue a dup of the libvirt bug?

Comment 8 Federico Simoncelli 2013-02-07 14:54:32 UTC
(In reply to comment #7)
> (In reply to comment #4)
> > libvirtError: internal error unable to execute QEMU command
> > '__com.redhat_drive-reopen': Could not open
> > '/rhev/data-center/f1c4c67b-0647-40c7-975a-837777656129/a68a5097-3032-4e99-
> > bc43-fdebf4f46df2/images/83dd02e0-2e19-4179-827d-7d49cded
> > ccb5/3b6d4f67-594a-4736-87e0-aed140a8ef5c': Operation not permitted
> > 
> > 
> > full logs will be attached
> 
> Fede, isn't this issue a dup of the libvirt bug?

Yes it looks like a duplicate of bug 903248

Comment 9 Ayal Baron 2013-02-07 15:09:14 UTC

*** This bug has been marked as a duplicate of bug 903248 ***


Note You need to log in before you can comment on or make changes to this bug.