Bug 856203 - [engine-core] During deleting action multiple floating disks and restart vdsm service action at SPM server some disks get locked status
Summary: [engine-core] During deleting action multiple floating disks and restart vdsm...
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: Red Hat Enterprise Virtualization Manager
Classification: Red Hat
Component: ovirt-engine
Version: 3.1.0
Hardware: x86_64
OS: Linux
unspecified
high
Target Milestone: ---
: 3.2.0
Assignee: Tal Nisan
QA Contact: Elad
URL:
Whiteboard: storage
: 856135 866886 (view as bug list)
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2012-09-11 12:39 UTC by vvyazmin@redhat.com
Modified: 2016-02-10 17:35 UTC (History)
13 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed:
oVirt Team: Storage
Target Upstream Version:
Embargoed:
scohen: Triaged+


Attachments (Terms of Use)
## Logs vdsm, rhevm (1.18 MB, application/x-gzip)
2012-09-11 12:39 UTC, vvyazmin@redhat.com
no flags Details

Description vvyazmin@redhat.com 2012-09-11 12:39:11 UTC
Created attachment 611756 [details]
## Logs vdsm, rhevm

Description of problem: 
During deleting action multiple floating disks and restart vdsm service action, some disks get "locked" status


Version-Release number of selected component (if applicable):
RHEVM 3.1 - SI17 

RHEVM: rhevm-3.1.0-15.el6ev.noarch 
VDSM: vdsm-4.9.6-32.0.el6_3.x86_64 
LIBVIRT: libvirt-0.9.10-21.el6_3.4.x86_64 
QEMU & KVM: qemu-kvm-rhev-0.12.1.2-2.295.el6_3.2.x86_64 
SANLOCK: sanlock-2.3-3.el6_3.x86_64

How reproducible:
Occasionally

Steps to Reproduce:
1. Create iSCSI DC with 2 hosts
2. Create 12 floating disk
3. Select them all, and delete
4. During deleting restart “vdsm” service on SPM server (run command #: service vdsmd stop && service vdsmd start)
  
Actual results:
1. Same disk get “Locked” status
2. No option delete them
3. In DB deleting task is exist (see attached log)
4. No tasks are running on both hosts (run command #: vdsClient -s 0 getAllTasksInfo)

Expected results:
No “Locked” disks
Option remove (delete them) them 
Clean task from DB

Additional info:
In DB “async_tasks” table: “delete” task are running
In VDSM (vdsClient -s 0 getAllTasksInfo) : no tasks are running
In DB “all_disks” table, disk have status: imagestatus == 2

Comment 1 vvyazmin@redhat.com 2012-09-11 12:41:52 UTC
psql -U postgres engine -c 'select disk_id,disk_alias,imagestatus  from all_disks where imagestatus = 2;'  | less -S 

               disk_id                | disk_alias | imagestatus 
--------------------------------------+------------+-------------
 efad1726-9da3-4937-a4f2-8e9f2e9ed37b | A-02       |           2

Comment 2 vvyazmin@redhat.com 2012-09-12 07:54:57 UTC
After 15 hours, “delete” task still running, and not released

Comment 3 Ayal Baron 2012-09-12 08:19:33 UTC
(In reply to comment #2)
> After 15 hours, “delete” task still running, and not released

the task reaper runs after 30 or 50 hours, I don't recall which.

Comment 4 Dafna Ron 2012-09-12 13:50:10 UTC
for dead tasks in vdsm it will run after 50-60 hours.
for engine db clean up it should be about 5 hours.

Comment 5 vvyazmin@redhat.com 2012-10-09 12:15:23 UTC
*** Bug 856135 has been marked as a duplicate of this bug. ***

Comment 6 vvyazmin@redhat.com 2012-10-09 12:20:42 UTC
On verification this bug please run scenario from BZ856135

Comment 7 Haim 2012-10-16 10:12:37 UTC
*** Bug 866886 has been marked as a duplicate of this bug. ***

Comment 8 Tal Nisan 2013-03-11 10:51:43 UTC
Could not reproduce, this patch seems to solve it:
http://gerrit.ovirt.org/#/c/11075/

Comment 9 Elad 2013-03-14 14:44:50 UTC
Verified on SF10. When reproduced, vdsm has been restarted during the removing of the 12 disks. The disks became 'illegal' and then I was manage to remove them.

Comment 10 Itamar Heim 2013-06-11 08:32:54 UTC
3.2 has been released

Comment 11 Itamar Heim 2013-06-11 08:32:58 UTC
3.2 has been released

Comment 12 Itamar Heim 2013-06-11 08:33:50 UTC
3.2 has been released

Comment 13 Itamar Heim 2013-06-11 08:42:21 UTC
3.2 has been released


Note You need to log in before you can comment on or make changes to this bug.