Bug 1245923 - [Snapshot] Scheduler should check vol-name exists or not before adding scheduled jobs
Summary: [Snapshot] Scheduler should check vol-name exists or not before adding sched...
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: GlusterFS
Classification: Community
Component: snapshot
Version: 3.7.3
Hardware: x86_64
OS: Linux
unspecified
urgent
Target Milestone: ---
Assignee: Avra Sengupta
QA Contact:
URL:
Whiteboard:
Depends On: 1213349 1245924
Blocks: qe_tracker_everglades
TreeView+ depends on / blocked
 
Reported: 2015-07-23 06:48 UTC by Avra Sengupta
Modified: 2015-09-09 09:38 UTC (History)
4 users (show)

Fixed In Version: glusterfs-3.7.4
Clone Of: 1213349
Environment:
Last Closed: 2015-09-09 09:38:24 UTC
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Embargoed:


Attachments (Terms of Use)

Description Avra Sengupta 2015-07-23 06:48:14 UTC
+++ This bug was initially created as a clone of Bug #1213349 +++

Description of problem:

When adding jobs to scheduler,  it should check whether volume with that volume name exist or not. If volume does not exists, it through message volume volume doesn't exists.   

Version-Release number of selected component (if applicable):

[root@localhost core]# rpm -qa | grep glusterfs
glusterfs-api-3.8dev-0.12.gitaa87c31.el6.x86_64
glusterfs-geo-replication-3.8dev-0.12.gitaa87c31.el6.x86_64
samba-glusterfs-3.6.509-169.4.el6rhs.x86_64
glusterfs-cli-3.8dev-0.12.gitaa87c31.el6.x86_64
glusterfs-fuse-3.8dev-0.12.gitaa87c31.el6.x86_64
glusterfs-server-3.8dev-0.12.gitaa87c31.el6.x86_64
glusterfs-rdma-3.8dev-0.12.gitaa87c31.el6.x86_64
glusterfs-debuginfo-3.7dev-0.994.gitf522001.el6.x86_64
glusterfs-libs-3.8dev-0.12.gitaa87c31.el6.x86_64
glusterfs-3.8dev-0.12.gitaa87c31.el6.x86_64


How reproducible:

100%

Steps to Reproduce:
1. Create 6*2 distributed replicate volume
2. create shared storage and do fuse mount on each storage node on path /var/run/gluster/shared_storage
mount -t glusterfs 10.70.47.143:meta /var/run/gluster/shared_storage/
3. initialize scheduler on each storage node e.g run snap_scheduler.py init  command
4. Enable scheduler on storage nodes e.g run snap_scheduler.py enable
5. Add jobs to scheduler providing volume name which doesn't exists


Actual results:

scheduler accepts the jobs.

Expected results:

Scheduler should check whether volume exists or not.

Additional info:

Comment 1 Avra Sengupta 2015-08-14 08:36:07 UTC
Fixed with http://review.gluster.org/#/c/11917/

Comment 2 Kaushal 2015-09-09 09:38:24 UTC
This bug is getting closed because a release has been made available that should address the reported issue. In case the problem is still not fixed with glusterfs-3.7.4, please open a new bug report.

glusterfs-3.7.4 has been announced on the Gluster mailinglists [1], packages for several distributions should become available in the near future. Keep an eye on the Gluster Users mailinglist [2] and the update infrastructure for your distribution.

[1] http://thread.gmane.org/gmane.comp.file-systems.gluster.devel/12496
[2] http://thread.gmane.org/gmane.comp.file-systems.gluster.user


Note You need to log in before you can comment on or make changes to this bug.