Bug 1330029 - 'Volume self heal info' continues to be 'OK' when heal is disabled for a volume
Summary: 'Volume self heal info' continues to be 'OK' when heal is disabled for a volume
Keywords:
Status: CLOSED NOTABUG
Alias: None
Product: Red Hat Gluster Storage
Classification: Red Hat
Component: nagios-server-addons
Version: rhgs-3.1
Hardware: Unspecified
OS: Unspecified
unspecified
medium
Target Milestone: ---
: ---
Assignee: Sahina Bose
QA Contact: Sweta Anandpara
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2016-04-25 10:16 UTC by Sweta Anandpara
Modified: 2016-04-26 09:22 UTC (History)
1 user (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2016-04-25 11:48:59 UTC
Target Upstream Version:


Attachments (Terms of Use)
screenshot of nagios UI (227.89 KB, image/png)
2016-04-25 10:16 UTC, Sweta Anandpara
no flags Details

Description Sweta Anandpara 2016-04-25 10:16:43 UTC
Created attachment 1150356 [details]
screenshot of nagios UI

Description of problem:
'Volume self-heal info' the option that is newly introduced in nagios web UI, shows the status/progress of self heal of a volume. When heal is disabled, it should ideally go to 'unknown' or better still, to a 'warning' with a note that healing is not taking place for that volume. 

Presently, it continues to remain in 'OK' state, giving the impression that everything is fine, even in the presence of unsynced entries in the volume. 

Version-Release number of selected component (if applicable):
Gluster-server 3.7.9-2
Nagios-server-addons 0.2.4-1

How reproducible: Always


Steps to Reproduce:
1. Have a 4 node setup, and a replica2/replica3/distribute-replicate volume 
2. Create a few files and verify that the 'volume heal info' stays 'green' in nagios web UI
3. Disable volume heal for the volume using the command 'gluster v heal <volname> disable'
4. Verify the 'volume self-heal info' attribute in the UI, and its state.

Actual results:
It continues to show 'OK'

Expected results:
'Volume self-heal info' should go to warning, with a note mentioning that self heal is disabled for the said volume

Additional info:

[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# gluster v list 
dist
dist-rep2
dist-rep3
rep2
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# gluster peer status
Number of Peers: 3

Hostname: 10.70.47.134
Uuid: fbc2256b-de25-49b2-a46a-b8d3c821b558
State: Peer in Cluster (Connected)

Hostname: 10.70.46.231
Uuid: e5cd7626-c7fa-4afe-a0d9-db38bc9b506e
State: Peer in Cluster (Connected)

Hostname: 10.70.47.131
Uuid: 27399a0b-06fa-4e3e-b270-9fc0884d126c
State: Peer in Cluster (Connected)
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# gluster v heal dist-rep3 info
Brick 10.70.47.116:/brick/brick2/dist-rep3
Number of entries: 0

Brick 10.70.47.131:/brick/brick3/dist-rep3
Number of entries: 0

Brick 10.70.46.231:/brick/brick3/dist-rep3
Number of entries: 0

Brick 10.70.47.116:/brick/brick1/dist-rep3
Number of entries: 0

Brick 10.70.47.131:/brick/brick4/dist-rep3
Number of entries: 0

Brick 10.70.46.231:/brick/brick4/dist-rep3
Number of entries: 0

[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# gluster v info dist-rep3
 
Volume Name: dist-rep3
Type: Distributed-Replicate
Volume ID: 8f152c8b-9fba-4cc2-9e07-a6dd1ee02c94
Status: Started
Number of Bricks: 2 x 3 = 6
Transport-type: tcp
Bricks:
Brick1: 10.70.47.116:/brick/brick2/dist-rep3
Brick2: 10.70.47.131:/brick/brick3/dist-rep3
Brick3: 10.70.46.231:/brick/brick3/dist-rep3
Brick4: 10.70.47.116:/brick/brick1/dist-rep3
Brick5: 10.70.47.131:/brick/brick4/dist-rep3
Brick6: 10.70.46.231:/brick/brick4/dist-rep3
Options Reconfigured:
cluster.self-heal-daemon: disable
features.soft-timeout: 0
features.quota-deem-statfs: on
features.inode-quota: on
features.quota: on
cluster.server-quorum-type: server
performance.readdir-ahead: on
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# rpm -qa | grep gluster
glusterfs-client-xlators-3.7.9-2.el6rhs.x86_64
glusterfs-cli-3.7.9-2.el6rhs.x86_64
gluster-nagios-common-0.2.4-1.el6rhs.noarch
gluster-nagios-addons-0.2.6-1.el6rhs.x86_64
glusterfs-libs-3.7.9-2.el6rhs.x86_64
glusterfs-3.7.9-2.el6rhs.x86_64
glusterfs-fuse-3.7.9-2.el6rhs.x86_64
glusterfs-server-3.7.9-2.el6rhs.x86_64
glusterfs-api-3.7.9-2.el6rhs.x86_64
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# 
[root@dhcp47-116 ~]# rpm -qa | grep nagios
nagios-common-3.5.1-6.el6.x86_64
gluster-nagios-common-0.2.4-1.el6rhs.noarch
gluster-nagios-addons-0.2.6-1.el6rhs.x86_64
nagios-plugins-1.4.16-12.el6rhs.x86_64
nagios-plugins-ping-1.4.16-12.el6rhs.x86_64
nagios-server-addons-0.2.4-1.el6rhs.noarch
nagios-plugins-procs-1.4.16-12.el6rhs.x86_64
nagios-plugins-dummy-1.4.16-12.el6rhs.x86_64
nagios-3.5.1-6.el6.x86_64
nagios-plugins-nrpe-2.15-4.2.el6rhs.x86_64
pnp4nagios-0.6.22-2.1.el6rhs.x86_64
nagios-plugins-ide_smart-1.4.16-12.el6rhs.x86_64
[root@dhcp47-116 ~]#

Comment 2 Sweta Anandpara 2016-04-25 11:48:59 UTC
As per discussion, 'Volume heal info' shows as per the cli 'gluster volume heal <volname> info', and the cli does detect the presence of unsynced entries, if there are any. The nagios UI service 'volume heal info' takes that info and reflects it in its state. 

Closing this as 'not a bug'.


Note You need to log in before you can comment on or make changes to this bug.