| Summary: | 'Volume self heal info' continues to be 'OK' when heal is disabled for a volume | ||||||
|---|---|---|---|---|---|---|---|
| Product: | Red Hat Gluster Storage | Reporter: | Sweta Anandpara <sanandpa> | ||||
| Component: | nagios-server-addons | Assignee: | Sahina Bose <sabose> | ||||
| Status: | CLOSED NOTABUG | QA Contact: | Sweta Anandpara <sanandpa> | ||||
| Severity: | medium | Docs Contact: | |||||
| Priority: | unspecified | ||||||
| Version: | rhgs-3.1 | CC: | rhinduja | ||||
| Target Milestone: | --- | Keywords: | ZStream | ||||
| Target Release: | --- | ||||||
| Hardware: | Unspecified | ||||||
| OS: | Unspecified | ||||||
| Whiteboard: | |||||||
| Fixed In Version: | Doc Type: | Bug Fix | |||||
| Doc Text: | Story Points: | --- | |||||
| Clone Of: | Environment: | ||||||
| Last Closed: | 2016-04-25 11:48:59 UTC | Type: | Bug | ||||
| Regression: | --- | Mount Type: | --- | ||||
| Documentation: | --- | CRM: | |||||
| Verified Versions: | Category: | --- | |||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||
| Attachments: |
|
||||||
As per discussion, 'Volume heal info' shows as per the cli 'gluster volume heal <volname> info', and the cli does detect the presence of unsynced entries, if there are any. The nagios UI service 'volume heal info' takes that info and reflects it in its state. Closing this as 'not a bug'. |
Created attachment 1150356 [details] screenshot of nagios UI Description of problem: 'Volume self-heal info' the option that is newly introduced in nagios web UI, shows the status/progress of self heal of a volume. When heal is disabled, it should ideally go to 'unknown' or better still, to a 'warning' with a note that healing is not taking place for that volume. Presently, it continues to remain in 'OK' state, giving the impression that everything is fine, even in the presence of unsynced entries in the volume. Version-Release number of selected component (if applicable): Gluster-server 3.7.9-2 Nagios-server-addons 0.2.4-1 How reproducible: Always Steps to Reproduce: 1. Have a 4 node setup, and a replica2/replica3/distribute-replicate volume 2. Create a few files and verify that the 'volume heal info' stays 'green' in nagios web UI 3. Disable volume heal for the volume using the command 'gluster v heal <volname> disable' 4. Verify the 'volume self-heal info' attribute in the UI, and its state. Actual results: It continues to show 'OK' Expected results: 'Volume self-heal info' should go to warning, with a note mentioning that self heal is disabled for the said volume Additional info: [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# gluster v list dist dist-rep2 dist-rep3 rep2 [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# gluster peer status Number of Peers: 3 Hostname: 10.70.47.134 Uuid: fbc2256b-de25-49b2-a46a-b8d3c821b558 State: Peer in Cluster (Connected) Hostname: 10.70.46.231 Uuid: e5cd7626-c7fa-4afe-a0d9-db38bc9b506e State: Peer in Cluster (Connected) Hostname: 10.70.47.131 Uuid: 27399a0b-06fa-4e3e-b270-9fc0884d126c State: Peer in Cluster (Connected) [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# gluster v heal dist-rep3 info Brick 10.70.47.116:/brick/brick2/dist-rep3 Number of entries: 0 Brick 10.70.47.131:/brick/brick3/dist-rep3 Number of entries: 0 Brick 10.70.46.231:/brick/brick3/dist-rep3 Number of entries: 0 Brick 10.70.47.116:/brick/brick1/dist-rep3 Number of entries: 0 Brick 10.70.47.131:/brick/brick4/dist-rep3 Number of entries: 0 Brick 10.70.46.231:/brick/brick4/dist-rep3 Number of entries: 0 [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# gluster v info dist-rep3 Volume Name: dist-rep3 Type: Distributed-Replicate Volume ID: 8f152c8b-9fba-4cc2-9e07-a6dd1ee02c94 Status: Started Number of Bricks: 2 x 3 = 6 Transport-type: tcp Bricks: Brick1: 10.70.47.116:/brick/brick2/dist-rep3 Brick2: 10.70.47.131:/brick/brick3/dist-rep3 Brick3: 10.70.46.231:/brick/brick3/dist-rep3 Brick4: 10.70.47.116:/brick/brick1/dist-rep3 Brick5: 10.70.47.131:/brick/brick4/dist-rep3 Brick6: 10.70.46.231:/brick/brick4/dist-rep3 Options Reconfigured: cluster.self-heal-daemon: disable features.soft-timeout: 0 features.quota-deem-statfs: on features.inode-quota: on features.quota: on cluster.server-quorum-type: server performance.readdir-ahead: on [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# rpm -qa | grep gluster glusterfs-client-xlators-3.7.9-2.el6rhs.x86_64 glusterfs-cli-3.7.9-2.el6rhs.x86_64 gluster-nagios-common-0.2.4-1.el6rhs.noarch gluster-nagios-addons-0.2.6-1.el6rhs.x86_64 glusterfs-libs-3.7.9-2.el6rhs.x86_64 glusterfs-3.7.9-2.el6rhs.x86_64 glusterfs-fuse-3.7.9-2.el6rhs.x86_64 glusterfs-server-3.7.9-2.el6rhs.x86_64 glusterfs-api-3.7.9-2.el6rhs.x86_64 [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# [root@dhcp47-116 ~]# rpm -qa | grep nagios nagios-common-3.5.1-6.el6.x86_64 gluster-nagios-common-0.2.4-1.el6rhs.noarch gluster-nagios-addons-0.2.6-1.el6rhs.x86_64 nagios-plugins-1.4.16-12.el6rhs.x86_64 nagios-plugins-ping-1.4.16-12.el6rhs.x86_64 nagios-server-addons-0.2.4-1.el6rhs.noarch nagios-plugins-procs-1.4.16-12.el6rhs.x86_64 nagios-plugins-dummy-1.4.16-12.el6rhs.x86_64 nagios-3.5.1-6.el6.x86_64 nagios-plugins-nrpe-2.15-4.2.el6rhs.x86_64 pnp4nagios-0.6.22-2.1.el6rhs.x86_64 nagios-plugins-ide_smart-1.4.16-12.el6rhs.x86_64 [root@dhcp47-116 ~]#