Bugzilla will be upgraded to version 5.0. The upgrade date is tentatively scheduled for 2 December 2018, pending final testing and feedback.
Bug 1089636 - [Nagios] Cluster status information says "None of the volumes are in critical state" even when a volume is utilized beyond critical level. [NEEDINFO]
[Nagios] Cluster status information says "None of the volumes are in critical...
Status: CLOSED CANTFIX
Product: Red Hat Gluster Storage
Classification: Red Hat
Component: gluster-nagios-addons (Show other bugs)
3.0
Unspecified Unspecified
medium Severity medium
: ---
: ---
Assigned To: Nishanth Thomas
RHS-C QE
: ZStream
Depends On:
Blocks: 1087818
  Show dependency treegraph
 
Reported: 2014-04-21 05:55 EDT by Shruti Sampat
Modified: 2018-01-30 06:12 EST (History)
4 users (show)

See Also:
Fixed In Version:
Doc Type: Known Issue
Doc Text:
In the Nagios UI, incorrect status information is displayed as "Cluster Status OK : None of the Volumes are in Critical State", when volume is utilized beyond critical level.
Story Points: ---
Clone Of:
Environment:
Last Closed: 2018-01-30 06:12:16 EST
Type: Bug
Regression: ---
Mount Type: ---
Documentation: ---
CRM:
Verified Versions:
Category: ---
oVirt Team: ---
RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: ---
kmayilsa: needinfo? (nthomas)


Attachments (Terms of Use)

  None (edit)
Description Shruti Sampat 2014-04-21 05:55:01 EDT
Description of problem:
-----------------------

Consider a cluster having a single volume, being monitored be Nagios. When the volume is utilized beyond the critical level, the volume utilization service reports the volume to be critical. The "host" that represents the cluster in the Nagios UI, reports "Cluster Status OK : None of the Volumes are in Critical State" as part of its status information.

Version-Release number of selected component (if applicable):
gluster-nagios-addons-0.1.0-25.git25f0bba.el6.x86_64

How reproducible:
Saw it once.

Steps to Reproduce:
1. Create a cluster of RHS nodes, and create a volume, start it and fill the volume with data, such that it crosses 90% capacity.
2. Configure nagios server to run on one of the RHS nodes, and configure this cluster to be monitored by nagios.
3. Check the volume utilization service for the volume created, it should show critical status.
4. Check the state information for the "host" that represents the cluster in the Nagios UI.

Actual results:
The status information for the host says "Cluster Status OK : None of the Volumes are in Critical State".

Expected results:
The status information should indicate that one of the volumes in the cluster is utilized above critical level.

Additional info:
Comment 1 Shruti Sampat 2014-04-21 06:52:16 EDT
The performance data for the cluster also shows the number of volumes in critical state as zero - 

noOfVolumesInCriticalState=0
Comment 2 Nishanth Thomas 2014-05-08 05:23:23 EDT
As per the current design, cluster status is determined from the status of 'Volume Status' services of all the volumes under that cluster. The 'Cluster Utilization' service will reflect if any of volume's utilization goes beyond critical level. When the cascading gets implemented(in future) the cluster utilization service's will get propagated to cluster status

This is inline with the implementation of volume Status as well. Similarly we don't consider the Brick Utilization while determining the Volume status.

So as per my opinion this is not a bug and works as per the design
Comment 3 Dusmant 2014-05-16 07:26:40 EDT
As discussed in the triage meeting : we will take it up in the 3.1 release
Comment 4 Shalaka 2014-06-26 10:43:37 EDT
Please review and signoff edited doc text.
Comment 5 Nishanth Thomas 2014-06-27 01:12:56 EDT
Doctext is fine
Comment 6 Kanagaraj 2014-07-15 00:02:57 EDT
What would be the expected behavior?

- Cluster status will be aggregation of status of all volumes

or 

- Cluster status will be aggregation of both status and utilization of volumes
Comment 11 Sahina Bose 2018-01-30 06:12:16 EST
Thank you for your report. However, this bug is being closed as it's logged against gluster-nagios monitoring for which no further new development is being undertaken.

Note You need to log in before you can comment on or make changes to this bug.