Bug 1349792

Summary: FutureFeature: Enhance Volume status to account for brick state and lifecycle operations (grow and shrink)
Product: [Community] GlusterFS Reporter: Atin Mukherjee <amukherj>
Component: glusterd2Assignee: bugs <bugs>
Status: CLOSED DEFERRED QA Contact:
Severity: unspecified Docs Contact:
Priority: medium    
Version: mainlineCC: amukherj, atumball, jschrode, pcuzner, rhs-bugs, sasundar, vbellur
Target Milestone: ---Keywords: Triaged
Target Release: ---   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: Enhancement
Doc Text:
Story Points: ---
Clone Of: 958419 Environment:
Last Closed: 2019-05-09 19:55:56 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 958419    
Bug Blocks:    

Description Atin Mukherjee 2016-06-24 09:36:46 UTC
+++ This bug was initially created as a clone of Bug #958419 +++

Description of problem:
gluster does not show a true volume status field. The status of the volume should account for the state of the bricks that make up the volume, and also activities that impact the volume like brick removal and addition.

This BZ is therefore an RFE to change the way the volume state is derived in the CLI.

Version-Release number of selected component (if applicable):
2.0 update 4 onwards

How reproducible:
RFE request

Steps to Reproduce:
1. vol info -  to get a brick list
2. vol status - to show the status of the bricks
3. take a node down and repeat the volume status command
  
Actual results:
At the moment, vol info shows the bricks that should be available to the volume. volume status shows those bricks - but if a node is down, the status doesn't show the missing brick at all (see BZ886865) - and more importantly the volume's state does not reflect the fact that the volume is effectively degraded (lost resiliency).

Expected results:
The volume status field should account for the components of the volume and any lifecycle operations that may be affecting the volume. 

For example, in a dist-repl volume 
- when all bricks are online - status = healthy
- when one brick/node in a replica set is down, status is "degraded"
- when multiple bricks are offline, but the offline bricks still leave an alternate brick within the replica set available - status remains "degraded"
- when a 2 bricks in the same replica set go offline - status should show something like "restricted" to indicate that data access will be impaired.
- when all bricks are inaccessible for a volume - status should show "Offline"

for a distributed volume it's a little simpler - it's either healthy, or restricted

In addition, when bricks are added to a volume the rebalance process should trigger a volume state change to 'Growing', and when removing bricks - "Shrinking"..

Additional info:

--- Additional comment from RHEL Product and Program Management on 2013-05-01 08:24:32 EDT ---

Since this issue was entered in bugzilla, the release flag has been
set to ? to ensure that it is properly evaluated for this release.

--- Additional comment from Simon Green on 2013-07-28 19:18:42 EDT ---

User pcuzner's account has been closed

--- Additional comment from John Skeoch on 2015-11-03 18:07:52 EST ---

User kparthas's account has been closed

--- Additional comment from Atin Mukherjee on 2015-12-24 04:13:37 EST ---

Agreed Paul, this is a limitation in volume status implementation. We have plans to fix this in GlusterD 2.0 as we can rely on centralized store, volume status can still be a local glusterd operation with out talking to all other nodes. We plan to re target this bug for RHGS 4.0. 

I am still keeping this bug open till we have rhgs 4.0 flag available.

--- Additional comment from Atin Mukherjee on 2016-06-24 05:36:16 EDT ---

I am going to clone this bug under glusterd2 component and close this now.

Comment 1 Amar Tumballi 2019-05-09 19:55:56 UTC
We will not be working on this anytime soon, hence marking it DEFERRED. Will revisit this after couple of releases.