Description of problem: When a cluster has no OSDs or no managers health is reported as HEALTH_OK cluster: id: 97ce8ce8-811c-46ce-9682-ce535d9859ab health: HEALTH_OK services: mon: 3 daemons, quorum a,b,c (age 11m) mgr: no daemons active osd: 0 osds: 0 up, 0 in data: pools: 0 pools, 0 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: Version-Release number of selected component (if applicable): 14.2.X (any including the latest 14.2.4) How reproducible: all the time Steps to Reproduce: 1. Deploy Ceph with no managers or OSDs 2. 3. Actual results: report is HEALTH_OK Expected results: report is HEALTH_WARN or HEALTH_ERR Additional info:
Please specify the severity of this bug. Severity is defined here: https://bugzilla.redhat.com/page.cgi?id=fields.html#bug_severity.
Severity?
This is more of a core ceph/rados thing, re-targetting.
Has anyone looked at this? I assume it's because the mgr is not available?
I don't know if anyone has looked at this and I am not sure why this is happening.
This was by design when ceph-mgr was created - the idea at the time was to avoid spurious warnings during cluster setup. and at that point ceph-mgr was not necessary for much functionality. At this point, ceph-mgr is doing much more. Currently the health status is only affected if there was ever a mgr running - it seems removing this condition so you get an error after there's no mgr for some time would resolve this. Moving to 4.1 since this is not a blocker for 4.0 (same behavior as 3.x).
included in 14.2.8 rebase
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2020:2231