Description of problem: Ceph df normally reports the MAX AVAIL space considering the OSDs in the ruleset, but when on of the OSDs is down and out it just reports 0 instead of the real MAX AVAIL space for the pools using that ruleset. Version-Release number of selected component (if applicable): ceph-0.80.8-4.el7cp.x86_64 Additional info: this bug seems to have been reported and fixed upstream at http://tracker.ceph.com/issues/10257
merged upstream to 0.80.x in https://github.com/ceph/ceph/pull/3826
Ken, can this bug be fixed in 1.3.0 too? The information given to user in this defect report is incorrect.
This issue significantly affects Openstack deployments: https://bugzilla.redhat.com/show_bug.cgi?id=1224978 If any of OSD nodes goes down, user can not create new cinder volumes in Overcloud.
Ken, QE is not seeing this issue in recent 1.3.0 builds. Not a blocker for 1.3.0 release.
Thanks Harish! Looking at this again, I see now that https://github.com/ceph/ceph/pull/3408 was merged back in January, so the fix has been in Hammer from day one. I'm removing the ceph-1.3.0 release flag in order to clarify this. So this bug is only for Firefly / RHCS 1.2.z. The upstream fix there is https://github.com/ceph/ceph/pull/3826 .
I can confirm that this issue is solved in ceph-0.94.1-11.el7cp.x86_64, with this version "ceph df" returns reasonable values after removing an OSD node and "cinder create" command works.
(In reply to Ken Dreyer (Red Hat) from comment #7) > So this bug is only for Firefly / RHCS 1.2.z. The upstream fix there is > https://github.com/ceph/ceph/pull/3826 . This fix has shipped in v0.80.9 upstream, so we'll take in in as part of the RHCS 1.2 rebase to Ceph v0.80.11 (bz 1271774)
Closing all RHCS 1.2 bugs, since we are not planning to do a formal RHCS 1.2.4 release. This particular issue was already fixed in the initial RHCS 1.3.0 release.