Bug 1225081 - Ceph df doesn't report MAX AVAIL correctly when an OSD in ruleset is down and out
Summary: Ceph df doesn't report MAX AVAIL correctly when an OSD in ruleset is down and...
Keywords:
Status: CLOSED WONTFIX
Alias: None
Product: Red Hat Ceph Storage
Classification: Red Hat Storage
Component: RADOS
Version: 1.2.3
Hardware: Unspecified
OS: Unspecified
unspecified
high
Target Milestone: rc
: 1.2.4
Assignee: Samuel Just
QA Contact: ceph-qe-bugs
URL:
Whiteboard:
Depends On: 1271774
Blocks:
TreeView+ depends on / blocked
 
Reported: 2015-05-26 14:22 UTC by Giulio Fidente
Modified: 2017-07-30 15:12 UTC (History)
9 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Clone Of:
Environment:
Last Closed: 2015-12-08 16:31:11 UTC
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Ceph Project Bug Tracker 10257 0 None None None Never

Description Giulio Fidente 2015-05-26 14:22:36 UTC
Description of problem:
Ceph df normally reports the MAX AVAIL space considering the OSDs in the ruleset, but when on of the OSDs is down and out it just reports 0 instead of the real MAX AVAIL space for the pools using that ruleset.


Version-Release number of selected component (if applicable):
ceph-0.80.8-4.el7cp.x86_64


Additional info:
this bug seems to have been reported and fixed upstream at http://tracker.ceph.com/issues/10257

Comment 2 Ken Dreyer (Red Hat) 2015-05-26 17:33:25 UTC
merged upstream to 0.80.x in https://github.com/ceph/ceph/pull/3826

Comment 3 Harish NV Rao 2015-05-27 15:37:14 UTC
Ken, can this bug be fixed in 1.3.0 too? The information given to user in this defect report is incorrect.

Comment 5 Jan Provaznik 2015-05-28 09:37:07 UTC
This issue significantly affects Openstack deployments:
https://bugzilla.redhat.com/show_bug.cgi?id=1224978

If any of OSD nodes goes down, user can not create new cinder volumes in Overcloud.

Comment 6 Harish NV Rao 2015-05-28 10:23:19 UTC
Ken, QE is not seeing this issue in recent 1.3.0 builds. Not a blocker for 1.3.0 release.

Comment 7 Ken Dreyer (Red Hat) 2015-05-28 14:24:41 UTC
Thanks Harish! Looking at this again, I see now that https://github.com/ceph/ceph/pull/3408 was merged back in January, so the fix has been in Hammer from day one. I'm removing the ceph-1.3.0 release flag in order to clarify this.

So this bug is only for Firefly / RHCS 1.2.z. The upstream fix there is https://github.com/ceph/ceph/pull/3826 .

Comment 8 Jan Provaznik 2015-06-02 13:56:50 UTC
I can confirm that this issue is solved in ceph-0.94.1-11.el7cp.x86_64, with this version "ceph df" returns reasonable values after removing an OSD node and "cinder create" command works.

Comment 9 Ken Dreyer (Red Hat) 2015-10-19 17:21:43 UTC
(In reply to Ken Dreyer (Red Hat) from comment #7)
> So this bug is only for Firefly / RHCS 1.2.z. The upstream fix there is
> https://github.com/ceph/ceph/pull/3826 .

This fix has shipped in v0.80.9 upstream, so we'll take in in as part of the RHCS 1.2 rebase to Ceph v0.80.11 (bz 1271774)

Comment 10 Ken Dreyer (Red Hat) 2015-12-08 16:31:11 UTC
Closing all RHCS 1.2 bugs, since we are not planning to do a formal RHCS 1.2.4 release.

This particular issue was already fixed in the initial RHCS 1.3.0 release.


Note You need to log in before you can comment on or make changes to this bug.