* Description of problem (please be detailed as possible and provide log snippests): Once the Ceph cluster reaches 'osd-full-ratio' threshold, the deletion of data is not allowed, to free-up space. This works as expected as deletion also requires "write" access which is denied when the cluster is Full. However, A user shouldn't need to increase the full-ratio in order to allow deletion. This is not a good user experience. * Version of all relevant components (if applicable): RHOCS 4.2.2 * Does this issue impact your ability to continue to work with the product (please explain in detail what is the user impact)? Partially * Is there any workaround available to the best of your knowledge? Yes, increase the threshold manually and delete data. * Rate from 1 - 5 the complexity of the scenario you performed that caused this bug (1 - very simple, 5 - very complex)? 3 * Can this issue reproducible? Yes * Can this issue reproduce from the UI? No * If this is a regression, please provide more details to justify this: No * Steps to Reproduce: NA * Actual results: NA * Expected results: NA * Additional info: The same behavior is present in RHCS as well. https://access.redhat.com/solutions/3001761
As Ashish Singh is from the GSS team, re-adding the [GSS] tag. Or should it only be used for BZs with customer cases attached? Also moving this out of 4.3
We can keep the title as is.
Is this something we could or would have to address in ceph itself?
(In reply to Michael Adam from comment #4) > Is this something we could or would have to address in ceph itself? Ceph itself can't add more capacity, OCS may be able to, so it should be addressed there.
FYI deletion in Ceph is allowed when it is full. What version and which commands are not working?
Doesn't seem like this RFE was priortized in 4.6 and now when we are nearly approaching dev freeze, I don't think we have a chance to fix it. Moving it out, please retarget if some one thinks otherwise. Also, we are in early phase of 4.7 so if we don't want this BZ to drag further now is the time to priortize it.
In order to enable deletion when cluster is we need to enable it in Ceph MGR: https://bugzilla.redhat.com/show_bug.cgi?id=1910272 We require change is Ceph-CSI as well. As we won't need any changes in OCS operator I am moving the BZ to Ceph-CSI.
(In reply to Orit Wasserman from comment #9) > In order to enable deletion when cluster is we need to enable it in Ceph > MGR: https://bugzilla.redhat.com/show_bug.cgi?id=1910272 > We require change is Ceph-CSI as well. > As we won't need any changes in OCS operator I am moving the BZ to Ceph-CSI. Marking this bug for Ceph CSI component as Tracker till we get it addressed in Ceph Core components.
https://bugzilla.redhat.com/show_bug.cgi?id=1910272 is acked for 5.0z1
This is getting fixed in 5.0z1
tested the workflow: wrote data to the cluster using ocs-ci function write_data_via_fio() until cluster was almost full sh-4.4$ ceph -s cluster: id: 1ae93eb4-edd9-4942-a27e-13dba341f1f2 health: HEALTH_ERR 3 full osd(s) 3 pool(s) full then deleted the data using delete_fio_data() data was deleted as expected sh-4.4$ ceph -s cluster: id: 1ae93eb4-edd9-4942-a27e-13dba341f1f2 health: HEALTH_OK moving to verified.
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat OpenShift Data Foundation 4.9.0 enhancement, security, and bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:5086