Bug 1986175

Summary: [Tracker for BZ #1995906] 1 MDSs report oversized cache keeps reappearing
Product: [Red Hat Storage] Red Hat OpenShift Data Foundation Reporter: kelwhite
Component: cephAssignee: Venky Shankar <vshankar>
ceph sub component: CephFS QA Contact: Elad <ebenahar>
Status: CLOSED CURRENTRELEASE Docs Contact:
Severity: urgent    
Priority: urgent CC: akrai, bniver, hnallurv, hyelloji, jbiao, madam, mduasope, muagarwa, nberry, ocs-bugs, odf-bz-bot, pdhange, pdonnell, shan
Version: 4.7   
Target Milestone: ---   
Target Release: ODF 4.11.0   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: No Doc Update
Doc Text:
Story Points: ---
Clone Of:
: 1995906 (view as bug list) Environment:
Last Closed: 2022-06-07 09:15:17 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 1995906    
Bug Blocks:    

Comment 3 Travis Nielsen 2021-08-02 15:47:15 UTC
Patrick, can someone from cephfs take a look at this health warning?

Comment 7 Patrick Donnelly 2021-08-02 21:08:00 UTC
If it reoccurs, please collect `ceph health detail`, `ceph fs dump`, and a perf dump of the mds `ceph daemon mds.<X> perf dump` (in a debug sidecar container).

Comment 21 Patrick Donnelly 2021-10-01 00:16:04 UTC
> Q. Is there any workaround we can try on this OCS platform to get ceph status healthy?

You can try disabling standby replay:

> ceph fs set <fs_name> allow_standby_replay false

Rook may re-enable it though but it will hopefully get you to a state that the cluster can be upgraded.