Bug 2151591

Summary: [GSS] ODF 4.10 Ceph daemon crashes occurring once per day
Product: [Red Hat Storage] Red Hat OpenShift Data Foundation Reporter: Levy Sant'Anna <lsantann>
Component: cephAssignee: Prashant Dhange <pdhange>
ceph sub component: RADOS QA Contact: Elad <ebenahar>
Status: CLOSED NOTABUG Docs Contact:
Severity: medium    
Priority: unspecified CC: bniver, khover, muagarwa, nojha, ocs-bugs, odf-bz-bot, pdhange, pdhiran, tdesala, vumrao
Version: 4.10   
Target Milestone: ---   
Target Release: ---   
Hardware: All   
OS: All   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2023-03-15 19:45:42 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Comment 13 khover 2023-01-19 22:12:57 UTC
Regarding:

This crash in _scrub() is indicative of a corruption in the monstore, see below.


    bufferlist bl;
    int err = store->get(k.first, k.second, bl);
    ceph_assert(err == 0);






Do we need to rebuild this MON with corrupted db ?

That is not a easy process for customers.


https://access.redhat.com/documentation/fr-fr/red_hat_openshift_data_foundation/4.10/html/troubleshooting_openshift_data_foundation/restoring-the-monitor-pods-in-openshift-data-foundation_rhodf#doc-wrapper