Bug 2116551

Summary: How to determine excessive crc error rate
Product: [Red Hat Storage] Red Hat OpenShift Data Foundation Reporter: Jenifer Abrams <jhopper>
Component: ceph-monitoringAssignee: Juan Miguel Olmo <jolmomar>
Status: ASSIGNED --- QA Contact: Neha Berry <nberry>
Severity: high Docs Contact:
Priority: unspecified    
Version: 4.11CC: akamra, alitke, dkamboj, fdeutsch, idryomov, jdurgin, muagarwa, nojha, nthomas, odf-bz-bot, pelauter, tnielsen
Target Milestone: ---Keywords: FutureFeature
Target Release: ---Flags: jolmomar: needinfo? (jdurgin)
jolmomar: needinfo? (nojha)
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Comment 2 Nitin Goyal 2022-08-09 04:56:42 UTC
It looks like it has to be done on the rook side, Moving it to rook.

Comment 3 Travis Nielsen 2022-08-09 22:31:12 UTC
Perhaps there is already a metric being collected related to the crc error rate. 
Paul, are you aware of any related metric or this would need to be a new metric?

Comment 4 Paul Cuzner 2022-08-10 03:56:21 UTC
No. The current osd perf schema includes 3 crc cache related metrics only - nothing that would be useful to indicate CRC issues in the data path. 

Sounds like it would likely be a new metric exposed within the osd perf counter scheme.

Josh, thoughts?

Comment 18 Juan Miguel Olmo 2023-04-05 07:53:31 UTC
Moved to ODF 4.14.0 because we need to evaluate how to get and impact of the new crc errors perf counters requested

Comment 19 Jenifer Abrams 2023-05-31 21:38:18 UTC
Could this be considered a ceph health warn issue, similar to the discussion in: https://issues.redhat.com/browse/RHSTOR-3276 ?