Bug 2183294
| Summary: | [RFE] Catch MDS damage to the dentry's first snapid | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat Ceph Storage | Reporter: | Patrick Donnelly <pdonnell> |
| Component: | CephFS | Assignee: | Patrick Donnelly <pdonnell> |
| Status: | CLOSED ERRATA | QA Contact: | Hemanth Kumar <hyelloji> |
| Severity: | urgent | Docs Contact: | lysanche |
| Priority: | urgent | ||
| Version: | 5.3 | CC: | amk, ceph-eng-bugs, cephqe-warriors, gfarnum, hyelloji, rmandyam, tserlin, vereddy, vshankar, vumrao |
| Target Milestone: | --- | Keywords: | FutureFeature |
| Target Release: | 5.3z3 | ||
| Hardware: | All | ||
| OS: | All | ||
| Whiteboard: | |||
| Fixed In Version: | ceph-16.2.10-164.el8cp | Doc Type: | Bug Fix |
| Doc Text: |
.A code assert is added to the Ceph Manager daemon service to detect metadata corruption
Previously, a type of snapshot-related metadata corruption would be introduced by the manager daemon service for workloads running Postgres, and possibly others.
With this fix, a code assert is added to the manager daemon service which is triggered if a new corruption is detected. This reduces the proliferation of the damage, and allows the collection of logs to ascertain the cause.
[NOTE]
====
If daemons crash after the cluster is upgraded to {storage-product} 5.3z3, contact link:https://access.redhat.com/support/contact/technicalSupport/[_Red Hat support_] for analysis and corrective action.
====
|
Story Points: | --- |
| Clone Of: | Environment: | ||
| Last Closed: | 2023-05-23 00:19:10 UTC | Type: | --- |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
| Bug Depends On: | |||
| Bug Blocks: | 2203283 | ||
|
Description
Patrick Donnelly
2023-03-30 19:20:31 UTC
Hi Patrick, we are analyzing the issue. Can you please help with reproduction steps for the same. Regards, Amarnath As initial backport tests did not pass and we are on a tight deadline, I am moving this to 5.3z3 for now. :( We may issue an async (or even grab it back) if we figure out and resolve the test issues fast enough. (In reply to Amarnath from comment #2) > Hi Patrick, > > we are analyzing the issue. > Can you please help with reproduction steps for the same. You would run these tests: https://gitlab.cee.redhat.com/ceph/ceph/-/merge_requests/278/diffs#7b2dc3f617cfcca3e13c38ef537cd6355175ac6b_565_567 Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Red Hat Ceph Storage 5.3 Bug Fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2023:3259 |