Bug 1851347 - [RFE][DOC] Disaster recovery steps to restore ceph-monitor quorum when 2 out of 3 monitors are lost
Summary: [RFE][DOC] Disaster recovery steps to restore ceph-monitor quorum when 2 out ...
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: Red Hat OpenShift Container Storage
Classification: Red Hat Storage
Component: documentation
Version: 4.3
Hardware: x86_64
OS: Linux
urgent
high
Target Milestone: ---
: ---
Assignee: Kusuma
QA Contact: Elad
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2020-06-26 09:37 UTC by Karun Josy
Modified: 2023-08-04 12:11 UTC (History)
11 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2022-03-02 07:11:51 UTC
Embargoed:
kbg: needinfo-
kbg: needinfo-


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Github red-hat-storage ocs-ci pull 5167 0 None Merged Mon recovery system test automation 2022-06-20 13:05:26 UTC

Comment 5 Elad 2020-09-24 12:36:10 UTC
This is not part of 4.6 content. Can we move this out?

Comment 20 Travis Nielsen 2021-10-04 17:18:03 UTC
After the mons scale up to 3 mons again, we should still see mon-a in the quorum. It's unexpected that mon-a is not showing up in quorum. Since all the daemons are crashing, this likely indicates that a new mon quorum has been created and the original mon quorum was lost. Something must not have worked when restoring the quorum.

After mon-b and mon-c are taken down, and the mon quorum is reset to only include mon-a, is the ceph status seen to be healthy with the single mon? Before continuing with the guide to restore up to 3 mons again, we need to see that the cluster is healthy with a single mon


Note You need to log in before you can comment on or make changes to this bug.