Description of problem (please be detailed as possible and provide log snippests): After OCP upgrade one monitor is reported as down and it doesn't recover in time. Version of all relevant components (if applicable): OCS 4.8.18-2 Rate from 1 - 5 the complexity of the scenario you performed that caused this bug (1 - very simple, 5 - very complex)? 2 Can this issue reproducible? yes Steps to Reproduce: 1. Prepare AWS IPI 3AZ RHCOS 3M 3W Cluster. 2. Perform OCP upgrade from OCP 4.8 to OCP 4.9. 3. Check ceph health. Actual results: Upgrade is done but a ceph monitor is down (Ceph cluster health is not OK. Health: HEALTH_WARN 1/3 mons down, quorum a,c). Expected results: Ceph should be healthy after upgrade. Additional info: Runs where this was observed: https://ocs4-jenkins-csb-odf-qe.apps.ocp-c1.prod.psi.redhat.com/job/qe-deploy-ocs-cluster-prod/7177/#showFailuresLink https://ocs4-jenkins-csb-odf-qe.apps.ocp-c1.prod.psi.redhat.com/job/qe-deploy-ocs-cluster-prod/7139/#showFailuresLink
Filip, Can you pls give us the working cluster, Above both clusters are gone.