Description of problem (please be detailed as possible and provide log snippests): Upgrade from OCS 4.6 to 4.7 is failing to succeed and CSV hangs in installing phase cause of noobaa core pod is in CrashLoopBackOff. Version of all relevant components (if applicable): OCS upgrade from v4.6.0-160.ci to 4.7.0-163.ci OCP 4.7.0-0.nightly-2020-11-22-204912 Does this issue impact your ability to continue to work with the product (please explain in detail what is the user impact)? Yes, cannot upgrade. Is there any workaround available to the best of your knowledge? No Rate from 1 - 5 the complexity of the scenario you performed that caused this bug (1 - very simple, 5 - very complex)? 1 Can this issue reproducible? Need to find out. Can this issue reproduce from the UI? Haven't tried If this is a regression, please provide more details to justify this: This worked before. Steps to Reproduce: 1. Install OCS 4.6 on top of OCP 4.7 2. Upgrade OCS to 4.7 build 3. Upgrade will not finish Actual results: CSV in installing phase, noba core pod in CrashLoopBackOff Expected results: Have upgrade succeeded Additional info: Job: https://ocs4-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/job/qe-deploy-ocs-cluster/14920/consoleFull Must gather: http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/j010vu1cs33-uan/j010vu1cs33-uan_20201123T080920/logs/failed_testcase_ocs_logs_1606123450/test_upgrade_ocs_logs/
I see it got reproduced also in this run: https://ocs4-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/job/qe-deploy-ocs-cluster/14923/ Must gather: http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/j011vu1cs33-uan/j011vu1cs33-uan_20201123T091220/logs/failed_testcase_ocs_logs_1606127218/test_upgrade_ocs_logs/
This is a 4.7 issue (the problem is in the new version) and would be fixed in 4.7 Moving
The issue still persist: https://ocs4-jenkins-csb-ocsqe.cloud.paas.psi.redhat.com/job/qe-deploy-ocs-cluster/85/console noobaa-core-0 0/1 CrashLoopBackOff 10 29m
Ran verification job here: https://ocs4-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/job/qe-deploy-ocs-cluster/15583/ Upgrade from 4.6 RC 7 to 4.7.0-192.ci which I see should have a fix.
Running new upgrade job here: https://ocs4-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/job/qe-trigger-aws-ipi-3az-rhcos-3m-3w-upgrade-ocs-auto-nightly/2/console
Running another verification jobs here: Running 2 verification jobs here: vSphere OCP 4.6 https://ocs4-jenkins-csb-ocsqe.apps.ocp4.prod.psi.redhat.com/job/qe-deploy-ocs-cluster/217/ AWS OCP 4.7: https://ocs4-jenkins-csb-ocsqe.apps.ocp4.prod.psi.redhat.com/job/qe-deploy-ocs-cluster/218/
Upgrade passed but then we saw some failures in our infra cause of: https://bugzilla.redhat.com/show_bug.cgi?id=1919967 So I am marking this BZ as verified. Petr
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat OpenShift Container Storage 4.7.0 security, bug fix, and enhancement update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:2041