Bug 1885971
Summary: | ocs-storagecluster-cephobjectstore doesn't report true state of RGW | ||
---|---|---|---|
Product: | [Red Hat Storage] Red Hat OpenShift Container Storage | Reporter: | Filip Balák <fbalak> |
Component: | rook | Assignee: | Sébastien Han <shan> |
Status: | CLOSED ERRATA | QA Contact: | Filip Balák <fbalak> |
Severity: | high | Docs Contact: | |
Priority: | unspecified | ||
Version: | 4.6 | CC: | ebenahar, madam, muagarwa, ocs-bugs |
Target Milestone: | --- | Keywords: | AutomationBackLog |
Target Release: | OCS 4.6.0 | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | 4.6.0-127.ci | Doc Type: | No Doc Update |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2020-12-17 06:24:44 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Filip Balák
2020-10-07 11:37:50 UTC
Filip, I'm not sure what's going on but I see rgw deployments here http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/fbalak-vm7/fbalak-vm7_20201007T061706/logs/testcases_1602073024/ocs_must_gather/quay-io-rhceph-dev-ocs-must-gather-sha256-9dfb64c63dd8f8ee033aff511a4ffd2906ffe2a7b637deb5c81d50b8c20eaffa/namespaces/openshift-storage/apps/deployments.yaml However they never get ready: NAME READY UP-TO-DATE AVAILABLE AGE deployment.apps/rook-ceph-rgw-ocs-storagecluster-cephobjectstore-a 0/0 0 0 5h26m deployment.apps/rook-ceph-rgw-ocs-storagecluster-cephobjectstore-b 0/0 0 0 5h26m So they are no pods running... Is it possible to access the env? Thanks Hi, there are no pods because I scaled those deployments to 0 (as part of reproducer '2. Scale all 'rgw' deployments to 0 pods.'). The problem is that there is nowhere reported that RGW is unavailable. I don't have the env at the moment and I am not sure that I will have it this week but if I will then I will ping you. Oh ok, that helps a lot actually. How long did you wait? The check runs every minute so one minute after scaling down the deployment, the status should be updated properly. Monitoring for RGW health seems fixed now. When RGW pods are scaled to 0 then RGW is displayed as red and in error state on Object Service dashboard and alert ClusterObjectStoreState is triggered. Tested with: OCP: 4.6.0-0.nightly-2020-10-13-064047 OCS: ocs-operator.v4.6.0-131.ci Thanks for verifying Filip. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat OpenShift Container Storage 4.6.0 security, bug fix, enhancement update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2020:5605 |