Bug 2001539
| Summary: | [UI] ODF Overview showing two different status for the same storage system | ||||||
|---|---|---|---|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat OpenShift Data Foundation | Reporter: | Jilju Joy <jijoy> | ||||
| Component: | odf-operator | Assignee: | umanga <uchapaga> | ||||
| Status: | CLOSED ERRATA | QA Contact: | Mugdha Soni <musoni> | ||||
| Severity: | high | Docs Contact: | |||||
| Priority: | medium | ||||||
| Version: | 4.9 | CC: | amagrawa, ebenahar, gshanmug, jefbrown, jrivera, madam, mashetty, muagarwa, musoni, nthomas, ocs-bugs, odf-bz-bot, skatiyar | ||||
| Target Milestone: | --- | Keywords: | AutomationBackLog | ||||
| Target Release: | ODF 4.9.0 | ||||||
| Hardware: | Unspecified | ||||||
| OS: | Unspecified | ||||||
| Whiteboard: | |||||||
| Fixed In Version: | v4.9.0-230.ci | Doc Type: | No Doc Update | ||||
| Doc Text: | Story Points: | --- | |||||
| Clone Of: | Environment: | ||||||
| Last Closed: | 2021-12-13 17:46:04 UTC | Type: | Bug | ||||
| Regression: | --- | Mount Type: | --- | ||||
| Documentation: | --- | CRM: | |||||
| Verified Versions: | Category: | --- | |||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||
| Embargoed: | |||||||
| Bug Depends On: | |||||||
| Bug Blocks: | 2019652 | ||||||
| Attachments: |
|
||||||
Fix is available in the latest builds On version 4.9.0-191.ci still 2 storagesystem can be seen in odf page. Please check attachment above. *** Bug 2012722 has been marked as a duplicate of this bug. *** Marking as a blocker? for the sake of not moving this out of 4.9.0 Hi **Tested the following with the ODF "4.9.0-230.ci" and OCP "4.9.0-0.nightly-2021-11-10-215111" **Steps to reproduce were performed same as mentioned in comment#0. ** Scenarios:- LSO and Non-lso cluster . ** OBSERVATIONS:- (a)One status is shown in the overview page for storage system. (b)One storage system is present in overview. (c) Ceph health was OK. sh-4.4$ ceph status cluster: id: 946078ae-63fb-4e84-9d38-8bb0863fa73a health: HEALTH_OK services: mon: 3 daemons, quorum a,c,b (age 35m) mgr: a(active, since 35m) mds: 1/1 daemons up, 1 hot standby osd: 3 osds: 3 up (since 34m), 3 in (since 35m) data: volumes: 1/1 healthy pools: 4 pools, 97 pgs objects: 93 objects, 134 MiB usage: 275 MiB used, 1.5 TiB / 1.5 TiB avail pgs: 97 active+clean io: client: 1.2 KiB/s rd, 6.3 KiB/s wr, 2 op/s rd, 0 op/s wr ** Screenshots :- https://docs.google.com/document/d/11PGHiz9POZFTzXQ_pysFSYFS-61YtPOm46SMCqUsbdc/edit?usp=sharing Moving the bug to verified state . Thanks Mugdha Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat OpenShift Data Foundation 4.9.0 enhancement, security, and bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:5086 |
Created attachment 1820806 [details] Screen recording which shows the issue Description of problem (please be detailed as possible and provide log snippests): OpenShift Data Foundation Overview page is showing two different status for the same storage system. Only one storage system "ocs-storagecluster-storagesystem" is present in the cluster. One status is showing the storage system "ocs-storagecluster-storagesystem" as healthy and another status on the same page is showing that the storage system "ocs-storagecluster-storagesystem" as degraded. The storage system is actually healthy. Testing was done on VMware LSO configuration. $ oc get storagesystem NAME STORAGE-SYSTEM-KIND STORAGE-SYSTEM-NAME ocs-storagecluster-storagesystem storagecluster.ocs.openshift.io/v1 ocs-storagecluster $ oc get storagecluster NAME AGE PHASE EXTERNAL CREATED AT VERSION ocs-storagecluster 10m Ready 2021-09-06T10:12:31Z 4.9.0 Screen recording is attached. ======================================================================= Version of all relevant components (if applicable): $ oc get clusterversion NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.9.0-0.nightly-2021-09-06-004132 True False 3h47m Cluster version is 4.9.0-0.nightly-2021-09-06-004132 $ oc get csv NAME DISPLAY VERSION REPLACES PHASE noobaa-operator.v4.9.0-125.ci NooBaa Operator 4.9.0-125.ci Succeeded ocs-operator.v4.9.0-125.ci OpenShift Container Storage 4.9.0-125.ci Succeeded odf-operator.v4.9.0-125.ci OpenShift Data Foundation 4.9.0-125.ci Succeeded ======================================================================= Does this issue impact your ability to continue to work with the product (please explain in detail what is the user impact)? UI is showing wrong status Is there any workaround available to the best of your knowledge? No Rate from 1 - 5 the complexity of the scenario you performed that caused this bug (1 - very simple, 5 - very complex)? 1 Can this issue reproducible? Reporting the first occurrence Can this issue reproduce from the UI? UI issue If this is a regression, please provide more details to justify this: Steps to Reproduce: 1. Install ODF operator and and create a storagesystem from UI using LSO configuration. 2. Go to Storage --> OpenShift Data Foundation. 3. Check the status of storagesystem Actual results: Two different status is shown for one storage system. Expected results: One one status should be shown in overview page. Additional info: