Bug 2213739
| Summary: | Disable topology view for external mode | |||
|---|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat OpenShift Data Foundation | Reporter: | Aaruni Aggarwal <aaaggarw> | |
| Component: | management-console | Assignee: | Bipul Adhikari <badhikar> | |
| Status: | CLOSED ERRATA | QA Contact: | Aaruni Aggarwal <aaaggarw> | |
| Severity: | unspecified | Docs Contact: | Kusuma <kbg> | |
| Priority: | unspecified | |||
| Version: | 4.13 | CC: | amagrawa, asriram, badhikar, dosypenk, ebenahar, etamir, kramdoss, nthomas, ocs-bugs, odf-bz-bot, sheggodu, skatiyar | |
| Target Milestone: | --- | |||
| Target Release: | ODF 4.13.1 | |||
| Hardware: | ppc64le | |||
| OS: | Linux | |||
| Whiteboard: | ||||
| Fixed In Version: | 4.13.1-3 | Doc Type: | Bug Fix | |
| Doc Text: |
Previously, an empty screen was seen for the topology view of the external mode because in external mode, the nodes are not labelled with the OCS label and hence, the topology view did not show the nodes at the first level.
With this fix, the topology view is disabled for the external mode clusters and as a result, the confusing empty screen is not displayed.
|
Story Points: | --- | |
| Clone Of: | ||||
| : | 2216707 (view as bug list) | Environment: | ||
| Last Closed: | 2023-08-02 16:07:42 UTC | Type: | Bug | |
| Regression: | --- | Mount Type: | --- | |
| Documentation: | --- | CRM: | ||
| Verified Versions: | Category: | --- | ||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
| Cloudforms Team: | --- | Target Upstream Version: | ||
| Embargoed: | ||||
| Bug Depends On: | ||||
| Bug Blocks: | 2154341, 2216707 | |||
Storagecluster and Cephcluster:
[root@rdr-cicd-odf-ef1f-bastion-0 ~]# oc get storagecluster -n openshift-storage
NAME AGE PHASE EXTERNAL CREATED AT VERSION
ocs-external-storagecluster 15m Ready true 2023-06-08T06:06:04Z 4.13.0
[root@rdr-cicd-odf-ef1f-bastion-0 ~]# oc get cephcluster -n openshift-storage
NAME DATADIRHOSTPATH MONCOUNT AGE PHASE MESSAGE HEALTH EXTERNAL FSID
ocs-external-storagecluster-cephcluster 15m Connected Cluster connected successfully HEALTH_OK true 46ab550e-ca6e-11ed-af21-00000a0b13a5
[root@rdr-cicd-odf-ef1f-bastion-0 ~]# oc describe storagecluster -n openshift-storage
Name: ocs-external-storagecluster
Namespace: openshift-storage
Labels: <none>
Annotations: uninstall.ocs.openshift.io/cleanup-policy: delete
uninstall.ocs.openshift.io/mode: graceful
API Version: ocs.openshift.io/v1
Kind: StorageCluster
Metadata:
Creation Timestamp: 2023-06-08T06:06:04Z
Finalizers:
storagecluster.ocs.openshift.io
Generation: 2
Managed Fields:
API Version: ocs.openshift.io/v1
Fields Type: FieldsV1
fieldsV1:
f:spec:
.:
f:externalStorage:
.:
f:enable:
f:network:
.:
f:connections:
.:
f:encryption:
Manager: Mozilla
Operation: Update
Time: 2023-06-08T06:06:04Z
API Version: ocs.openshift.io/v1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:ownerReferences:
.:
k:{"uid":"712d0159-3df2-43b1-a611-eae1e9de5f0b"}:
Manager: manager
Operation: Update
Time: 2023-06-08T06:06:04Z
API Version: ocs.openshift.io/v1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:annotations:
.:
f:uninstall.ocs.openshift.io/cleanup-policy:
f:uninstall.ocs.openshift.io/mode:
f:finalizers:
.:
v:"storagecluster.ocs.openshift.io":
f:spec:
f:arbiter:
f:encryption:
.:
f:kms:
f:labelSelector:
f:managedResources:
.:
f:cephBlockPools:
f:cephCluster:
f:cephConfig:
f:cephDashboard:
f:cephFilesystems:
f:cephNonResilientPools:
f:cephObjectStoreUsers:
f:cephObjectStores:
f:cephRBDMirror:
f:cephToolbox:
f:mirroring:
f:network:
f:multiClusterService:
Manager: ocs-operator
Operation: Update
Time: 2023-06-08T06:06:04Z
API Version: ocs.openshift.io/v1
Fields Type: FieldsV1
fieldsV1:
f:status:
.:
f:conditions:
f:externalSecretHash:
f:images:
.:
f:ceph:
.:
f:desiredImage:
f:noobaaCore:
.:
f:actualImage:
f:desiredImage:
f:noobaaDB:
.:
f:actualImage:
f:desiredImage:
f:kmsServerConnection:
f:phase:
f:relatedObjects:
f:version:
Manager: ocs-operator
Operation: Update
Subresource: status
Time: 2023-06-08T06:25:06Z
Owner References:
API Version: odf.openshift.io/v1alpha1
Kind: StorageSystem
Name: ocs-external-storagecluster-storagesystem
UID: 712d0159-3df2-43b1-a611-eae1e9de5f0b
Resource Version: 1402178
UID: dd67158d-dc54-4499-aa6c-38c0e70c75d4
Spec:
Arbiter:
Encryption:
Kms:
External Storage:
Enable: true
Label Selector:
Managed Resources:
Ceph Block Pools:
Ceph Cluster:
Ceph Config:
Ceph Dashboard:
Ceph Filesystems:
Ceph Non Resilient Pools:
Ceph Object Store Users:
Ceph Object Stores:
Ceph RBD Mirror:
Ceph Toolbox:
Mirroring:
Network:
Connections:
Encryption:
Multi Cluster Service:
Status:
Conditions:
Last Heartbeat Time: 2023-06-08T06:06:06Z
Last Transition Time: 2023-06-08T06:06:06Z
Message: Version check successful
Reason: VersionMatched
Status: False
Type: VersionMismatch
Last Heartbeat Time: 2023-06-08T06:25:06Z
Last Transition Time: 2023-06-08T06:06:06Z
Message: Reconcile completed successfully
Reason: ReconcileCompleted
Status: True
Type: ReconcileComplete
Last Heartbeat Time: 2023-06-08T06:25:06Z
Last Transition Time: 2023-06-08T06:08:28Z
Message: Reconcile completed successfully
Reason: ReconcileCompleted
Status: True
Type: Available
Last Heartbeat Time: 2023-06-08T06:25:06Z
Last Transition Time: 2023-06-08T06:08:28Z
Message: Reconcile completed successfully
Reason: ReconcileCompleted
Status: False
Type: Progressing
Last Heartbeat Time: 2023-06-08T06:25:06Z
Last Transition Time: 2023-06-08T06:06:06Z
Message: Reconcile completed successfully
Reason: ReconcileCompleted
Status: False
Type: Degraded
Last Heartbeat Time: 2023-06-08T06:25:06Z
Last Transition Time: 2023-06-08T06:08:28Z
Message: Reconcile completed successfully
Reason: ReconcileCompleted
Status: True
Type: Upgradeable
External Secret Hash: 5fe48ad90d8bf08a712e2f5a67384c5ebf2348fca08124574afb4a77f0a5eb97e7841ad23c23ab7f180437ec01356789b94ca5aa6a73dca853a730cde0a63fc0
Images:
Ceph:
Desired Image: quay.io/rhceph-dev/rhceph@sha256:fa6d01cdef17bc32d2b95b8121b02f4d41adccc5ba8a9b95f38c97797ff6621f
Noobaa Core:
Actual Image: quay.io/rhceph-dev/odf4-mcg-core-rhel9@sha256:574b6258ee4d7ac2532b9143390a130acf89b34b77abc6c446a329286bcd27a5
Desired Image: quay.io/rhceph-dev/odf4-mcg-core-rhel9@sha256:574b6258ee4d7ac2532b9143390a130acf89b34b77abc6c446a329286bcd27a5
Noobaa DB:
Actual Image: quay.io/rhceph-dev/rhel8-postgresql-12@sha256:f7f678d44d5934ed3d95c83b4428fee4b616f37e8eadc5049778f133b4ce3713
Desired Image: quay.io/rhceph-dev/rhel8-postgresql-12@sha256:f7f678d44d5934ed3d95c83b4428fee4b616f37e8eadc5049778f133b4ce3713
Kms Server Connection:
Phase: Ready
Related Objects:
API Version: ceph.rook.io/v1
Kind: CephCluster
Name: ocs-external-storagecluster-cephcluster
Namespace: openshift-storage
Resource Version: 1401943
UID: fe6a1e03-0a05-4ecf-9aae-c7b6565cd7b7
API Version: noobaa.io/v1alpha1
Kind: NooBaa
Name: noobaa
Namespace: openshift-storage
Resource Version: 1402174
UID: 39ed0313-cb11-4020-8bad-e8afd812a047
Version: 4.13.0
Events: <none>
[root@rdr-cicd-odf-ef1f-bastion-0 ~]# oc describe cephcluster -n openshift-storage
Name: ocs-external-storagecluster-cephcluster
Namespace: openshift-storage
Labels: app=ocs-external-storagecluster
Annotations: <none>
API Version: ceph.rook.io/v1
Kind: CephCluster
Metadata:
Creation Timestamp: 2023-06-08T06:06:05Z
Finalizers:
cephcluster.ceph.rook.io
Generation: 1
Managed Fields:
API Version: ceph.rook.io/v1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:labels:
.:
f:app:
f:ownerReferences:
.:
k:{"uid":"dd67158d-dc54-4499-aa6c-38c0e70c75d4"}:
f:spec:
.:
f:cephVersion:
f:cleanupPolicy:
.:
f:sanitizeDisks:
f:crashCollector:
.:
f:disable:
f:dashboard:
f:disruptionManagement:
f:external:
.:
f:enable:
f:healthCheck:
.:
f:daemonHealth:
.:
f:mon:
f:osd:
f:status:
f:labels:
.:
f:monitoring:
.:
f:rook.io/managedBy:
f:logCollector:
f:mgr:
f:mon:
f:monitoring:
.:
f:enabled:
f:externalMgrEndpoints:
f:externalMgrPrometheusPort:
f:network:
.:
f:connections:
.:
f:encryption:
f:multiClusterService:
f:security:
.:
f:keyRotation:
.:
f:enabled:
f:kms:
f:storage:
Manager: ocs-operator
Operation: Update
Time: 2023-06-08T06:06:05Z
API Version: ceph.rook.io/v1
Fields Type: FieldsV1
fieldsV1:
f:metadata:
f:finalizers:
.:
v:"cephcluster.ceph.rook.io":
Manager: rook
Operation: Update
Time: 2023-06-08T06:06:12Z
API Version: ceph.rook.io/v1
Fields Type: FieldsV1
fieldsV1:
f:status:
.:
f:ceph:
.:
f:capacity:
.:
f:bytesAvailable:
f:bytesTotal:
f:bytesUsed:
f:lastUpdated:
f:fsid:
f:health:
f:lastChecked:
f:versions:
.:
f:mds:
.:
f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable):
f:mgr:
.:
f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable):
f:mon:
.:
f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable):
f:osd:
.:
f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable):
f:overall:
.:
f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable):
f:rgw:
.:
f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable):
f:conditions:
f:message:
f:phase:
f:state:
f:version:
.:
f:version:
Manager: rook
Operation: Update
Subresource: status
Time: 2023-06-08T06:24:39Z
Owner References:
API Version: ocs.openshift.io/v1
Block Owner Deletion: true
Controller: true
Kind: StorageCluster
Name: ocs-external-storagecluster
UID: dd67158d-dc54-4499-aa6c-38c0e70c75d4
Resource Version: 1401943
UID: fe6a1e03-0a05-4ecf-9aae-c7b6565cd7b7
Spec:
Ceph Version:
Cleanup Policy:
Sanitize Disks:
Crash Collector:
Disable: true
Dashboard:
Disruption Management:
External:
Enable: true
Health Check:
Daemon Health:
Mon:
Osd:
Status:
Labels:
Monitoring:
rook.io/managedBy: ocs-external-storagecluster
Log Collector:
Mgr:
Mon:
Monitoring:
Enabled: true
External Mgr Endpoints:
Ip: 9.46.254.57
External Mgr Prometheus Port: 9283
Network:
Connections:
Encryption:
Multi Cluster Service:
Security:
Key Rotation:
Enabled: false
Kms:
Storage:
Status:
Ceph:
Capacity:
Bytes Available: 3198717943808
Bytes Total: 3221200306176
Bytes Used: 22482362368
Last Updated: 2023-06-08T06:24:37Z
Fsid: 46ab550e-ca6e-11ed-af21-00000a0b13a5
Health: HEALTH_OK
Last Checked: 2023-06-08T06:24:37Z
Versions:
Mds:
ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 1
Mgr:
ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 2
Mon:
ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 5
Osd:
ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 6
Overall:
ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 16
Rgw:
ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 2
Conditions:
Last Heartbeat Time: 2023-06-08T06:06:12Z
Last Transition Time: 2023-06-08T06:06:12Z
Message: Attempting to connect to an external Ceph cluster
Reason: ClusterConnecting
Status: True
Type: Connecting
Last Heartbeat Time: 2023-06-08T06:24:39Z
Last Transition Time: 2023-06-08T06:06:31Z
Message: Cluster connected successfully
Reason: ClusterConnected
Status: True
Type: Connected
Message: Cluster connected successfully
Phase: Connected
State: Connected
Version:
Version: 16.2.10-138
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal ReconcileSucceeded 18m rook-ceph-cluster-controller successfully configured CephCluster "openshift-storage/ocs-external-storagecluster-cephcluster"
vSphere based deployment, external mode has the same issue: OC version: Client Version: 4.12.0-202208031327 Kustomize Version: v4.5.4 Server Version: 4.13.0-0.nightly-2023-06-12-231643 Kubernetes Version: v1.26.5+7d22122 OCS verison: ocs-operator.v4.13.0-rhodf OpenShift Container Storage 4.13.0-rhodf ocs-operator.v4.12.4-rhodf Succeeded Cluster version NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.13.0-0.nightly-2023-06-12-231643 True False 16h Cluster version is 4.13.0-0.nightly-2023-06-12-231643 Rook version: rook: v4.12.4-0.bc1e9806c3281090b58872e303e947ff5437c078 go: go1.18.10 Ceph version: ceph version 16.2.10-172.el8cp (00a157ecd158911ece116ae43095de793ed9f389) pacific (stable) ---------- screen-recording https://drive.google.com/file/d/1MtVp4PoPcY-rX40G0fV9jvdNhI2Tea6C/view?usp=sharing web console output http://pastebin.test.redhat.com/1102577 The first idea was to infact hide the tab but if we hide the tab then again we would get questions and possibly GSS bugs of customers asking why the topology tab is not showing. Giving them a message should eliminate that possible issue. WDYT @etamir @chandkum ? Topology is disabled in External mode cluster for IBM Power. Tried with build: 4.13.1-4 Attaching the screenshot as well. (In reply to Aaruni Aggarwal from comment #13) > Topology is disabled in External mode cluster for IBM Power. > Tried with build: 4.13.1-4 > > Attaching the screenshot as well. Moving the bug to verified based on this comment. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat OpenShift Data Foundation 4.13.1 security and bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2023:4437 |
Created attachment 1969855 [details] topology tab in external mode odf cluster Description of problem (please be detailed as possible and provide log snippests): Topology is missing inside Storage system in External mode ODF cluster. Version of all relevant components (if applicable): ODF: 4.13.0-207 OCP: 4.13 Does this issue impact your ability to continue to work with the product (please explain in detail what is the user impact)? Is there any workaround available to the best of your knowledge? No Rate from 1 - 5 the complexity of the scenario you performed that caused this bug (1 - very simple, 5 - very complex)? Can this issue reproducible? Yes Can this issue reproduce from the UI? Yes If this is a regression, please provide more details to justify this: Steps to Reproduce: 1.Create RHCS Cluster 2. CReate OCP cluster, install ODF Operator on it. 3. create storage system using external RHCS. 4. Navigate to Storage -> Data Foundation -> Topology Actual results: Topology is not appearing in topology tab. Expected results: Topology should be appearing in topology tab. Additional info: