Bug 2213739
Summary: | Disable topology view for external mode | |||
---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat OpenShift Data Foundation | Reporter: | Aaruni Aggarwal <aaaggarw> | |
Component: | management-console | Assignee: | Bipul Adhikari <badhikar> | |
Status: | CLOSED ERRATA | QA Contact: | Aaruni Aggarwal <aaaggarw> | |
Severity: | unspecified | Docs Contact: | Kusuma <kbg> | |
Priority: | unspecified | |||
Version: | 4.13 | CC: | amagrawa, asriram, badhikar, dosypenk, ebenahar, etamir, kramdoss, nthomas, ocs-bugs, odf-bz-bot, sheggodu, skatiyar | |
Target Milestone: | --- | |||
Target Release: | ODF 4.13.1 | |||
Hardware: | ppc64le | |||
OS: | Linux | |||
Whiteboard: | ||||
Fixed In Version: | 4.13.1-3 | Doc Type: | Bug Fix | |
Doc Text: |
Previously, an empty screen was seen for the topology view of the external mode because in external mode, the nodes are not labelled with the OCS label and hence, the topology view did not show the nodes at the first level.
With this fix, the topology view is disabled for the external mode clusters and as a result, the confusing empty screen is not displayed.
|
Story Points: | --- | |
Clone Of: | ||||
: | 2216707 (view as bug list) | Environment: | ||
Last Closed: | 2023-08-02 16:07:42 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | ||||
Bug Blocks: | 2154341, 2216707 |
Storagecluster and Cephcluster: [root@rdr-cicd-odf-ef1f-bastion-0 ~]# oc get storagecluster -n openshift-storage NAME AGE PHASE EXTERNAL CREATED AT VERSION ocs-external-storagecluster 15m Ready true 2023-06-08T06:06:04Z 4.13.0 [root@rdr-cicd-odf-ef1f-bastion-0 ~]# oc get cephcluster -n openshift-storage NAME DATADIRHOSTPATH MONCOUNT AGE PHASE MESSAGE HEALTH EXTERNAL FSID ocs-external-storagecluster-cephcluster 15m Connected Cluster connected successfully HEALTH_OK true 46ab550e-ca6e-11ed-af21-00000a0b13a5 [root@rdr-cicd-odf-ef1f-bastion-0 ~]# oc describe storagecluster -n openshift-storage Name: ocs-external-storagecluster Namespace: openshift-storage Labels: <none> Annotations: uninstall.ocs.openshift.io/cleanup-policy: delete uninstall.ocs.openshift.io/mode: graceful API Version: ocs.openshift.io/v1 Kind: StorageCluster Metadata: Creation Timestamp: 2023-06-08T06:06:04Z Finalizers: storagecluster.ocs.openshift.io Generation: 2 Managed Fields: API Version: ocs.openshift.io/v1 Fields Type: FieldsV1 fieldsV1: f:spec: .: f:externalStorage: .: f:enable: f:network: .: f:connections: .: f:encryption: Manager: Mozilla Operation: Update Time: 2023-06-08T06:06:04Z API Version: ocs.openshift.io/v1 Fields Type: FieldsV1 fieldsV1: f:metadata: f:ownerReferences: .: k:{"uid":"712d0159-3df2-43b1-a611-eae1e9de5f0b"}: Manager: manager Operation: Update Time: 2023-06-08T06:06:04Z API Version: ocs.openshift.io/v1 Fields Type: FieldsV1 fieldsV1: f:metadata: f:annotations: .: f:uninstall.ocs.openshift.io/cleanup-policy: f:uninstall.ocs.openshift.io/mode: f:finalizers: .: v:"storagecluster.ocs.openshift.io": f:spec: f:arbiter: f:encryption: .: f:kms: f:labelSelector: f:managedResources: .: f:cephBlockPools: f:cephCluster: f:cephConfig: f:cephDashboard: f:cephFilesystems: f:cephNonResilientPools: f:cephObjectStoreUsers: f:cephObjectStores: f:cephRBDMirror: f:cephToolbox: f:mirroring: f:network: f:multiClusterService: Manager: ocs-operator Operation: Update Time: 2023-06-08T06:06:04Z API Version: ocs.openshift.io/v1 Fields Type: FieldsV1 fieldsV1: f:status: .: f:conditions: f:externalSecretHash: f:images: .: f:ceph: .: f:desiredImage: f:noobaaCore: .: f:actualImage: f:desiredImage: f:noobaaDB: .: f:actualImage: f:desiredImage: f:kmsServerConnection: f:phase: f:relatedObjects: f:version: Manager: ocs-operator Operation: Update Subresource: status Time: 2023-06-08T06:25:06Z Owner References: API Version: odf.openshift.io/v1alpha1 Kind: StorageSystem Name: ocs-external-storagecluster-storagesystem UID: 712d0159-3df2-43b1-a611-eae1e9de5f0b Resource Version: 1402178 UID: dd67158d-dc54-4499-aa6c-38c0e70c75d4 Spec: Arbiter: Encryption: Kms: External Storage: Enable: true Label Selector: Managed Resources: Ceph Block Pools: Ceph Cluster: Ceph Config: Ceph Dashboard: Ceph Filesystems: Ceph Non Resilient Pools: Ceph Object Store Users: Ceph Object Stores: Ceph RBD Mirror: Ceph Toolbox: Mirroring: Network: Connections: Encryption: Multi Cluster Service: Status: Conditions: Last Heartbeat Time: 2023-06-08T06:06:06Z Last Transition Time: 2023-06-08T06:06:06Z Message: Version check successful Reason: VersionMatched Status: False Type: VersionMismatch Last Heartbeat Time: 2023-06-08T06:25:06Z Last Transition Time: 2023-06-08T06:06:06Z Message: Reconcile completed successfully Reason: ReconcileCompleted Status: True Type: ReconcileComplete Last Heartbeat Time: 2023-06-08T06:25:06Z Last Transition Time: 2023-06-08T06:08:28Z Message: Reconcile completed successfully Reason: ReconcileCompleted Status: True Type: Available Last Heartbeat Time: 2023-06-08T06:25:06Z Last Transition Time: 2023-06-08T06:08:28Z Message: Reconcile completed successfully Reason: ReconcileCompleted Status: False Type: Progressing Last Heartbeat Time: 2023-06-08T06:25:06Z Last Transition Time: 2023-06-08T06:06:06Z Message: Reconcile completed successfully Reason: ReconcileCompleted Status: False Type: Degraded Last Heartbeat Time: 2023-06-08T06:25:06Z Last Transition Time: 2023-06-08T06:08:28Z Message: Reconcile completed successfully Reason: ReconcileCompleted Status: True Type: Upgradeable External Secret Hash: 5fe48ad90d8bf08a712e2f5a67384c5ebf2348fca08124574afb4a77f0a5eb97e7841ad23c23ab7f180437ec01356789b94ca5aa6a73dca853a730cde0a63fc0 Images: Ceph: Desired Image: quay.io/rhceph-dev/rhceph@sha256:fa6d01cdef17bc32d2b95b8121b02f4d41adccc5ba8a9b95f38c97797ff6621f Noobaa Core: Actual Image: quay.io/rhceph-dev/odf4-mcg-core-rhel9@sha256:574b6258ee4d7ac2532b9143390a130acf89b34b77abc6c446a329286bcd27a5 Desired Image: quay.io/rhceph-dev/odf4-mcg-core-rhel9@sha256:574b6258ee4d7ac2532b9143390a130acf89b34b77abc6c446a329286bcd27a5 Noobaa DB: Actual Image: quay.io/rhceph-dev/rhel8-postgresql-12@sha256:f7f678d44d5934ed3d95c83b4428fee4b616f37e8eadc5049778f133b4ce3713 Desired Image: quay.io/rhceph-dev/rhel8-postgresql-12@sha256:f7f678d44d5934ed3d95c83b4428fee4b616f37e8eadc5049778f133b4ce3713 Kms Server Connection: Phase: Ready Related Objects: API Version: ceph.rook.io/v1 Kind: CephCluster Name: ocs-external-storagecluster-cephcluster Namespace: openshift-storage Resource Version: 1401943 UID: fe6a1e03-0a05-4ecf-9aae-c7b6565cd7b7 API Version: noobaa.io/v1alpha1 Kind: NooBaa Name: noobaa Namespace: openshift-storage Resource Version: 1402174 UID: 39ed0313-cb11-4020-8bad-e8afd812a047 Version: 4.13.0 Events: <none> [root@rdr-cicd-odf-ef1f-bastion-0 ~]# oc describe cephcluster -n openshift-storage Name: ocs-external-storagecluster-cephcluster Namespace: openshift-storage Labels: app=ocs-external-storagecluster Annotations: <none> API Version: ceph.rook.io/v1 Kind: CephCluster Metadata: Creation Timestamp: 2023-06-08T06:06:05Z Finalizers: cephcluster.ceph.rook.io Generation: 1 Managed Fields: API Version: ceph.rook.io/v1 Fields Type: FieldsV1 fieldsV1: f:metadata: f:labels: .: f:app: f:ownerReferences: .: k:{"uid":"dd67158d-dc54-4499-aa6c-38c0e70c75d4"}: f:spec: .: f:cephVersion: f:cleanupPolicy: .: f:sanitizeDisks: f:crashCollector: .: f:disable: f:dashboard: f:disruptionManagement: f:external: .: f:enable: f:healthCheck: .: f:daemonHealth: .: f:mon: f:osd: f:status: f:labels: .: f:monitoring: .: f:rook.io/managedBy: f:logCollector: f:mgr: f:mon: f:monitoring: .: f:enabled: f:externalMgrEndpoints: f:externalMgrPrometheusPort: f:network: .: f:connections: .: f:encryption: f:multiClusterService: f:security: .: f:keyRotation: .: f:enabled: f:kms: f:storage: Manager: ocs-operator Operation: Update Time: 2023-06-08T06:06:05Z API Version: ceph.rook.io/v1 Fields Type: FieldsV1 fieldsV1: f:metadata: f:finalizers: .: v:"cephcluster.ceph.rook.io": Manager: rook Operation: Update Time: 2023-06-08T06:06:12Z API Version: ceph.rook.io/v1 Fields Type: FieldsV1 fieldsV1: f:status: .: f:ceph: .: f:capacity: .: f:bytesAvailable: f:bytesTotal: f:bytesUsed: f:lastUpdated: f:fsid: f:health: f:lastChecked: f:versions: .: f:mds: .: f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): f:mgr: .: f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): f:mon: .: f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): f:osd: .: f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): f:overall: .: f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): f:rgw: .: f:ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): f:conditions: f:message: f:phase: f:state: f:version: .: f:version: Manager: rook Operation: Update Subresource: status Time: 2023-06-08T06:24:39Z Owner References: API Version: ocs.openshift.io/v1 Block Owner Deletion: true Controller: true Kind: StorageCluster Name: ocs-external-storagecluster UID: dd67158d-dc54-4499-aa6c-38c0e70c75d4 Resource Version: 1401943 UID: fe6a1e03-0a05-4ecf-9aae-c7b6565cd7b7 Spec: Ceph Version: Cleanup Policy: Sanitize Disks: Crash Collector: Disable: true Dashboard: Disruption Management: External: Enable: true Health Check: Daemon Health: Mon: Osd: Status: Labels: Monitoring: rook.io/managedBy: ocs-external-storagecluster Log Collector: Mgr: Mon: Monitoring: Enabled: true External Mgr Endpoints: Ip: 9.46.254.57 External Mgr Prometheus Port: 9283 Network: Connections: Encryption: Multi Cluster Service: Security: Key Rotation: Enabled: false Kms: Storage: Status: Ceph: Capacity: Bytes Available: 3198717943808 Bytes Total: 3221200306176 Bytes Used: 22482362368 Last Updated: 2023-06-08T06:24:37Z Fsid: 46ab550e-ca6e-11ed-af21-00000a0b13a5 Health: HEALTH_OK Last Checked: 2023-06-08T06:24:37Z Versions: Mds: ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 1 Mgr: ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 2 Mon: ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 5 Osd: ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 6 Overall: ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 16 Rgw: ceph version 16.2.10-138.el8cp (a63ae467c8e1f7503ea3855893f1e5ca189a71b9) pacific (stable): 2 Conditions: Last Heartbeat Time: 2023-06-08T06:06:12Z Last Transition Time: 2023-06-08T06:06:12Z Message: Attempting to connect to an external Ceph cluster Reason: ClusterConnecting Status: True Type: Connecting Last Heartbeat Time: 2023-06-08T06:24:39Z Last Transition Time: 2023-06-08T06:06:31Z Message: Cluster connected successfully Reason: ClusterConnected Status: True Type: Connected Message: Cluster connected successfully Phase: Connected State: Connected Version: Version: 16.2.10-138 Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal ReconcileSucceeded 18m rook-ceph-cluster-controller successfully configured CephCluster "openshift-storage/ocs-external-storagecluster-cephcluster" vSphere based deployment, external mode has the same issue: OC version: Client Version: 4.12.0-202208031327 Kustomize Version: v4.5.4 Server Version: 4.13.0-0.nightly-2023-06-12-231643 Kubernetes Version: v1.26.5+7d22122 OCS verison: ocs-operator.v4.13.0-rhodf OpenShift Container Storage 4.13.0-rhodf ocs-operator.v4.12.4-rhodf Succeeded Cluster version NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.13.0-0.nightly-2023-06-12-231643 True False 16h Cluster version is 4.13.0-0.nightly-2023-06-12-231643 Rook version: rook: v4.12.4-0.bc1e9806c3281090b58872e303e947ff5437c078 go: go1.18.10 Ceph version: ceph version 16.2.10-172.el8cp (00a157ecd158911ece116ae43095de793ed9f389) pacific (stable) ---------- screen-recording https://drive.google.com/file/d/1MtVp4PoPcY-rX40G0fV9jvdNhI2Tea6C/view?usp=sharing web console output http://pastebin.test.redhat.com/1102577 The first idea was to infact hide the tab but if we hide the tab then again we would get questions and possibly GSS bugs of customers asking why the topology tab is not showing. Giving them a message should eliminate that possible issue. WDYT @etamir @chandkum ? Topology is disabled in External mode cluster for IBM Power. Tried with build: 4.13.1-4 Attaching the screenshot as well. (In reply to Aaruni Aggarwal from comment #13) > Topology is disabled in External mode cluster for IBM Power. > Tried with build: 4.13.1-4 > > Attaching the screenshot as well. Moving the bug to verified based on this comment. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat OpenShift Data Foundation 4.13.1 security and bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2023:4437 |
Created attachment 1969855 [details] topology tab in external mode odf cluster Description of problem (please be detailed as possible and provide log snippests): Topology is missing inside Storage system in External mode ODF cluster. Version of all relevant components (if applicable): ODF: 4.13.0-207 OCP: 4.13 Does this issue impact your ability to continue to work with the product (please explain in detail what is the user impact)? Is there any workaround available to the best of your knowledge? No Rate from 1 - 5 the complexity of the scenario you performed that caused this bug (1 - very simple, 5 - very complex)? Can this issue reproducible? Yes Can this issue reproduce from the UI? Yes If this is a regression, please provide more details to justify this: Steps to Reproduce: 1.Create RHCS Cluster 2. CReate OCP cluster, install ODF Operator on it. 3. create storage system using external RHCS. 4. Navigate to Storage -> Data Foundation -> Topology Actual results: Topology is not appearing in topology tab. Expected results: Topology should be appearing in topology tab. Additional info: