Bug 2239101 - [RDR] DR monitoring dashboard is inconsistent in replicating data for subscription based workloads
Summary: [RDR] DR monitoring dashboard is inconsistent in replicating data for subscri...
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat OpenShift Data Foundation
Classification: Red Hat Storage
Component: management-console
Version: 4.14
Hardware: Unspecified
OS: Unspecified
unspecified
urgent
Target Milestone: ---
: ODF 4.14.0
Assignee: Timothy Asir
QA Contact: Aman Agrawal
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2023-09-15 08:05 UTC by Aman Agrawal
Modified: 2023-11-08 18:55 UTC (History)
5 users (show)

Fixed In Version: 4.14.0-148
Doc Type: No Doc Update
Doc Text:
Clone Of:
Environment:
Last Closed: 2023-11-08 18:54:31 UTC
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Github red-hat-storage odf-console pull 1061 0 None Merged Fix tooltips messages in the graphs at DR Policy dashboard 2023-10-11 06:57:44 UTC
Github red-hat-storage odf-console pull 1074 0 None Merged Bug 2239093: [release-4.14] Fix tooltips messages in the graphs at DR Policy dashboard 2023-10-11 07:19:28 UTC
Github red-hat-storage odf-console pull 1079 0 None Merged Bug 2239093: [release-4.14-compatibility] Fix tooltips messages in DR dashboard 2023-10-11 07:19:25 UTC
Red Hat Product Errata RHSA-2023:6832 0 None None None 2023-11-08 18:55:10 UTC

Description Aman Agrawal 2023-09-15 08:05:09 UTC
Description of problem (please be detailed as possible and provide log
snippests):


Version of all relevant components (if applicable):
ODF 4.14.0-132.stable
OCP 4.14.0-0.nightly-2023-09-02-132842
ACM 2.9.0-DOWNSTREAM-2023-08-24-09-30-12
ceph version 17.2.6-138.el9cp (b488c8dad42b2ecffcd96f3d76eeeecce48b8590) quincy (stable)


Does this issue impact your ability to continue to work with the product
(please explain in detail what is the user impact)?


Is there any workaround available to the best of your knowledge?


Rate from 1 - 5 the complexity of the scenario you performed that caused this
bug (1 - very simple, 5 - very complex)?


Can this issue reproducible?


Can this issue reproduce from the UI?


If this is a regression, please provide more details to justify this:


Steps to Reproduce:
1.On a RDR setup, configure DR monitoring dashboard by whitelisting below 2 metrics
ceph_rbd_mirror_snapshot_sync_bytes
ceph_rbd_mirror_snapshot_snapshots
2. Deploy both subscription and app-set based DR protected workloads on any of the managed clusters or both of them
3. On the ACM console, Data Services --> Data policies page
Select the cluster name under the cluster dropdown where workloads are deployed. Now under All applications dropdown, you will note that the subscription based application isn't listed (be it backed by rbd or cephfs). 
Also under the "Applications" label on the 2nd tile on the same page, the count of DR protected apps doesn't take subscription into consideration.

This is contradictory to the other data being shown on the dashboard as we raise VolumeSynchronizationDelay alert regardless of the app type, the 2 graphs- Snapshots synced and Replication throughput reflects data for both subscription as well as app-set based apps and doesn't filter out subscription based app data.

Please note, most of the RDR customers are using/and likely to use subscription based applications and this missing piece could lead to customer cases because if ** only ** subscription based applications are deployed, it would be missing from the list of app dropdown as well as count of DR protected app on the dashboard, however, the data will still reflect on the 2 graphs and alert will be raised (if applicable) and we should avoid being in situations like this. 


Actual results: DR monitoring dashboard is inconsistent in replicating data for subscription based workloads


Expected results: DR monitoring dashboard should be consistent in replicating data for subscription based workloads


Additional info:

Comment 7 Aman Agrawal 2023-09-22 08:12:27 UTC
It's a blocker hence providing qa_ack and requesting devel_ack.

Comment 8 Mudit Agarwal 2023-09-26 09:34:58 UTC
Can we have some ETA for this?

I don't understand the reasoning behind making a bug blocker for some cosmetic fixes but I was not part of the discussion mentioned in the above comment also.
If we can't fix this in the timelines then let's move it out.

Comment 12 Aman Agrawal 2023-10-18 06:39:25 UTC
With ODF 4.14.0-150.stable, the changes done on the dashboard are as below:

1. There is a question mark icon in the Alerts tile which has a helper text *** Alerts are displayed for both ApplicationSet and Subscription type applications. ***.

2. There is a question mark icon in the Applications tile which has a helper text *** The applications count displays the total number of ApplicationSet type applications in all disaster recovery configured clusters. ***.

3. The application dropdown next to cluster name now says *** Application: All ApplicationSet *** to convey that only ApplicationSet types applications would be listed for each cluster, sum of which matches to the Applications count from the above Applications tile. 

4. Text *** Total applications (ApplicationSet) *** conveys the right message that it only shows count for ApplicationSet types applications.

5. Text *** Protected PVCs (ApplicationSet) *** conveys the right message that it only shows Protected PVCs count for ApplicationSet types applications.

6. Text *** Volume replication health (ApplicationSet) *** conveys the right message that it shows replication health for ApplicationSet types applications only.

All these changes LTGM.

Comment 14 errata-xmlrpc 2023-11-08 18:54:31 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Important: Red Hat OpenShift Data Foundation 4.14.0 security, enhancement & bug fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHSA-2023:6832


Note You need to log in before you can comment on or make changes to this bug.