Bug 2297384 - [RDR] [Discovered Apps] Lack of clear error messaging for disabled volsync feature [NEEDINFO]
Summary: [RDR] [Discovered Apps] Lack of clear error messaging for disabled volsync fe...
Keywords:
Status: ASSIGNED
Alias: None
Product: Red Hat OpenShift Data Foundation
Classification: Red Hat Storage
Component: odf-dr
Version: 4.16
Hardware: Unspecified
OS: Unspecified
unspecified
medium
Target Milestone: ---
: ---
Assignee: Raghavendra Talur
QA Contact: krishnaram Karthick
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2024-07-11 17:29 UTC by Sidhant Agrawal
Modified: 2024-10-09 12:17 UTC (History)
4 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed:
Embargoed:
sheggodu: needinfo? (rtalur)


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Issue Tracker OCSBZM-8681 0 None None None 2024-10-09 12:17:27 UTC

Description Sidhant Agrawal 2024-07-11 17:29:46 UTC
Description of problem (please be detailed as possible and provide log
snippests):
Currently, in fresh RDR setup where volsync with CephFS is disabled by default for discovered apps (Refer https://bugzilla.redhat.com/show_bug.cgi?id=2291255), there is a lack of clear error messaging which says that volsync is disabled.

Version of all relevant components (if applicable):
ODF: 4.16.0-rhodf (full_version=4.16.0-134)

Does this issue impact your ability to continue to work with the product
(please explain in detail what is the user impact)?
No

Is there any workaround available to the best of your knowledge?
No

Rate from 1 - 5 the complexity of the scenario you performed that caused this
bug (1 - very simple, 5 - very complex)?
2

Can this issue reproducible?
Yes

Can this issue reproduce from the UI?


If this is a regression, please provide more details to justify this:


Steps to Reproduce:
1. Configure RDR
2. Deploy cephfs based application without using the RHACM create application capability
3. Enroll the application for DR protection using Discovered app UI via Protected applications tab

Observation: Workload remains unprotected. In drpc yaml output Protected condition reported as False. VRG state remains Unknown.
However, there is no clear error message for user that volsync functionality has been disabled for discovered apps.

$ oc -n openshift-dr-ops get drpc test-cephfs-discover -o wide
NAME                   AGE   PREFERREDCLUSTER   FAILOVERCLUSTER   DESIREDSTATE   CURRENTSTATE   PROGRESSION   START TIME             DURATION       PEER READY
test-cephfs-discover   16m   sagrawal-c1                                         Deployed       Completed     2024-07-10T09:41:16Z   18.05208435s   True

- lastTransitionTime: "2024-07-10T09:41:05Z"
    message: VolumeReplicationGroup (openshift-dr-ops/test-cephfs-discover) on cluster
      sagrawal-c1 is reporting errors (All PVCs of the VolumeReplicationGroup are
      not ready) readying workload data, retrying till DataReady condition is met
    observedGeneration: 2
    reason: Error
    status: "False"
    type: Protected


$ oc -n openshift-dr-ops get vrg -o wide
NAME                   DESIREDSTATE   CURRENTSTATE
test-cephfs-discover   primary        Unknown


Actual results:
No clear error message that volsync is disabled for discovered apps.

Expected results:
Appropriate error message when volsync is disabled for discovered apps.

Additional info:

DRPC and VRG output:

$ oc -n openshift-dr-ops get drpc test-cephfs-discover -o yaml
apiVersion: ramendr.openshift.io/v1alpha1
kind: DRPlacementControl
metadata:
  annotations:
    drplacementcontrol.ramendr.openshift.io/app-namespace: openshift-dr-ops
    drplacementcontrol.ramendr.openshift.io/last-app-deployment-cluster: sagrawal-c1
  creationTimestamp: "2024-07-10T09:41:03Z"
  finalizers:
  - drpc.ramendr.openshift.io/finalizer
  generation: 2
  labels:
    cluster.open-cluster-management.io/backup: ramen
  name: test-cephfs-discover
  namespace: openshift-dr-ops
  ownerReferences:
  - apiVersion: cluster.open-cluster-management.io/v1beta1
    blockOwnerDeletion: true
    controller: true
    kind: Placement
    name: test-cephfs-discover-placement-1
    uid: 741dd5e0-6bf3-4634-bed6-325113e33775
  resourceVersion: "14849404"
  uid: 2a0d3072-9490-4de8-9814-fde529fe7567
spec:
  drPolicyRef:
    apiVersion: ramendr.openshift.io/v1alpha1
    kind: DRPolicy
    name: odr-policy-5m
  kubeObjectProtection:
    captureInterval: 5m0s
    kubeObjectSelector:
      matchExpressions:
      - key: app-name
        operator: In
        values:
        - app-busybox-cephfs-1
  placementRef:
    apiVersion: cluster.open-cluster-management.io/v1beta1
    kind: Placement
    name: test-cephfs-discover-placement-1
    namespace: openshift-dr-ops
  preferredCluster: sagrawal-c1
  protectedNamespaces:
  - test-cephfs-discover
  pvcSelector:
    matchExpressions:
    - key: app-name
      operator: In
      values:
      - app-busybox-cephfs-1
status:
  actionDuration: 18.05208435s
  actionStartTime: "2024-07-10T09:41:16Z"
  conditions:
  - lastTransitionTime: "2024-07-10T09:41:04Z"
    message: Initial deployment completed
    observedGeneration: 2
    reason: Deployed
    status: "True"
    type: Available
  - lastTransitionTime: "2024-07-10T09:41:04Z"
    message: Ready
    observedGeneration: 2
    reason: Success
    status: "True"
    type: PeerReady
  - lastTransitionTime: "2024-07-10T09:41:05Z"
    message: VolumeReplicationGroup (openshift-dr-ops/test-cephfs-discover) on cluster
      sagrawal-c1 is reporting errors (All PVCs of the VolumeReplicationGroup are
      not ready) readying workload data, retrying till DataReady condition is met
    observedGeneration: 2
    reason: Error
    status: "False"
    type: Protected
  lastKubeObjectProtectionTime: "2024-07-10T09:46:23Z"
  lastUpdateTime: "2024-07-10T09:50:34Z"
  observedGeneration: 2
  phase: Deployed
  preferredDecision:
    clusterName: sagrawal-c1
    clusterNamespace: sagrawal-c1
  progression: Completed
  resourceConditions:
    conditions:
    - lastTransitionTime: "2024-07-10T09:41:05Z"
      message: All PVCs of the VolumeReplicationGroup are not ready
      observedGeneration: 1
      reason: Error
      status: "False"
      type: DataReady
    - lastTransitionTime: "2024-07-10T09:41:05Z"
      message: 'PVCs in the VolumeReplicationGroup are data protected '
      observedGeneration: 1
      reason: DataProtected
      status: "True"
      type: DataProtected
    - lastTransitionTime: "2024-07-10T09:41:04Z"
      message: Nothing to restore
      observedGeneration: 1
      reason: Restored
      status: "True"
      type: ClusterDataReady
    - lastTransitionTime: "2024-07-10T09:41:12Z"
      message: Cluster data of all PVs are protected. Kube objects protected
      observedGeneration: 1
      reason: Uploaded
      status: "True"
      type: ClusterDataProtected
    resourceMeta:
      generation: 1
      kind: VolumeReplicationGroup
      name: test-cephfs-discover
      namespace: openshift-dr-ops
      resourceVersion: "16910300"

$ oc -n openshift-dr-ops get vrg test-cephfs-discover -o yaml
apiVersion: ramendr.openshift.io/v1alpha1
kind: VolumeReplicationGroup
metadata:
  annotations:
    drplacementcontrol.ramendr.openshift.io/destination-cluster: sagrawal-c1
    drplacementcontrol.ramendr.openshift.io/do-not-delete-pvc: ""
    drplacementcontrol.ramendr.openshift.io/drpc-uid: 2a0d3072-9490-4de8-9814-fde529fe7567
  creationTimestamp: "2024-07-10T09:41:04Z"
  finalizers:
  - volumereplicationgroups.ramendr.openshift.io/vrg-protection
  generation: 1
  name: test-cephfs-discover
  namespace: openshift-dr-ops
  ownerReferences:
  - apiVersion: work.open-cluster-management.io/v1
    kind: AppliedManifestWork
    name: e6773b354f76c66c8ce2dc1d425ba2fd1de7369c07b57a46ffd25d0de1c499cd-test-cephfs-discover-openshift-dr-ops-vrg-mw
    uid: 5e55d307-ac00-4dd6-9a2c-3e3446d6f82e
  resourceVersion: "16916422"
  uid: 1fac4453-71a4-446d-977e-a29125952cac
spec:
  async:
    replicationClassSelector: {}
    schedulingInterval: 5m
    volumeSnapshotClassSelector: {}
  kubeObjectProtection:
    captureInterval: 5m0s
    kubeObjectSelector:
      matchExpressions:
      - key: app-name
        operator: In
        values:
        - app-busybox-cephfs-1
  protectedNamespaces:
  - test-cephfs-discover
  pvcSelector:
    matchExpressions:
    - key: app-name
      operator: In
      values:
      - app-busybox-cephfs-1
  replicationState: primary
  s3Profiles:
  - s3profile-sagrawal-c1-ocs-storagecluster
  - s3profile-sagrawal-c2-ocs-storagecluster
  volSync:
    disabled: true
status:
  conditions:
  - lastTransitionTime: "2024-07-10T09:41:05Z"
    message: All PVCs of the VolumeReplicationGroup are not ready
    observedGeneration: 1
    reason: Error
    status: "False"
    type: DataReady
  - lastTransitionTime: "2024-07-10T09:41:05Z"
    message: 'PVCs in the VolumeReplicationGroup are data protected '
    observedGeneration: 1
    reason: DataProtected
    status: "True"
    type: DataProtected
  - lastTransitionTime: "2024-07-10T09:41:04Z"
    message: Nothing to restore
    observedGeneration: 1
    reason: Restored
    status: "True"
    type: ClusterDataReady
  - lastTransitionTime: "2024-07-10T09:41:12Z"
    message: Cluster data of all PVs are protected. Kube objects protected
    observedGeneration: 1
    reason: Uploaded
    status: "True"
    type: ClusterDataProtected
  kubeObjectProtection:
    captureToRecoverFrom:
      endTime: "2024-07-10T09:51:24Z"
      number: 1
      startGeneration: 1
      startTime: "2024-07-10T09:51:18Z"
  lastUpdateTime: "2024-07-10T09:55:19Z"
  observedGeneration: 1
  state: Unknown


Note You need to log in before you can comment on or make changes to this bug.