Bug 2089347

Summary: ODF to ODF: provider cluster should not have the default rbd and cephfs Storageclasses
Product: [Red Hat Storage] Red Hat OpenShift Data Foundation Reporter: Neha Berry <nberry>
Component: odf-managed-serviceAssignee: Dhruv Bindra <dbindra>
Status: CLOSED CURRENTRELEASE QA Contact: Neha Berry <nberry>
Severity: high Docs Contact:
Priority: unspecified    
Version: 4.10CC: aeyal, dbindra, ocs-bugs, odf-bz-bot
Target Milestone: ---   
Target Release: ---   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2022-09-28 06:49:06 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Description Neha Berry 2022-05-23 13:31:20 UTC
Description of problem:
============================
In ODF to ODF provider cluster, one must not be allowed to create PVCs using the default rbd and cephfs SC which get created on ODF install

With deployer 2.0.0 , this issue was addressed via PR https://github.com/red-hat-storage/ocs-osd-deployer/pull/133/files , however the fix worked partially and below were the observations

1. the CSI pods(provisioner and plugin) are no longer created in the provider
2. the default SC and cephblockpool still exists

NAME                          PROVISIONER                             RECLAIMPOLICY   VOLUMEBINDINGMODE      ALLOWVOLUMEEXPANSION   AGE
gp2 (default)                 kubernetes.io/aws-ebs                   Delete          WaitForFirstConsumer   true                   6h16m
gp2-csi                       ebs.csi.aws.com                         Delete          WaitForFirstConsumer   true                   6h16m
ocs-storagecluster-ceph-rbd   openshift-storage.rbd.csi.ceph.com      Delete          Immediate              true                   4h52m
ocs-storagecluster-cephfs     openshift-storage.cephfs.csi.ceph.com   Delete          Immediate              true                   4h52m

#oc get cephblockpool
cephblockpool-storageconsumer-35b116bc-7b87-4e20-922f-2a9af20a2563.yaml*
ocs-storagecluster-cephblockpool.yaml*

Though as part of this PR, both SC and CSI pods should not have got created

Version-Release number of selected component (if applicable):
=================================================================
deployer v2.0.0 onwards 

 oc get clusterversion
NAME      VERSION   AVAILABLE   PROGRESSING   SINCE   STATUS
version   4.10.13   True        False         4d9h    Error while reconciling 4.10.13: the cluster operator insights is degraded
➜  ~ oc get csv -n openshift-storage -o json ocs-operator.v4.10.2 | jq '.metadata.labels["full_version"]'                                                                                                                      
"4.10.2-3"

➜  ~ echo "Deployer"; oc describe csv ocs-osd-deployer.v2.0.2|grep -i image
Deployer
    Mediatype:   image/svg+xml
                Image:  quay.io/openshift/origin-kube-rbac-proxy:4.10.0
                Image:             quay.io/osd-addons/ocs-osd-deployer:2.0.2-2
                Image:             quay.io/osd-addons/ocs-osd-deployer:2.0.2-2



How reproducible:
======================
Always


Steps to Reproduce:
=======================
1.IN ROSA for a ODF to ODF offering, install a provider cluster using :
A) Addon based deployment - install OCP and then install provder-qe addon
b) Install a managed services appliance mode provider using rosa create service

2. Check the presence of RBD and cephfs storageclasses
3.

Actual results:
======================
Till deployer 2.0.1 -> both SC present
From deployer 2.0.2 + ODF 4.10.2 -> cephFS SC present but RBD SC is not created as default pool "ocs-storagecluster-cephblockpool" is no longer created due to fix of Bug 2078715
Storagecluster stays in Error state - https://bugzilla.redhat.com/show_bug.cgi?id=2089296

Expected results:
========================
Default SC + CSI pods both should not get created


Additional info:
===========================
                   



oc get sc,cephblockpool,pods
NAME                                                      PROVISIONER                             RECLAIMPOLICY   VOLUMEBINDINGMODE      ALLOWVOLUMEEXPANSION   AGE
storageclass.storage.k8s.io/gp2 (default)                 kubernetes.io/aws-ebs                   Delete          WaitForFirstConsumer   true                   4d9h
storageclass.storage.k8s.io/gp2-csi                       ebs.csi.aws.com                         Delete          WaitForFirstConsumer   true                   4d9h
storageclass.storage.k8s.io/gp3-csi                       ebs.csi.aws.com                         Delete          WaitForFirstConsumer   true                   4d9h
storageclass.storage.k8s.io/ocs-storagecluster-ceph-rbd   openshift-storage.rbd.csi.ceph.com      Delete          Immediate              true                   4d9h
storageclass.storage.k8s.io/ocs-storagecluster-cephfs     openshift-storage.cephfs.csi.ceph.com   Delete          Immediate              true                   4d9h

NAME                                                                                            AGE
cephblockpool.ceph.rook.io/cephblockpool-storageconsumer-5b716509-9514-499d-83b8-02653f013014   4d7h
cephblockpool.ceph.rook.io/cephblockpool-storageconsumer-c00d3d25-e898-4bbd-96a3-18465f0f7d46   4d7h
cephblockpool.ceph.rook.io/ocs-storagecluster-cephblockpool                                     4d9h

NAME                                                                  READY   STATUS      RESTARTS   AGE
pod/8b9812e03cb80021cb5974ee1876bd40009a23b678f9e6b1e3e63b98787l4fw   0/1     Completed   0          3d22h
pod/94963c92ac0decbfa826f438b1d9e5f555b46a84bda92fc43de6a5ed7a5w2sg   0/1     Completed   0          4d9h
pod/96f4060767c209c4e921c86f13b9ad5e1e84606de95ede4f448af2bdeekv7vp   0/1     Completed   0          4d9h
pod/addon-ocs-provider-qe-catalog-vcrnz                               1/1     Running     0          4d6h
pod/alertmanager-managed-ocs-alertmanager-0                           2/2     Running     0          4d9h
pod/alertmanager-managed-ocs-alertmanager-1                           2/2     Running     0          4d9h
pod/alertmanager-managed-ocs-alertmanager-2                           2/2     Running     0          4d9h
pod/csi-addons-controller-manager-7d99f546c9-bl9sn                    2/2     Running     0          3d22h
pod/ocs-metrics-exporter-5dcf6f88df-zlhsd                             1/1     Running     0          3d22h
pod/ocs-operator-5985b8b5f4-tn5ss                                     1/1     Running     0          3d22h
pod/ocs-osd-controller-manager-6b74c4cc67-4j5lr                       3/3     Running     0          3d22h
pod/ocs-provider-server-86d8bf774-5p67d                               1/1     Running     0          3d22h
pod/odf-console-58f6b6f5bb-lm95n                                      1/1     Running     0          3d22h
pod/odf-operator-controller-manager-584df64f8-mwpk6                   2/2     Running     0          3d22h
pod/prometheus-managed-ocs-prometheus-0                               3/3     Running     0          3d22h
pod/prometheus-operator-8547cc9f89-7mmbb                              1/1     Running     0          4d9h
pod/rook-ceph-crashcollector-ip-10-0-131-67.ec2.internal-bd6b5ll5q9   1/1     Running     0          3d22h
pod/rook-ceph-crashcollector-ip-10-0-159-28.ec2.internal-7d488dnhd8   1/1     Running     0          3d22h
pod/rook-ceph-crashcollector-ip-10-0-173-158.ec2.internal-56559j565   1/1     Running     0          3d22h
pod/rook-ceph-mds-ocs-storagecluster-cephfilesystem-a-865579858m9s9   2/2     Running     0          4d9h
pod/rook-ceph-mds-ocs-storagecluster-cephfilesystem-b-6869bf77p2tgc   2/2     Running     0          4d9h
pod/rook-ceph-mgr-a-5ff964d8b-msfxg                                   2/2     Running     0          4d9h
pod/rook-ceph-mon-a-5d766555cb-k7d7m                                  2/2     Running     0          4d9h
pod/rook-ceph-mon-b-656774f7dc-88bh9                                  2/2     Running     0          4d9h
pod/rook-ceph-mon-c-5bcb75c5b5-p5mkn                                  2/2     Running     0          4d9h
pod/rook-ceph-operator-5678fcf74-p72j6                                1/1     Running     0          3d22h
pod/rook-ceph-osd-0-6b57fbff84-b6svt                                  2/2     Running     0          3d22h
pod/rook-ceph-osd-1-646698bddc-zr8tk                                  2/2     Running     0          3d22h
pod/rook-ceph-osd-10-b77dc4df4-q464f                                  2/2     Running     0          3d22h
pod/rook-ceph-osd-11-85dff77676-j4f6c                                 2/2     Running     0          3d22h
pod/rook-ceph-osd-12-64cfb98575-zkczt                                 2/2     Running     0          3d22h
pod/rook-ceph-osd-13-7c449bbccd-stwsq                                 2/2     Running     0          3d22h
pod/rook-ceph-osd-14-5594c64997-ndx4r                                 2/2     Running     0          3d22h
pod/rook-ceph-osd-2-685c5699db-w6g4b                                  2/2     Running     0          3d22h
pod/rook-ceph-osd-3-7d7884d954-lpm4d                                  2/2     Running     0          3d22h
pod/rook-ceph-osd-4-7ff64fbbdc-lx4qv                                  2/2     Running     0          3d22h
pod/rook-ceph-osd-5-54896756d8-gjt9v                                  2/2     Running     0          3d22h
pod/rook-ceph-osd-6-74cb644645-xx7s4                                  2/2     Running     0          3d22h
pod/rook-ceph-osd-7-787cdbc4b-kgh7q                                   2/2     Running     0          3d22h
pod/rook-ceph-osd-8-66576858-gx8kn                                    2/2     Running     0          3d22h
pod/rook-ceph-osd-9-85cbf74b95-lh6hk                                  2/2     Running     0          3d22h