Bug 1993918
| Summary: | CSI drivers are not getting created | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat OpenShift Data Foundation | Reporter: | Bipul Adhikari <badhikar> |
| Component: | csi-driver | Assignee: | Yug Gupta <ygupta> |
| Status: | CLOSED NOTABUG | QA Contact: | Elad <ebenahar> |
| Severity: | medium | Docs Contact: | |
| Priority: | unspecified | ||
| Version: | 4.9 | CC: | hchiramm, madam, ocs-bugs, odf-bz-bot, ygupta |
| Target Milestone: | --- | ||
| Target Release: | --- | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2021-08-17 06:59:38 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Bipul Adhikari
2021-08-16 12:11:12 UTC
Just to conclude quick findings from this cluster: 1) #oc get csidriver does not list the ceph csi drivers But it list the aws csi driver ( note that its on v1 version) 2) We have done enhancements in Rook, where we recreate the CSI driver object based on the cluster version It delete the Betav1 csi driver and create v1 version if cluster is >1.17 3) We can manually create the CSI driver with v1 yaml which get picked up here. betav1 CSIDriver object is not supported from Kubernetes v1.22. For that, we recently added a check in rook so that:
1. If Kubernetes version >= v1.18 and Kubernetes version <= v1.21 then we delete the betav1 CSIDriver (if present no-op otherwise)
2. For If Kubernetes version >= v1.22, directly create the v1 CSIDriver
Although the csi driver is present, the CSIDriver object is not listed when `oc get CSIDriver` is done.
When I did local testing on minikube with k8s v1.22 and Rook 1.7.0, the v1CSIDriver object comes up properly without any issues.
```
$ kubectl version Client Version: version.Info{Major:"1", Minor:"22", GitVersion:"v1.22.0", GitCommit:"c2b5237ccd9c0f1d600d3072634ca66cefdf272f", GitTreeState:"clean", BuildDate:"2021-08-04T18:03:20Z", GoVersion:"go1.16.6", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"22", GitVersion:"v1.22.0", GitCommit:"c2b5237ccd9c0f1d600d3072634ca66cefdf272f", GitTreeState:"clean", BuildDate:"2021-08-04T17:57:25Z", GoVersion:"go1.16.6", Compiler:"gc", Platform:"linux/amd64"}
```
```
[yuggupta@fedora ceph-csi](test-rook-dep)$ kubectl get CSIDriver NAME ATTACHREQUIRED PODINFOONMOUNT STORAGECAPACITY TOKENREQUESTS REQUIRESREPUBLISH MODES AGE rook-ceph.cephfs.csi.ceph.com true false false <unset> false Persistent 2m16s rook-ceph.rbd.csi.ceph.com true false false <unset> false Persistent 2m17s
```
Bipul can you please share the exact OCP build and the rook-ceph-operator logs so that we can debug it further?
Looks like the cluster got destroyed while we started to dig further. @Bipul, please provide the access details, if you are hitting this again in your next test run? On verifying the rook version, the rook version being used is `rook/ceph:v1.6.5-2.gb78358e` which is a 2 months old image and doesn't contain my fix for the same in rook upstream https://github.com/rook/rook/pull/8029 . Since this image doesn't have the fix, this is expected to happen. Let me know if you face any issues in the rook downstream image based on v1.7. Closing this a not a bug, feel free to re-open if similar behavior is seen for downstream rook images which has the fix https://github.com/rook/rook/pull/8029. |