Created attachment 2076120 [details] odf must gather *Description of problem -* rook-ceph-nfs-ocs-storagecluster-cephnfs-X pod is in CrashLoopBackOff state after enabling nfs. *The OCP platform infrastructure and deployment type (AWS, Bare Metal, VMware, etc. Please clarify if it is platform agnostic deployment), (IPI/UPI):* VMware UPI *The ODF deployment type (Internal, External, Internal-Attached (LSO), Multicluster, DR, Provider, etc):* Internal *The version of all relevant components (OCP, ODF, RHCS, ACM whichever is applicable):* OCP: 4.18.0-0.nightly-2025-02-06-092713 ODF: 4.18 {{ceph version 19.2}} *Does this issue impact your ability to continue to work with the product?* *Is there any workaround available to the best of your knowledge?* *Can this issue be reproduced? If so, please provide the hit rate* Yes, 100% *Can this issue be reproduced from the UI?* *If this is a regression, please provide more details to justify this:* Yes, it has passed in 4.17 *Steps to Reproduce:* 1. Create ODF cluster 2. Enable nfs feature using patch command oc patch -n openshift-storage storageclusters.ocs.openshift.io ocs-storagecluster --patch '{"spec": {"nfs": {"enable": true} }}' --type merge Check csi-nfs pods came up - {code:java} oc get pods | grep nfs csi-nfsplugin-m8hkm 3/3 Running 0 18s csi-nfsplugin-phqpb 3/3 Running 0 18s csi-nfsplugin-provisioner-5f94bb6f64-7htn8 6/6 Running 0 18s csi-nfsplugin-provisioner-5f94bb6f64-dw68c 6/6 Running 0 18s csi-nfsplugin-vgcj4 3/3 Running 0 18s {code} 3. Enable csi_nfs oc patch cm rook-ceph-operator-config -n openshift-storage -p $'data:\n "ROOK_CSI_ENABLE_NFS": "true"' 4. Check rook-ceph-nfs-ocs-storagecluster-cephnfs pod came up - {code:java} oc get pods | grep nfs csi-nfsplugin-m8hkm 3/3 Running 0 62s csi-nfsplugin-phqpb 3/3 Running 0 62s csi-nfsplugin-provisioner-5f94bb6f64-7htn8 6/6 Running 0 62s csi-nfsplugin-provisioner-5f94bb6f64-dw68c 6/6 Running 0 62s csi-nfsplugin-vgcj4 3/3 Running 0 62s rook-ceph-nfs-ocs-storagecluster-cephnfs-a-585f8b9796-pwcdw 1/2 CrashLoopBackOff 2 (19s ago) 43s {code} *The exact date and time when the issue was observed, including timezone details:* *Actual results:* rook-ceph-nfs-ocs-storagecluster-cephnfs pod is in CrashLoopBackOff state *Expected results:* rook-ceph-nfs-ocs-storagecluster-cephnfs pod should be in running state Additional Info: {{From regression runs we see that this issue was not faced for,}} {{ceph version--18.2.1-278.el9cp (2ae16095654f99a1a043ca3f0c7befcb78080058) reef (stable)}} {{ }} {{[regression-test-logs|http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/j393ai3c33ua/j393ai3c33ua_20250131T072303/logs/test_report_1738325989.html]}} {{ }} {{For odf 4.18 also this ceph version is working fine.}} {{ocp: 4.18.0-0.nightly-2025-02-11-134047}} {{odf: 4.18.0-124}} {{ceph image used : registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:cea147a49680256b8f3ec9b1e3a723f5e040cd0d722b1f18086f6959f71ffaf8}} {{oc get pods -n openshift-storage| grep nfs}} {{rook-ceph-nfs-ocs-storagecluster-cephnfs-a-57776f94df-9r5q8 2/2 Running 0 17m}} ODF bug: https://issues.redhat.com/browse/DFBUGS-1592
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Important: Red Hat Ceph Storage 8.1 security, bug fix, and enhancement updates), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2025:9775