Bug 2345193

Summary: rook-ceph-nfs pod is in CrashLoopBackOff state after enabling nfs.
Product: [Red Hat Storage] Red Hat Ceph Storage Reporter: Mudit Agarwal <muagarwa>
Component: NFS-GaneshaAssignee: Sachin Punadikar <spunadik>
Status: CLOSED ERRATA QA Contact: Manisha Saini <msaini>
Severity: urgent Docs Contact:
Priority: unspecified    
Version: 8.0CC: assingh, cephqe-warriors, kkeithle, mrajanna, muagarwa, tserlin
Target Milestone: ---Keywords: Regression
Target Release: 8.1   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: nfs-ganesha-6.5-9.el9cp Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2025-06-26 12:25:26 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Attachments:
Description Flags
odf must gather none

Description Mudit Agarwal 2025-02-12 13:45:05 UTC
Created attachment 2076120 [details]
odf must gather

*Description of problem -* 

rook-ceph-nfs-ocs-storagecluster-cephnfs-X pod is in CrashLoopBackOff state after enabling nfs.

 

*The OCP platform infrastructure and deployment type (AWS, Bare Metal, VMware, etc. Please clarify if it is platform agnostic deployment), (IPI/UPI):*

VMware UPI

 

*The ODF deployment type (Internal, External, Internal-Attached (LSO), Multicluster, DR, Provider, etc):*

Internal

*The version of all relevant components (OCP, ODF, RHCS, ACM whichever is applicable):*

OCP: 4.18.0-0.nightly-2025-02-06-092713

ODF: 4.18
{{ceph version 19.2}}

*Does this issue impact your ability to continue to work with the product?*

 

*Is there any workaround available to the best of your knowledge?*

 

*Can this issue be reproduced? If so, please provide the hit rate*

Yes, 100%

*Can this issue be reproduced from the UI?*

 

*If this is a regression, please provide more details to justify this:*

Yes, it has passed in 4.17

*Steps to Reproduce:*

1. Create ODF cluster

2. Enable nfs feature using patch command

oc patch -n openshift-storage storageclusters.ocs.openshift.io ocs-storagecluster --patch '{"spec": {"nfs":

{"enable": true}

}}' --type merge

Check csi-nfs pods came up - 

 
{code:java}
oc get pods | grep nfs
csi-nfsplugin-m8hkm                                               3/3     Running     0               18s
csi-nfsplugin-phqpb                                               3/3     Running     0               18s
csi-nfsplugin-provisioner-5f94bb6f64-7htn8                        6/6     Running     0               18s
csi-nfsplugin-provisioner-5f94bb6f64-dw68c                        6/6     Running     0               18s
csi-nfsplugin-vgcj4                                               3/3     Running     0               18s
 
{code}
3. Enable csi_nfs
oc patch cm rook-ceph-operator-config -n openshift-storage -p $'data:\n "ROOK_CSI_ENABLE_NFS":  "true"'

4. Check rook-ceph-nfs-ocs-storagecluster-cephnfs pod came up -

 
{code:java}
oc get pods | grep nfs
csi-nfsplugin-m8hkm                                               3/3     Running            0               62s
csi-nfsplugin-phqpb                                               3/3     Running            0               62s
csi-nfsplugin-provisioner-5f94bb6f64-7htn8                        6/6     Running            0               62s
csi-nfsplugin-provisioner-5f94bb6f64-dw68c                        6/6     Running            0               62s
csi-nfsplugin-vgcj4                                               3/3     Running            0               62s
rook-ceph-nfs-ocs-storagecluster-cephnfs-a-585f8b9796-pwcdw       1/2     CrashLoopBackOff   2 (19s ago)     43s
{code}
 

*The exact date and time when the issue was observed, including timezone details:*

 

*Actual results:*

rook-ceph-nfs-ocs-storagecluster-cephnfs pod is in CrashLoopBackOff state

 

*Expected results:*

rook-ceph-nfs-ocs-storagecluster-cephnfs pod should be in running state

Additional Info:

{{From regression runs we see that this issue was not faced for,}}
{{ceph version--18.2.1-278.el9cp (2ae16095654f99a1a043ca3f0c7befcb78080058) reef (stable)}}
{{ }}
{{[regression-test-logs|http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/j393ai3c33ua/j393ai3c33ua_20250131T072303/logs/test_report_1738325989.html]}}
{{ }}
{{For odf 4.18 also this ceph version is working fine.}}

{{ocp: 4.18.0-0.nightly-2025-02-11-134047}}
{{odf: 4.18.0-124}}
{{ceph image used : registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:cea147a49680256b8f3ec9b1e3a723f5e040cd0d722b1f18086f6959f71ffaf8}}

{{oc get pods -n openshift-storage| grep nfs}}
{{rook-ceph-nfs-ocs-storagecluster-cephnfs-a-57776f94df-9r5q8 2/2 Running 0 17m}}


ODF bug: https://issues.redhat.com/browse/DFBUGS-1592

Comment 12 errata-xmlrpc 2025-06-26 12:25:26 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Important: Red Hat Ceph Storage 8.1 security, bug fix, and enhancement updates), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHSA-2025:9775