Bug 2345193 - rook-ceph-nfs pod is in CrashLoopBackOff state after enabling nfs.
Summary: rook-ceph-nfs pod is in CrashLoopBackOff state after enabling nfs.
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Ceph Storage
Classification: Red Hat Storage
Component: NFS-Ganesha
Version: 8.0
Hardware: Unspecified
OS: Unspecified
unspecified
urgent
Target Milestone: ---
: 8.1
Assignee: Sachin Punadikar
QA Contact: Manisha Saini
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2025-02-12 13:45 UTC by Mudit Agarwal
Modified: 2025-06-26 12:25 UTC (History)
6 users (show)

Fixed In Version: nfs-ganesha-6.5-9.el9cp
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2025-06-26 12:25:26 UTC
Embargoed:


Attachments (Terms of Use)
odf must gather (19.01 MB, application/gzip)
2025-02-12 13:45 UTC, Mudit Agarwal
no flags Details


Links
System ID Private Priority Status Summary Last Updated
Red Hat Issue Tracker RHCEPH-10600 0 None None None 2025-02-12 13:46:46 UTC
Red Hat Product Errata RHSA-2025:9775 0 None None None 2025-06-26 12:25:29 UTC

Description Mudit Agarwal 2025-02-12 13:45:05 UTC
Created attachment 2076120 [details]
odf must gather

*Description of problem -* 

rook-ceph-nfs-ocs-storagecluster-cephnfs-X pod is in CrashLoopBackOff state after enabling nfs.

 

*The OCP platform infrastructure and deployment type (AWS, Bare Metal, VMware, etc. Please clarify if it is platform agnostic deployment), (IPI/UPI):*

VMware UPI

 

*The ODF deployment type (Internal, External, Internal-Attached (LSO), Multicluster, DR, Provider, etc):*

Internal

*The version of all relevant components (OCP, ODF, RHCS, ACM whichever is applicable):*

OCP: 4.18.0-0.nightly-2025-02-06-092713

ODF: 4.18
{{ceph version 19.2}}

*Does this issue impact your ability to continue to work with the product?*

 

*Is there any workaround available to the best of your knowledge?*

 

*Can this issue be reproduced? If so, please provide the hit rate*

Yes, 100%

*Can this issue be reproduced from the UI?*

 

*If this is a regression, please provide more details to justify this:*

Yes, it has passed in 4.17

*Steps to Reproduce:*

1. Create ODF cluster

2. Enable nfs feature using patch command

oc patch -n openshift-storage storageclusters.ocs.openshift.io ocs-storagecluster --patch '{"spec": {"nfs":

{"enable": true}

}}' --type merge

Check csi-nfs pods came up - 

 
{code:java}
oc get pods | grep nfs
csi-nfsplugin-m8hkm                                               3/3     Running     0               18s
csi-nfsplugin-phqpb                                               3/3     Running     0               18s
csi-nfsplugin-provisioner-5f94bb6f64-7htn8                        6/6     Running     0               18s
csi-nfsplugin-provisioner-5f94bb6f64-dw68c                        6/6     Running     0               18s
csi-nfsplugin-vgcj4                                               3/3     Running     0               18s
 
{code}
3. Enable csi_nfs
oc patch cm rook-ceph-operator-config -n openshift-storage -p $'data:\n "ROOK_CSI_ENABLE_NFS":  "true"'

4. Check rook-ceph-nfs-ocs-storagecluster-cephnfs pod came up -

 
{code:java}
oc get pods | grep nfs
csi-nfsplugin-m8hkm                                               3/3     Running            0               62s
csi-nfsplugin-phqpb                                               3/3     Running            0               62s
csi-nfsplugin-provisioner-5f94bb6f64-7htn8                        6/6     Running            0               62s
csi-nfsplugin-provisioner-5f94bb6f64-dw68c                        6/6     Running            0               62s
csi-nfsplugin-vgcj4                                               3/3     Running            0               62s
rook-ceph-nfs-ocs-storagecluster-cephnfs-a-585f8b9796-pwcdw       1/2     CrashLoopBackOff   2 (19s ago)     43s
{code}
 

*The exact date and time when the issue was observed, including timezone details:*

 

*Actual results:*

rook-ceph-nfs-ocs-storagecluster-cephnfs pod is in CrashLoopBackOff state

 

*Expected results:*

rook-ceph-nfs-ocs-storagecluster-cephnfs pod should be in running state

Additional Info:

{{From regression runs we see that this issue was not faced for,}}
{{ceph version--18.2.1-278.el9cp (2ae16095654f99a1a043ca3f0c7befcb78080058) reef (stable)}}
{{ }}
{{[regression-test-logs|http://magna002.ceph.redhat.com/ocsci-jenkins/openshift-clusters/j393ai3c33ua/j393ai3c33ua_20250131T072303/logs/test_report_1738325989.html]}}
{{ }}
{{For odf 4.18 also this ceph version is working fine.}}

{{ocp: 4.18.0-0.nightly-2025-02-11-134047}}
{{odf: 4.18.0-124}}
{{ceph image used : registry.redhat.io/rhceph/rhceph-7-rhel9@sha256:cea147a49680256b8f3ec9b1e3a723f5e040cd0d722b1f18086f6959f71ffaf8}}

{{oc get pods -n openshift-storage| grep nfs}}
{{rook-ceph-nfs-ocs-storagecluster-cephnfs-a-57776f94df-9r5q8 2/2 Running 0 17m}}


ODF bug: https://issues.redhat.com/browse/DFBUGS-1592

Comment 12 errata-xmlrpc 2025-06-26 12:25:26 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Important: Red Hat Ceph Storage 8.1 security, bug fix, and enhancement updates), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHSA-2025:9775


Note You need to log in before you can comment on or make changes to this bug.