+++ This bug was initially created as a clone of Bug #2269663 +++ This config can result in a single client holding up mds to service other clients since once a client is deferred from eviction due to laggy OSD, a new clients cap acquire request can be possibly blocked until the other laggy client resumes operation, i.e., when the laggy OSD is considered non-laggy anymore.
Hi All, [ceph: root@mero017 /]# ceph config get mds defer_client_eviction_on_laggy_osds true [ceph: root@mero017 /]# ceph versions { "mon": { "ceph version 18.2.1-76.el9cp (2517f8a5ef5f5a6a22013b2fb11a591afd474668) reef (stable)": 3 }, "mgr": { "ceph version 18.2.1-76.el9cp (2517f8a5ef5f5a6a22013b2fb11a591afd474668) reef (stable)": 3 }, "osd": { "ceph version 18.2.1-76.el9cp (2517f8a5ef5f5a6a22013b2fb11a591afd474668) reef (stable)": 33 }, "mds": { "ceph version 18.2.1-76.el9cp (2517f8a5ef5f5a6a22013b2fb11a591afd474668) reef (stable)": 6 }, "overall": { "ceph version 18.2.1-76.el9cp (2517f8a5ef5f5a6a22013b2fb11a591afd474668) reef (stable)": 45 } } [ceph: root@mero017 /]# We are validating the clients eviction also. If clients are not evicted and if we still can access the mount points on the client then we are failing the test case Ref: https://github.com/red-hat-storage/cephci/blob/e01ff9a132697422bf8d320385aceed5140db553/tests/cephfs/cephfs_bugs/test_defer_client_evict_on_laggy_osd.py#L172 Log : http://magna002.ceph.redhat.com/cephci-jenkins/test-runs/18.2.1-77/Regression/cephfs/84/tier-2_cephfs_test-clients/Client_eviction_deferred_if_OSD_is_laggy_0.log Regards, Amarnath
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Critical: Red Hat Ceph Storage 7.1 security, enhancements, and bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2024:3925