Description of problem: add a command to dump directory information Version-Release number of selected component (if applicable): 6.0 How reproducible: $ ceph daemon mds.a dump dir /test-dir Expected results: [ { "value/bits": "0/0", "status": "dirfrag not in cache" } ] Additional info:
This is not required for 6.1, changing the target to 7.0z2.
Hi All, Setup details : [root@ceph-amk-weekly-wzejtc-node8 ~]# ceph -s cluster: id: cab92b0e-f624-11ee-9418-fa163ecee7e5 health: HEALTH_OK services: mon: 3 daemons, quorum ceph-amk-weekly-wzejtc-node1-installer,ceph-amk-weekly-wzejtc-node3,ceph-amk-weekly-wzejtc-node2 (age 46m) mgr: ceph-amk-weekly-wzejtc-node1-installer.mxsbyy(active, since 48m), standbys: ceph-amk-weekly-wzejtc-node2.kknief mds: 2/2 daemons up, 3 standby osd: 12 osds: 12 up (since 41m), 12 in (since 42m) data: volumes: 1/1 healthy pools: 3 pools, 49 pgs objects: 103 objects, 8.3 MiB usage: 725 MiB used, 179 GiB / 180 GiB avail pgs: 49 active+clean [root@ceph-amk-weekly-wzejtc-node8 ~]# ceph versions { "mon": { "ceph version 18.2.0-188.el9cp (639245962af16e2c342418f2318f86f7d1e34e24) reef (stable)": 3 }, "mgr": { "ceph version 18.2.0-188.el9cp (639245962af16e2c342418f2318f86f7d1e34e24) reef (stable)": 2 }, "osd": { "ceph version 18.2.0-188.el9cp (639245962af16e2c342418f2318f86f7d1e34e24) reef (stable)": 12 }, "mds": { "ceph version 18.2.0-188.el9cp (639245962af16e2c342418f2318f86f7d1e34e24) reef (stable)": 5 }, "overall": { "ceph version 18.2.0-188.el9cp (639245962af16e2c342418f2318f86f7d1e34e24) reef (stable)": 22 } } [root@ceph-amk-weekly-wzejtc-node8 ~]# ceph fs status cephfs - 2 clients ====== RANK STATE MDS ACTIVITY DNS INOS DIRS CAPS 0 active cephfs.ceph-amk-weekly-wzejtc-node3.zumhmb Reqs: 0 /s 86 21 20 2 1 active cephfs.ceph-amk-weekly-wzejtc-node4.rcdikx Reqs: 0 /s 27 20 18 0 POOL TYPE USED AVAIL cephfs.cephfs.meta metadata 24.0M 56.6G cephfs.cephfs.data data 0 56.6G STANDBY MDS cephfs.ceph-amk-weekly-wzejtc-node5.ckfdfo cephfs.ceph-amk-weekly-wzejtc-node7.qnmjub cephfs.ceph-amk-weekly-wzejtc-node6.zpjhbt MDS version: ceph version 18.2.0-188.el9cp (639245962af16e2c342418f2318f86f7d1e34e24) reef (stable) Command output: [root@ceph-amk-weekly-wzejtc-node8 ~]# ceph orch host ls HOST ADDR LABELS STATUS ceph-amk-weekly-wzejtc-node1-installer 10.0.210.48 _admin,mgr,mon,installer ceph-amk-weekly-wzejtc-node2 10.0.208.35 mgr,mon ceph-amk-weekly-wzejtc-node3 10.0.208.162 mon,mds ceph-amk-weekly-wzejtc-node4 10.0.208.179 mds,osd ceph-amk-weekly-wzejtc-node5 10.0.208.161 mds,osd ceph-amk-weekly-wzejtc-node6 10.0.210.4 nfs,mds,osd ceph-amk-weekly-wzejtc-node7 10.0.209.11 nfs,mds [root@ceph-amk-weekly-wzejtc-node8 ~]# ceph-fuse /mnt/fuse/ 2024-04-09T00:51:35.795-0400 7f1822438480 -1 init, newargv = 0x7f1810004df0 newargc=15 ceph-fuse[9898]: starting ceph client ceph-fuse[9898]: starting fuse [root@ceph-amk-weekly-wzejtc-node8 ~]# mkdir /mnt/fuse/test_dir [root@ceph-amk-weekly-wzejtc-node8 ~]# mkdir /mnt/fuse/test_dir/test [root@ceph-amk-weekly-wzejtc-node8 ~]# [ceph: root@ceph-amk-weekly-wzejtc-node3 /]# ceph daemon mds.cephfs.ceph-amk-weekly-wzejtc-node3.zumhmb dump dir / [ { "path": "", "dirfrag": "0x1", "snapid_first": 2, "projected_version": "1920", "version": "1920", "committing_version": "0", "committed_version": "0", "is_rep": false, "dir_auth": "0", "states": [ "auth", "dirty", "complete" ], "is_auth": true, "auth_state": { "replicas": {} }, "replica_state": { "authority": [ 0, -2 ], "replica_nonce": 0 }, "auth_pins": 0, "is_frozen": false, "is_freezing": false, "pins": { "child": 1, "subtree": 1, "subtreetemp": 0, "replicated": 0, "dirty": 1, "waiter": 0, "authpin": 0 }, "nref": 3 } ] [ceph: root@ceph-amk-weekly-wzejtc-node3 /]# ceph daemon mds.cephfs.ceph-amk-weekly-wzejtc-node3.zumhmb dump dir /test_dir [ { "path": "/test_dir", "dirfrag": "0x100000007e0", "snapid_first": 2, "projected_version": "1", "version": "1", "committing_version": "0", "committed_version": "0", "is_rep": false, "dir_auth": "", "states": [ "auth", "dirty", "complete" ], "is_auth": true, "auth_state": { "replicas": {} }, "replica_state": { "authority": [ 0, -2 ], "replica_nonce": 0 }, "auth_pins": 0, "is_frozen": false, "is_freezing": false, "pins": { "dirty": 1 }, "nref": 1 } ] [ceph: root@ceph-amk-weekly-wzejtc-node3 /]# ceph daemon mds.cephfs.ceph-amk-weekly-wzejtc-node3.zumhmb dump dir /test_dir_1 [ceph: root@ceph-amk-weekly-wzejtc-node3 /]# ceph daemon mds.cephfs.ceph-amk-weekly-wzejtc-node3.zumhmb dump dir /test_dir [ { "path": "/test_dir", "dirfrag": "0x100000007e0", "snapid_first": 2, "projected_version": "3", "version": "3", "committing_version": "0", "committed_version": "0", "is_rep": false, "dir_auth": "", "states": [ "auth", "dirty", "complete" ], "is_auth": true, "auth_state": { "replicas": {} }, "replica_state": { "authority": [ 0, -2 ], "replica_nonce": 0 }, "auth_pins": 0, "is_frozen": false, "is_freezing": false, "pins": { "child": 1, "dirty": 1, "authpin": 0 }, "nref": 2 } ] [ceph: root@ceph-amk-weekly-wzejtc-node3 /]# ceph daemon mds.cephfs.ceph-amk-weekly-wzejtc-node3.zumhmb dump dir /test_dir/test [ { "path": "/test_dir/test", "dirfrag": "0x100000007e1", "snapid_first": 2, "projected_version": "1", "version": "1", "committing_version": "0", "committed_version": "0", "is_rep": false, "dir_auth": "", "states": [ "auth", "dirty", "complete" ], "is_auth": true, "auth_state": { "replicas": {} }, "replica_state": { "authority": [ 0, -2 ], "replica_nonce": 0 }, "auth_pins": 0, "is_frozen": false, "is_freezing": false, "pins": { "dirty": 1 }, "nref": 1 } ] [ceph: root@ceph-amk-weekly-wzejtc-node3 /]# Do we need to verify anything apart from this ? Regards, Amarnath
The outputs are correct.
Hi Jos, Could you please add doc text for the BZ Regards, Amarnath
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Red Hat Ceph Storage 7.0 Bug Fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2024:2743