As monitoring user of RBD mirror daemon I need to have perf counters where context attributes will not be part of the performance counter name. This will allow to transorm these performance counter to prometheus metrics using the usual format. Example: rbd_mirror_snapshot_image_<image_spec>snapshots* to: ceph_rbd_mirror_snapshot_image_snapshot_*(pool: px, image: imgx)
Please specify the severity of this bug. Severity is defined here: https://bugzilla.redhat.com/page.cgi?id=fields.html#bug_severity.
can you try and run the admin socket command in the directory, instead of providing the path? looks like your path is longer than the system defined limit. also make sure that you're running the command on the secondary cluster to see the values.
i think the socket file path length is an os limit, i don't think we need to file a BZ for that
> Is this overall pool (Total number of pools) mirror usage for snapshot based? These values are afaik agnostic of pools, as this is the global data of all the snapshots that daemon has handled. `perf dump` will not show any labels > If we get these metrics using pool wise, wouldn't it be more helpful? Right now there's only 2 levels in which snapshot based mirroring reports perf counters, one is on the global daemon level, another is per image based(with the labels). If we want pool based data, we can use the per image counter to compile it. > why these metrics are not available from primary? These are not available on the primary, as the snapshots on the primary get handled by the rbd-mirror daemon on the secondary cluster. Thus we are only able to generate counters on the secondary side. This data also gets written on mirror image status, so we can observe this data on the primary cluster using that command. > Labels field is empty here, what value we should expect here and which scenario? The counter dump command you've posted is for rbd_mirror_snapshot, which is counters for all the snapshots being processed by that daemon (global). The labels are on the per image based counter `rbd_mirror_snapshot_image`. And will look something like this, ``` "rbd_mirror_snapshot_image": { "labels": { "image": "image1", "namespace": "", "pool": "data" }, "counters": { "snapshots": 106, "sync_time": { "avgcount": 106, "sum": 8.157710400, "avgtime": 0.076959532 }, "sync_bytes": 524288000, "remote_timestamp": 1682500200.777647685, "local_timestamp": 1682500200.777647685, "last_sync_time": 0.003713542, "last_sync_bytes": 0 } }``` can you check and see if you have that counter working for you? you can confirm the values using mirror image status command to verify if the values are correct for the images. > both are having same values, looks like an issue for me. I'm not sure I see the same values in the output you've posted, can you point out where the values are same, I might've missed it.
*** Bug 2145593 has been marked as a duplicate of this bug. ***
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Red Hat Ceph Storage 6.1 security and bug fix update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2023:3623
The needinfo request[s] on this closed bug have been removed as they have been unresolved for 120 days