Bug 2251055 - RGW labeled perf counter metrics sent by ceph-exporter dont difference between user and bucket
Summary: RGW labeled perf counter metrics sent by ceph-exporter dont difference betwee...
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Ceph Storage
Classification: Red Hat Storage
Component: Ceph-Metrics
Version: 7.0
Hardware: Unspecified
OS: Unspecified
high
urgent
Target Milestone: ---
: 7.0z2
Assignee: Ali Maredia
QA Contact: Chaithra
Disha Walvekar
URL:
Whiteboard:
Depends On:
Blocks: 2279771 2265574 2270485
TreeView+ depends on / blocked
 
Reported: 2023-11-22 15:45 UTC by daniel parkes
Modified: 2024-05-09 07:02 UTC (History)
9 users (show)

Fixed In Version: ceph-18.2.0-158.el9cp
Doc Type: Release Note
Doc Text:
Labeled perf counters for RGW Op metrics are now split into different sections in the output of `counter dump` for the user operation counters and bucket operation counters.
Clone Of:
: 2265574 (view as bug list)
Environment:
Last Closed: 2024-05-07 12:10:24 UTC
Embargoed:


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Issue Tracker RHCEPH-7948 0 None None None 2023-11-22 15:48:03 UTC
Red Hat Product Errata RHBA-2024:2743 0 None None None 2024-05-07 12:10:35 UTC

Description daniel parkes 2023-11-22 15:45:02 UTC
Current Metrics sent by the ceph-exporter don't include user or bucket on the metric name, for example:

ceph_rgw_op_del_obj_bytes{Bucket="bkt", instance="localhost:9926", instance_id="8000", job="radosgw"}
ceph_rgw_op_del_obj_bytes{User="anonymous", instance="localhost:9926", instance_id="8000", job="radosgw"}

This adds an extra dimension for Prometheus and Grafana to query and filter out the specific data, making it very complex and resource-expensive to configure Grafana dashboards.

The request is to modify the ceph-exporter so it constructs the metrics it sends out to Prometheus using the metric's key. In the future, they could be for anything else, like groups, accounts,etc.

Examples on what the final metric being sent to Prometheus should look like:

-- GLOBAL --
ceph_rgw_op_del_obj_bytes
ceph_rgw_op_del_obj_bytes{instance="localhost:9926", instance_id="8000", job="radosgw"}


-- Bucket operations --
ceph_rgw_op_del_bucket_obj_bytes

ceph_rgw_op_del_bucket_obj_bytes{Bucket="bkt", instance="localhost:9926", instance_id="8000", job="radosgw"}
ceph_rgw_op_del_bucket_obj_bytes{Bucket="bkt2", instance="localhost:9926", instance_id="8000", job="radosgw"}

-- User operations --
ceph_rgw_op_del_user_obj_bytes

ceph_rgw_op_del_user_obj_bytes{User="anonymous", instance="localhost:9926", instance_id="8000", job="radosgw"}
ceph_rgw_op_del_user_obj_bytes{User="test3", instance="localhost:9926", instance_id="8000", job="radosgw"}

Comment 10 errata-xmlrpc 2024-05-07 12:10:24 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Red Hat Ceph Storage 7.0 Bug Fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2024:2743


Note You need to log in before you can comment on or make changes to this bug.