Bug 2251055

Summary: RGW labeled perf counter metrics sent by ceph-exporter dont difference between user and bucket
Product: [Red Hat Storage] Red Hat Ceph Storage Reporter: daniel parkes <dparkes>
Component: Ceph-MetricsAssignee: Ali Maredia <amaredia>
Status: CLOSED ERRATA QA Contact: Chaithra <ckulal>
Severity: urgent Docs Contact: Disha Walvekar <dwalveka>
Priority: high    
Version: 7.0CC: amaredia, ceph-eng-bugs, cephqe-warriors, ckulal, dparkes, dwalveka, kdreyer, nravinas, tserlin
Target Milestone: ---   
Target Release: 7.0z2   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: ceph-18.2.0-158.el9cp Doc Type: Release Note
Doc Text:
Labeled perf counters for RGW Op metrics are now split into different sections in the output of `counter dump` for the user operation counters and bucket operation counters.
Story Points: ---
Clone Of:
: 2265574 (view as bug list) Environment:
Last Closed: 2024-05-07 12:10:24 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On:    
Bug Blocks: 2279771, 2265574, 2270485    

Description daniel parkes 2023-11-22 15:45:02 UTC
Current Metrics sent by the ceph-exporter don't include user or bucket on the metric name, for example:

ceph_rgw_op_del_obj_bytes{Bucket="bkt", instance="localhost:9926", instance_id="8000", job="radosgw"}
ceph_rgw_op_del_obj_bytes{User="anonymous", instance="localhost:9926", instance_id="8000", job="radosgw"}

This adds an extra dimension for Prometheus and Grafana to query and filter out the specific data, making it very complex and resource-expensive to configure Grafana dashboards.

The request is to modify the ceph-exporter so it constructs the metrics it sends out to Prometheus using the metric's key. In the future, they could be for anything else, like groups, accounts,etc.

Examples on what the final metric being sent to Prometheus should look like:

-- GLOBAL --
ceph_rgw_op_del_obj_bytes
ceph_rgw_op_del_obj_bytes{instance="localhost:9926", instance_id="8000", job="radosgw"}


-- Bucket operations --
ceph_rgw_op_del_bucket_obj_bytes

ceph_rgw_op_del_bucket_obj_bytes{Bucket="bkt", instance="localhost:9926", instance_id="8000", job="radosgw"}
ceph_rgw_op_del_bucket_obj_bytes{Bucket="bkt2", instance="localhost:9926", instance_id="8000", job="radosgw"}

-- User operations --
ceph_rgw_op_del_user_obj_bytes

ceph_rgw_op_del_user_obj_bytes{User="anonymous", instance="localhost:9926", instance_id="8000", job="radosgw"}
ceph_rgw_op_del_user_obj_bytes{User="test3", instance="localhost:9926", instance_id="8000", job="radosgw"}

Comment 10 errata-xmlrpc 2024-05-07 12:10:24 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Red Hat Ceph Storage 7.0 Bug Fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2024:2743