Bug 2265574

Summary: RGW labeled perf counter metrics sent by ceph-exporter dont difference between user and bucket
Product: [Red Hat Storage] Red Hat Ceph Storage Reporter: Ali Maredia <amaredia>
Component: RGWAssignee: Ali Maredia <amaredia>
Status: CLOSED ERRATA QA Contact: Chaithra <ckulal>
Severity: urgent Docs Contact: Akash Raj <akraj>
Priority: unspecified    
Version: 7.1CC: akraj, ceph-eng-bugs, cephqe-warriors, dparkes, mkasturi, nravinas, saraut, tserlin, vereddy
Target Milestone: ---   
Target Release: 7.1   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: ceph-18.2.1-37.el9cp Doc Type: Enhancement
Doc Text:
.RGW labeled user and bucket operation counters are now in different sections when the `ceph counter dump` is run Previously, all RGW labeled operation counters were in the ‘rgw_op’ section of the output of the `ceph counter dump` command but would either have a user label or a bucket label. With this enhancement, RGW labeled user and bucket operation counters are in `rgw_op_per_user` or `rgw_op_per_bucket` sections respectively when the `ceph counter dump` command is executed.
Story Points: ---
Clone Of: 2251055
: 2279771 (view as bug list) Environment:
Last Closed: 2024-06-13 14:27:25 UTC Type: ---
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:
Bug Depends On: 2251055    
Bug Blocks: 2279771, 2267614, 2298578, 2298579    

Description Ali Maredia 2024-02-22 19:58:16 UTC
+++ This bug was initially created as a clone of Bug #2251055 +++

Current Metrics sent by the ceph-exporter don't include user or bucket on the metric name, for example:

ceph_rgw_op_del_obj_bytes{Bucket="bkt", instance="localhost:9926", instance_id="8000", job="radosgw"}
ceph_rgw_op_del_obj_bytes{User="anonymous", instance="localhost:9926", instance_id="8000", job="radosgw"}

This adds an extra dimension for Prometheus and Grafana to query and filter out the specific data, making it very complex and resource-expensive to configure Grafana dashboards.

The request is to modify the ceph-exporter so it constructs the metrics it sends out to Prometheus using the metric's key. In the future, they could be for anything else, like groups, accounts,etc.

Examples on what the final metric being sent to Prometheus should look like:

-- GLOBAL --
ceph_rgw_op_del_obj_bytes
ceph_rgw_op_del_obj_bytes{instance="localhost:9926", instance_id="8000", job="radosgw"}


-- Bucket operations --
ceph_rgw_op_del_bucket_obj_bytes

ceph_rgw_op_del_bucket_obj_bytes{Bucket="bkt", instance="localhost:9926", instance_id="8000", job="radosgw"}
ceph_rgw_op_del_bucket_obj_bytes{Bucket="bkt2", instance="localhost:9926", instance_id="8000", job="radosgw"}

-- User operations --
ceph_rgw_op_del_user_obj_bytes

ceph_rgw_op_del_user_obj_bytes{User="anonymous", instance="localhost:9926", instance_id="8000", job="radosgw"}
ceph_rgw_op_del_user_obj_bytes{User="test3", instance="localhost:9926", instance_id="8000", job="radosgw"}

--- Additional comment from Ali Maredia on 2024-02-22 19:19:26 UTC ---

The solution for this issue has been merged upstream at https://github.com/ceph/ceph/pull/54623 (only the first commit) and that commit has been pushed downstream.

Comment 9 errata-xmlrpc 2024-06-13 14:27:25 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (Critical: Red Hat Ceph Storage 7.1 security, enhancements, and bug fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHSA-2024:3925