Description of problem: The container_memory_failures_total metric is in the top 10 of metrics with high cardinality. It isn't used in any rule or dashboard. Storing the metric in Prometheus increases memory usage for no benefit. Version-Release number of selected component (if applicable): 4.6 How reproducible: Always Steps to Reproduce: 1. Open the Prometheus UI, go to the Status > TSDB status page and look at the "Top 10 series count by metric names" section. Actual results: container_memory_failures_total is listed. Expected results: container_memory_failures_total isn't present. Additional info: N/A
checked with 4.8.0-0.nightly-2021-05-05-030749, no container_memory_failures_total metrics now # token=`oc sa get-token prometheus-k8s -n openshift-monitoring` # oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k -H "Authorization: Bearer $token" 'https://prometheus-k8s.openshift-monitoring.svc:9091/api/v1/label/__name__/values' | jq | grep container_memory_failures_total no result
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: OpenShift Container Platform 4.8.2 bug fix and security update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2021:2438