RHEL Engineering is moving the tracking of its product development work on RHEL 6 through RHEL 9 to Red Hat Jira (issues.redhat.com). If you're a Red Hat customer, please continue to file support cases via the Red Hat customer portal. If you're not, please head to the "RHEL project" in Red Hat Jira and file new tickets here. Individual Bugzilla bugs in the statuses "NEW", "ASSIGNED", and "POST" are being migrated throughout September 2023. Bugs of Red Hat partners with an assigned Engineering Partner Manager (EPM) are migrated in late September as per pre-agreed dates. Bugs against components "kernel", "kernel-rt", and "kpatch" are only migrated if still in "NEW" or "ASSIGNED". If you cannot log in to RH Jira, please consult article #7032570. That failing, please send an e-mail to the RH Jira admins at rh-issues@redhat.com to troubleshoot your issue as a user management inquiry. The email creates a ServiceNow ticket with Red Hat. Individual Bugzilla bugs that are migrated will be moved to status "CLOSED", resolution "MIGRATED", and set with "MigratedToJIRA" in "Keywords". The link to the successor Jira issue will be found under "Links", have a little "two-footprint" icon next to it, and direct you to the "RHEL project" in Red Hat Jira (issue links are of type "https://issues.redhat.com/browse/RHEL-XXXX", where "X" is a digit). This same link will be available in a blue banner at the top of the page informing you that that bug has been migrated.
Bug 1985523 - Memory metrics moved in /proc/zoneinfo in RHEL 8
Summary: Memory metrics moved in /proc/zoneinfo in RHEL 8
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Enterprise Linux 8
Classification: Red Hat
Component: pcp
Version: 8.6
Hardware: All
OS: Linux
high
high
Target Milestone: beta
: 8.6
Assignee: Nathan Scott
QA Contact: Jan Kurik
Apurva Bhide
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2021-07-23 19:39 UTC by Charles Haithcock
Modified: 2022-05-10 13:49 UTC (History)
3 users (show)

Fixed In Version: pcp-5.3.5-1.el8
Doc Type: No Doc Update
Doc Text:
Clone Of:
Environment:
Last Closed: 2022-05-10 13:30:52 UTC
Type: Bug
Target Upstream Version:
Embargoed:
pm-rhel: mirror+


Attachments (Terms of Use)


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHBA-2022:1765 0 None None None 2022-05-10 13:31:02 UTC

Description Charles Haithcock 2021-07-23 19:39:00 UTC
Description of problem:

/proc/zoneinfo changed in RHEL 8 to include aggregate per-node metrics. Some metrics are still per-zone as well as per-node now, but some are no longer per-zone. Slab metrics, for example, are per-node now and the kernel does not actually manage slab at the zone level. As such, mem.zoneinfo metrics are producing strange results. 


Version-Release number of selected component (if applicable):

pcp-5.2.5-4.el8.x86_64 
pcp-libs-5.2.5-4.el8.x86_64

How reproducible:

100%


Steps to Reproduce:
1. install pcp-zeroconf
2.
3.

Actual results:

Some metrics reported are larger than the upper limit of memory in the zone

Expected results:

accurate info

Additional info:


From my RHEL 8 vm:

 r8 # pminfo -ft | grep -B 1 'DMA::node0'
mem.zoneinfo.free [free space in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 15096
--
mem.zoneinfo.min [min space in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 376
--
mem.zoneinfo.low [low space in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 468
--
mem.zoneinfo.high [high space in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 560
--
mem.zoneinfo.scanned [scanned space in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 0
--
mem.zoneinfo.spanned [spanned space in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 16380
--
mem.zoneinfo.present [present space in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 15992
--
mem.zoneinfo.managed [managed space in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 15360
--
mem.zoneinfo.nr_free_pages [number of free pages in each zone for each NUMA node.]
    inst [0 or "DMA::node0"] value 15096
--
mem.zoneinfo.nr_alloc_batch [number of pages allocated to other zones due to insufficient memory]
    inst [0 or "DMA::node0"] value 0
--
mem.zoneinfo.nr_inactive_anon [number of inactive anonymous memory pages in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 382140          <---
--
mem.zoneinfo.nr_active_anon [number of active anonymous memory pages in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 1072
--
mem.zoneinfo.nr_inactive_file [number of inactive file memory pages in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 264124           <--- 
--
mem.zoneinfo.nr_active_file [number of active file memory memory pages in each zone for each NUMA node]
    inst [0 or "DMA::node0"] value 108376           <---
--
[...]

Comment 1 Nathan Scott 2021-10-19 06:38:59 UTC
Hi Charles,

Any chance you can help me out with some test data from a multi-node NUMA system with a newer kernel like this?  I need something like this to regression test the changes I'll be making here ...

$ tar tzf qa/linux/meminfo-root-003.tgz 
proc/
proc/cpuinfo
proc/slabinfo
proc/stat
proc/vmstat
proc/zoneinfo
proc/meminfo
sys/
sys/devices/
sys/devices/system/
sys/devices/system/node/
sys/devices/system/node/node0/
sys/devices/system/node/node0/numastat
sys/devices/system/node/node0/meminfo
sys/devices/system/node/node1/
sys/devices/system/node/node1/numastat
sys/devices/system/node/node1/meminfo

thanks!

Comment 2 Nathan Scott 2021-10-20 02:14:35 UTC
Not to worry - I've been able to create one locally with virsh.

Comment 3 Nathan Scott 2021-10-20 06:56:32 UTC
Resolved upstream now, will be included from pcp-5.3.5 onward.

commit 2ebee9724ed895480080fc7efcc407cdbe261ffa
Author: Nathan Scott <nathans>
Date:   Wed Oct 20 17:55:16 2021 +1100

    pmdalinux: additions and updates to mem.zoneinfo metrics
    
    New metrics below mem.zoneinfo to match those available from
    latest Linux kernel versions.  Updates to existing metrics too.
    
    Support for the per-node change to /proc/zoneinfo is added as
    well, where some memory metrics that were previously exported
    both per-node-per-zone are now only exported per-node.  This
    is handled in a backward compatible way in terms of metrics
    we export - old metrics give 'no values available' and newer
    ones appear as both new metrics and new instances.
    
    New test data from modern kernels is added and test qa/821 is
    updated for expected output on various kernel versions.
    
    Resolves Red Hat BZ #1985523.

Comment 11 errata-xmlrpc 2022-05-10 13:30:52 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (pcp bug fix and enhancement update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

https://access.redhat.com/errata/RHBA-2022:1765


Note You need to log in before you can comment on or make changes to this bug.