Note: This bug is displayed in read-only format because
the product is no longer active in Red Hat Bugzilla.
RHEL Engineering is moving the tracking of its product development work on RHEL 6 through RHEL 9 to Red Hat Jira (issues.redhat.com). If you're a Red Hat customer, please continue to file support cases via the Red Hat customer portal. If you're not, please head to the "RHEL project" in Red Hat Jira and file new tickets here. Individual Bugzilla bugs in the statuses "NEW", "ASSIGNED", and "POST" are being migrated throughout September 2023. Bugs of Red Hat partners with an assigned Engineering Partner Manager (EPM) are migrated in late September as per pre-agreed dates. Bugs against components "kernel", "kernel-rt", and "kpatch" are only migrated if still in "NEW" or "ASSIGNED". If you cannot log in to RH Jira, please consult article #7032570. That failing, please send an e-mail to the RH Jira admins at rh-issues@redhat.com to troubleshoot your issue as a user management inquiry. The email creates a ServiceNow ticket with Red Hat. Individual Bugzilla bugs that are migrated will be moved to status "CLOSED", resolution "MIGRATED", and set with "MigratedToJIRA" in "Keywords". The link to the successor Jira issue will be found under "Links", have a little "two-footprint" icon next to it, and direct you to the "RHEL project" in Red Hat Jira (issue links are of type "https://issues.redhat.com/browse/RHEL-XXXX", where "X" is a digit). This same link will be available in a blue banner at the top of the page informing you that that bug has been migrated.
Description of problem:
If the total size of hugepages allocated on a system is over half of the total memory size, commitlimit becomes a negative number. That calculation is used in __vm_enough_memory() in mm/mmap.c. This currently broken in the community.
Version-Release number of selected component (if applicable):
RHEL6.1 (and community)
How reproducible:
100%
Steps to Reproduce:
1. Allocate enough hugepages to consume over half of available memory
For example "default_hugepagesz=1G hugepagesz=1G hugepages=64" on the
linux kernel bootline.
2. After the system is booted, "cat /proc/meminfo". Look at the "CommitLimit:"
entry. If is is a huge number (ie "CommitLimit: 737869762947802600 kB"
it hit the problem.
Actual results:
CommitLimit: 737869762947802600 kB
Expected results:
A CommitLimit number less than the size of memory.
Additional info:
This is also broken in the community. Reported it http://marc.info/?l=linux-kernel&m=130515303205772&w=2
If the total size of hugepages allocated on a system is over half of the total memory size, commitlimit becomes a negative number.
What happens in fs/proc/meminfo.c is this calculation:
allowed = ((totalram_pages - hugetlb_total_pages())
* sysctl_overcommit_ratio / 100) + total_swap_pages;
The problem is that hugetlb_total_pages() is larger than totalram_pages resulting in a negative number. Since allowed is an unsigned long the
negative shows up as a big number.
A similar calculation occurs in __vm_enough_memory() in mm/mmap.c.
A symptom of this problem is that /proc/meminfo prints a very large
CommitLimit number.
CommitLimit: 737869762947802600 kB
To reproduce the problem reserve over half of memory as hugepages. For
example "default_hugepagesz=1G hugepagesz=1G hugepages=64". Then look at /proc/meminfo .
uv1-sys:~ # cat /proc/meminfo
MemTotal: 32395508 kB
MemFree: 32029276 kB
Buffers: 8656 kB
Cached: 89548 kB
SwapCached: 0 kB
Active: 55336 kB
Inactive: 73916 kB
Active(anon): 31220 kB
Inactive(anon): 36 kB
Active(file): 24116 kB
Inactive(file): 73880 kB
Unevictable: 0 kB
Mlocked: 0 kB
SwapTotal: 0 kB
SwapFree: 0 kB
Dirty: 1692 kB
Writeback: 0 kB
AnonPages: 31132 kB
Mapped: 15668 kB
Shmem: 152 kB
Slab: 70256 kB
SReclaimable: 17148 kB
SUnreclaim: 53108 kB
KernelStack: 6536 kB
PageTables: 3704 kB
NFS_Unstable: 0 kB
Bounce: 0 kB
WritebackTmp: 0 kB
CommitLimit: 737869762947802600 kB
Committed_AS: 394044 kB
VmallocTotal: 34359738367 kB
VmallocUsed: 713960 kB
VmallocChunk: 34325764204 kB
HardwareCorrupted: 0 kB
HugePages_Total: 32
HugePages_Free: 32
HugePages_Rsvd: 0
HugePages_Surp: 0
Hugepagesize: 1048576 kB
DirectMap4k: 16384 kB
DirectMap2M: 2064384 kB
DirectMap1G: 65011712 kB
Turns out that when hugepages are allocated totalram_pages gets decremented
so there is no need to subtract hugetlb_total_pages().
Sent a fix to the community.
http://marc.info/?l=linux-mm&m=130573288915598&w=2
I thought the first solution was not ok, the second looks better.
Comment 6RHEL Program Management
2011-06-07 18:19:54 UTC
This request was evaluated by Red Hat Product Management for inclusion
in a Red Hat Enterprise Linux maintenance release. Product Management has
requested further review of this request by Red Hat Engineering, for potential
inclusion in a Red Hat Enterprise Linux Update release for currently deployed
products. This request is not yet committed for inclusion in an Update release.
Howdy Peter,
Seems that your attempt to setup gigantic hugepages is being 'rejected', as your hugepage size still reported as the ordinary 2MB:
> Hugepagesize: 2048 kB
Check if the processor you're using for your tests supports 1GB pages. If the CPU supports 1GB pages it has the PDPE1GB flag.
Cheers!
--aquini
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.
For information on the advisory, and where to find the updated
files, follow the link below.
If the solution does not work for you, open a new bug report.
http://rhn.redhat.com/errata/RHSA-2011-1530.html