Bug 2006625
Summary: | Engine generates VDS_HIGH_MEM_USE events for empty hosts that have most memory reserved by huge pages | ||
---|---|---|---|
Product: | Red Hat Enterprise Virtualization Manager | Reporter: | Germano Veit Michel <gveitmic> |
Component: | ovirt-engine | Assignee: | Lucia Jelinkova <ljelinko> |
Status: | CLOSED ERRATA | QA Contact: | Qin Yuan <qiyuan> |
Severity: | medium | Docs Contact: | |
Priority: | high | ||
Version: | 4.4.8 | CC: | ahadas, apinnick, dfodor, emarcus, ljelinko, mavital, mburman, qiyuan, tgolembi |
Target Milestone: | ovirt-4.5.1 | Keywords: | ZStream |
Target Release: | --- | ||
Hardware: | x86_64 | ||
OS: | Linux | ||
Whiteboard: | |||
Fixed In Version: | ovirt-engine-4.5.1.2 | Doc Type: | Bug Fix |
Doc Text: |
Previously, memory allocated by hugepages was included in the host memory usage calculation, resulting in high memory usage in the Administration Portal, even with no running VMs, and false VDS_HIGH_MEM_USE warnings in the logs.
In this release, hugepages are not included in the memory usage. VDS_HIGH_MEM_USE warnings are logged only when normal (not hugepages) memory usage is above a defined threshold. Memory usage in the Administration Portal is calculated from the normal and hugepages used memory, not from allocated memory.
|
Story Points: | --- |
Clone Of: | Environment: | ||
Last Closed: | 2022-07-14 12:54:31 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | Virt | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Germano Veit Michel
2021-09-22 04:23:32 UTC
Due to memUsed = 71, the Admin Portal also shows the host with the graph bar at 71% and yellow. It is not really true... Maybe somehow take into account the free huge pages? Verified with: ovirt-engine-4.5.1.2-0.11.el8ev.noarch Steps and results: 1. Set the cluster memory threshold to 50% 2. On a host(not running any VM) with 62G total memory, reserve 40G (64.5%) with HugePages # egrep '^HugePages_|^Mem' /proc/meminfo MemTotal: 65366332 kB MemFree: 20537144 kB MemAvailable: 21274804 kB HugePages_Total: 40 HugePages_Free: 40 HugePages_Rsvd: 0 HugePages_Surp: 0 3. Check engine logs to see if there is no VDS_HIGH_MEM_USE warning: There is no VDS_HIGH_MEM_USE in engine.log 4. Create a VM with 16G memory, no hugepages, run the VM on the host, load memory: # free -m total used free shared buff/cache available Mem: 15798 15200 392 8 205 324 Swap: 0 0 0 5. Check engine logs to see if there is a VDS_HIGH_MEM_USE warning: There is a VDS_HIGH_MEM_USE warning saying: 2022-06-27 17:59:01,492+03 WARN [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-15) [] EVENT_ID: VDS_HIGH_MEM_USE(532), Used memory of host host_mixed_1 in cluster golden_env_mixed_1 [28%] exceeded defined threshold [50%]. 6. Check memory usage on UI to see if it's a total usage of normal memory and hugepages memory: The memory usage number is 28% 7. Create another VM with 40G memory, hugepages=1048576, run the VM also on the host, check memory usage on UI to see if it's a total usage of normal memory and hugepages memory: The memory usage number is 92% According to the test results, the VDS_HIGH_MEM_USE warning and the memory usage on UI work as expected, except the usage number in the warning should be the usage of normal memory, but not the total usage of normal memory and hugepages memory. Filed a bug tracking for the wrong usage number issue, see https://bugzilla.redhat.com/show_bug.cgi?id=2101503. Move this bug to VERIFIED. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: RHV Manager (ovirt-engine) [ovirt-4.5.1] security, bug fix and update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2022:5555 Due to QE capacity, we are not going to cover this issue in our automation |