Bug 1934304
Summary: | MemoryPressure Top Pod Consumers seems to be 2x expected value | ||||||||
---|---|---|---|---|---|---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | jhusta <jhusta> | ||||||
Component: | Management Console | Assignee: | Bipul Adhikari <badhikar> | ||||||
Status: | CLOSED ERRATA | QA Contact: | Yanping Zhang <yanpzhan> | ||||||
Severity: | medium | Docs Contact: | |||||||
Priority: | low | ||||||||
Version: | 4.7 | CC: | aos-bugs, badhikar, jhadvig, kdoberst, krmoser, nmukherj, spadgett, yanpzhan, yapei | ||||||
Target Milestone: | --- | ||||||||
Target Release: | 4.11.0 | ||||||||
Hardware: | s390x | ||||||||
OS: | Linux | ||||||||
Whiteboard: | Scrubbed | ||||||||
Fixed In Version: | Doc Type: | No Doc Update | |||||||
Doc Text: | Story Points: | --- | |||||||
Clone Of: | Environment: | ||||||||
Last Closed: | 2022-08-10 10:36:17 UTC | Type: | Bug | ||||||
Regression: | --- | Mount Type: | --- | ||||||
Documentation: | --- | CRM: | |||||||
Verified Versions: | Category: | --- | |||||||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||||
Cloudforms Team: | --- | Target Upstream Version: | |||||||
Embargoed: | |||||||||
Bug Depends On: | |||||||||
Bug Blocks: | 2055290 | ||||||||
Attachments: |
|
Description
jhusta
2021-03-02 21:55:48 UTC
Created attachment 1760291 [details]
Screen Shots from console
Opened per Samuel Padgett request from related MemoryPressure Defect. Hi jhusta, The bug is reported against hardware s390x, I'm doubt whether it is hardware related, by the way, could you share me the image for pod memstress shown in your screenshot? Hi @yanpzhan my repos and image are in ibm git and artifactory which you will not have access to. We are simply using an ubuntu container and using stress-ng. Here is the command "stress-ng", "-v", "--vm", "1", "--vm-bytes", "'$ALLOCATION'", "--vm-method", "all", "--verify", "--temp-path", "/tmp"]' . With bytes equal to some value. I chose s390x as that is what I am testing on. I don't have access to an x86 machine so I make no assumptions. Here is my dockerfile FROM docker.io/ubuntu RUN apt-get update -y && apt-get install -y stress-ng iperf3 USER 0 CMD stress-ng --mmap 1 Thanks Thanks jhusta, I built image successfully with the dockerfile. Checked on ocp 4.8 cluster with payload 4.8.0-0.nightly-2021-06-02-025513, the bug is still reproduced. The fix pr9030 is not contained in the payload. Waiting for new build with the fix. The fix is till not contained in payload 4.8.0-0.nightly-2021-06-06-164529 @yanpzhan Thanks for keep me posted! Created attachment 1789785 [details]
mem-pod-list
In the test, I created the deployment with pod to consume 8G memory, so that memory are used up. Tested on ocp 4.11 cluster with payload 4.11.0-0.nightly-2022-02-16-211105. 1. $ oc label machineconfigpool worker custom-kubelet=small-pods 2. Create kubeletconfig: apiVersion: machineconfiguration.openshift.io/v1 kind: KubeletConfig metadata: name: set-allocatable spec: machineConfigPoolSelector: matchLabels: custom-kubelet: small-pods kubeletConfig: systemReserved: cpu: 1000m memory: 3Gi 3. Create deployment with pods consume large memory. apiVersion: apps/v1 kind: Deployment metadata: name: memtest namespace: prozyp1 spec: selector: matchLabels: app: httpd replicas: 3 template: metadata: labels: app: httpd spec: containers: - name: httpd image: quay.io/yanpzhan/memstress:latest command: ["stress-ng", "-v", "--vm", "1", "--vm-bytes", "8G", "--vm-method", "all", "--verify", "--temp-path", "/tmp"] ports: - containerPort: 8080 4. Then check on nodes list page, when node show memory pressure info, check in the popover about the top pod info, compare it with the pod memory info on pods list page. The memory info is normal now. The bug is fixed. Thank you @yanpzhan I am still testing 4.10 but will verify this fix once we move to 4.11 Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Important: OpenShift Container Platform 4.11.0 bug fix and security update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2022:5069 The needinfo request[s] on this closed bug have been removed as they have been unresolved for 365 days |