Bug 1292964
Summary: | OpenShift doesn't notice that Docker Storage is, or is reaching that state of being, full | ||
---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | Eric Jones <erjones> |
Component: | Node | Assignee: | Derek Carr <decarr> |
Status: | CLOSED ERRATA | QA Contact: | DeShuai Ma <dma> |
Severity: | medium | Docs Contact: | |
Priority: | urgent | ||
Version: | 3.1.0 | CC: | agoldste, aos-bugs, dma, erich, jokerman, kurktchiev, mgandolf, mmccomas, pep, pweil, rhowe, tdawson |
Target Milestone: | --- | ||
Target Release: | --- | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: |
OpenShift Enterprise 3.1
Current issue found on a a node hosted in OpenStack
|
|
Last Closed: | 2017-01-18 12:38:53 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: | |||
Bug Depends On: | |||
Bug Blocks: | 1267746, 1292845 |
Description
Eric Jones
2015-12-18 21:23:38 UTC
So I reported this originally, and as pointed out it blocks on the image clean up. I had to manually run clean up steps (specifically the image ones) from the OSE 3 docs in order to get it to actually clean up after itself. I have not tried running the ansible playbooks, but overall it seems like the clean up process should be happening automatically at least semi scheduled. Which in my case did not seem to be the case Has this been fixed in 3.1 with this PR? https://github.com/openshift/origin/pull/5599 No, we have no fix yet. This work is being done upstream. According to a proposal [1], everything we need (volume accounting) shall be covered. As of now, only the volume interface [2] is in place. Unfortunatelly, accounting for the host_path volumes has been recently disabled [3] due to high CPU load. Neither NFS nor AWS nor GCE is supported yet. [1] https://github.com/kubernetes/kubernetes/blob/master/docs/proposals/disk-accounting.md#introduction [2] https://github.com/kubernetes/kubernetes/pull/18232 [3] https://github.com/kubernetes/kubernetes/pull/23446 This is a new feature in Kubernetes 1.4 that just got merged. You can read the feature description here: https://github.com/kubernetes/kubernetes/blob/master/docs/proposals/kubelet-eviction.md Users will be able to set thresholds for both the rootfs (i.e. nodefs) and the imagefs (i.e. docker storage). If those thresholds are met, the node will report disk pressure, perform image gc, and evict pods on the node to reduce disk pressure to a stable state. While the node reports disk pressure, no additional pods are admitted to the node for execution. Marking this upcoming release. I am assuming these are going to be exposed in some way in the node configs? Boris - correct, in 3.4, users will be able to configure the values in node-config. OCP 3.4 has added support to handle disk pressure based on the work we did in upstream Kubernetes 1.4. For details: http://kubernetes.io/docs/admin/out-of-resource/ I am moving this to ON_QA as a result. Test on openshift v3.4.0.15+9c963ec, disk pressure works as expected. detail in the card. https://trello.com/c/3LvGAHr3/371-5-kubelet-evicts-pods-when-low-on-disk-node-reliability Verify this bug. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2017:0066 |