Bug 1710404
| Summary: | Suggestion to adjust Elasticsearch OOTB cpu limit and memory request | ||
|---|---|---|---|
| Product: | OpenShift Container Platform | Reporter: | Mike Fiedler <mifiedle> |
| Component: | Logging | Assignee: | ewolinet |
| Status: | CLOSED ERRATA | QA Contact: | Mike Fiedler <mifiedle> |
| Severity: | medium | Docs Contact: | |
| Priority: | unspecified | ||
| Version: | 4.1.0 | CC: | anli, aos-bugs, ewolinet, jcantril, pweil, qitang, rmeggins |
| Target Milestone: | --- | ||
| Target Release: | 4.2.0 | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | No Doc Update | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2019-10-16 06:28:53 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Mike Fiedler
2019-05-15 13:55:22 UTC
None of our pods should have cpu limits, only memory. This came out of pportante's experience of pods getting kicked off nodes. cpu limits should be a bug. I hesitate to drop the memlimit as we explicitly set it to 16G in 3.x releases because of capacity issues. I defer to PM their desires here as now we have an alternate user experience which is similarly less then desirable. (In reply to Jeff Cantrill from comment #1) > None of our pods should have cpu limits, only memory. This came out of > pportante's experience of pods getting kicked off nodes. Was that setting the `requests` or the `limits`? resources: limits: cpu: 500m memory: 4Gi requests: cpu: 500m memory: 4Gi I would expect the requests settings would cause issues with the scheduler, if there is not a node which can spare 500m cpu or 4GB RAM. > cpu limits should > be a bug. I hesitate to drop the memlimit as we explicitly set it to 16G in > 3.x releases because of capacity issues. I defer to PM their desires here > as now we have an alternate user experience which is similarly less then > desirable. Right. The alternative to Elasticsearch not being scheduled at all, is Elasticsearch having poor performance (and the subsequent support calls about log records not showing up in Kibana, or Kibana not working at all). I opened bug 1710657 for the ES CPU limit issue. The gory details are in that bz, but here is the tl;dr request/limit excerpt from the Elasticsearch deployment YAML: resources: limits: cpu: "1" memory: 16Gi requests: cpu: "1" memory: 16Gi Mike, Is the result acceptable?
The default resource deployed by CLO:
resources
limits:
memory: 16Gi
requests:
cpu: "1"
memory: 16Gi
The default resource deployed by EO:
resources:
limits:
memory: 4Gi
requests:
cpu: 100m
memory: 1Gi
Verified on the cluster-logging-operator image from 4.2.0-0.nightly-2019-07-17-165351. Values are set as in comment 5. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2019:2922 |