Bug 2025811
Summary: | Upgrading to Satellite 6.9.6 and above introduces an increase in system memory consumption causing Pulp activities to fail with OOM at certain times | ||
---|---|---|---|
Product: | Red Hat Satellite | Reporter: | Sayan Das <saydas> |
Component: | Installation | Assignee: | satellite6-bugs <satellite6-bugs> |
Status: | CLOSED ERRATA | QA Contact: | isinghal |
Severity: | high | Docs Contact: | |
Priority: | high | ||
Version: | 6.9.6 | CC: | ahumbe, cdonnell, dalley, dhjoshi, ehelms, hakon.gislason, isinghal, jhutar, lvrtelov, mmccune, peter.vreman, pmendezh, sadas, tasander |
Target Milestone: | 6.11.0 | Keywords: | Performance, PrioBumpGSS, Triaged |
Target Release: | Unused | ||
Hardware: | All | ||
OS: | All | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | If docs needed, set a value | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2022-07-05 14:30:29 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Sayan Das
2021-11-23 05:54:17 UTC
I build a small script myself to write every 10 seconds the status-text from the foreman systemd unit that is providing the actual usage of pumma. while true; do echo "$(date +%Y%m%d-%H%M%S) $(systemctl show foreman --prop=StatusText)"; sleep 10; done This reveals that in my situation with 200-250 clients attached only max 30 threads are used. On an other Sat6 server having more content changes shows a neglitable increase due to API calls, because most API calls must be sequential due to functional dependencies of the steps to update content. And parallism in the API calls on the content part is therefor rare. And even in that case lets assuma 10 API calls concurrent are providing some headroom. There are also almost no long running API calls, they are all async and dynflowized. The highest parallelism I was able to trigger was from a daily cron-job without splay that was calling 'insighst-client --compliance' on ~50 servers, this showed a peak use of ~20 threads in puma at the start of the insights-client where it is querying the cloud.redhat.com for the host details on assigned compliance policies. After that the runtimes of the complaince clients is already starting to vary in duration and for puma there are almost no concurrent requests done anymore. The default from Sat6 with the 16 cpu was 24 workers x 5 threads = 120 threads in total. Extrapolating my real-world use on the foreman requests the 120 puma threads are maybe relevant for a 10000+ clients setup, but not for a setup <500 clients. What is missing here is a reporting functionality how much puma threads are actually used. Based on this input a more realistic value for the number of puma workers can be set. It is a known that puma needs 1.0-1.5GB per worker. If you chose as default 1.5x that means 1.5x1.5GB a mnemory consumption of 2.25GB/CPU. This ratio does not match the current resource requirements documented. Especially also thinking of all the other resource databases,candlepin,pulp,puppet needing to run ont he same server Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (Moderate: Satellite 6.11 Release), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2022:5498 The needinfo request[s] on this closed bug have been removed as they have been unresolved for 365 days |