Bug 2045972 - etcd and api server cpu mask interferes with a guaranteed workload
Summary: etcd and api server cpu mask interferes with a guaranteed workload
Alias: None
Product: OpenShift Container Platform
Classification: Red Hat
Component: Node
Version: 4.9
Hardware: Unspecified
OS: Unspecified
Target Milestone: ---
: 4.9.z
Assignee: Artyom
QA Contact: Walid A.
Depends On: 2041583
Blocks: 2050131
TreeView+ depends on / blocked
Reported: 2022-01-26 06:00 UTC by OpenShift BugZilla Robot
Modified: 2022-02-28 08:56 UTC (History)
8 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Last Closed: 2022-02-23 21:25:39 UTC
Target Upstream Version:

Attachments (Terms of Use)

System ID Private Priority Status Summary Last Updated
Github openshift kubernetes pull 1144 0 None open [release-4.9] Bug 2045972: UPSTREAM: <carry>: set correctly static pods CPUs when workload partitioning is disabled 2022-01-31 08:58:19 UTC
Red Hat Product Errata RHBA-2022:0488 0 None None None 2022-02-14 12:01:26 UTC

Comment 6 errata-xmlrpc 2022-02-14 12:00:57 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (OpenShift Container Platform 4.9.21 bug fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.


Comment 9 Martin Sivák 2022-02-22 14:12:16 UTC
> [core@ngdc-rcp-oe20-1 ~]$ kubectl -n openshift-cluster-node-tuning-operator exec  cluster-node-tuning-operator-647cbd9f67-vrqvl --  cat /sys/fs/cgroup/cpuset/cpuset.cpus

How is this the same issue? I clearly see the cpus 2-3 and 34-35 were removed from the cpuset likely due to a guaranteed workload consuming them.

Just to remind you, the original issue was etcd having:

0-63 - meaning guaranteed cpus were not removed at all form the cpu mask

> Can someone confirm if the fix of this issue "etcd and api server cpu mask interferes with a guaranteed workload" is to remove the pod annotation: target.workload.openshift.io/management: {"effect":"PreferredDuringScheduling"} ?

No, do not touch this annotation, it is necessary for workload partitioning to work properly.

This bug is worked around by using workload partitioning and fixed without it by using OCP 4.9.21.

Comment 10 Stephen Cuppett 2022-02-23 21:25:39 UTC
This has already been delivered. Please open a new bug (versus re-opening) and link for where the issue (or a new issue) is believed to exist.

Comment 12 Artyom 2022-02-28 08:56:19 UTC
I think Martin already answered it under https://bugzilla.redhat.com/show_bug.cgi?id=2045972#c9.

Note You need to log in before you can comment on or make changes to this bug.