Bug 1794493
| Summary: | Applying "ctrcfg" causes cri-o to fail to start on node reboot | |||
|---|---|---|---|---|
| Product: | OpenShift Container Platform | Reporter: | Urvashi Mohnani <umohnani> | |
| Component: | Node | Assignee: | Urvashi Mohnani <umohnani> | |
| Status: | CLOSED ERRATA | QA Contact: | Sunil Choudhary <schoudha> | |
| Severity: | unspecified | Docs Contact: | ||
| Priority: | unspecified | |||
| Version: | 4.4 | CC: | amurdaca, aos-bugs, fpaoline, jokerman, pruan, rphillips, wking | |
| Target Milestone: | --- | |||
| Target Release: | 4.4.0 | |||
| Hardware: | Unspecified | |||
| OS: | Unspecified | |||
| Whiteboard: | ||||
| Fixed In Version: | Doc Type: | Bug Fix | ||
| Doc Text: |
Cause: Applying a MachineConfig to a Machine Config Pool would sometimes cause cri-o to fail to start after the node rebooted.
Consequence: Pods would be non-functional.
Fix:
Result: Applying a MachineConfig will work properly on node reboot.
|
Story Points: | --- | |
| Clone Of: | ||||
| : | 1794495 (view as bug list) | Environment: | ||
| Last Closed: | 2020-05-04 11:26:35 UTC | Type: | Bug | |
| Regression: | --- | Mount Type: | --- | |
| Documentation: | --- | CRM: | ||
| Verified Versions: | Category: | --- | ||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
| Cloudforms Team: | --- | Target Upstream Version: | ||
| Embargoed: | ||||
| Bug Depends On: | ||||
| Bug Blocks: | 1794495 | |||
|
Description
Urvashi Mohnani
2020-01-23 17:36:15 UTC
verified with 4.4.0-0.nightly-2020-02-10-165717
1. oc edit machineconfigpool worker
labels:
custom-crio: high-pid-limit <-- add this line
machineconfiguration.openshift.io/mco-built-in: ""
2. create a config yaml
apiVersion: machineconfiguration.openshift.io/v1
kind: ContainerRuntimeConfig
metadata:
name: set-log-and-pid
spec:
machineConfigPoolSelector:
matchLabels:
custom-crio: high-pid-limit ### <-- this must match the label created in step #1
containerRuntimeConfig:
pidsLimit: 2048
logLevel: debug
3. oc create -f config.yaml
4. wait until all worker nodes come up
pruan@fedora-vm ~/workspace/testcases/1794493 $ oc get nodes
NAME STATUS ROLES AGE VERSION
ip-10-0-137-163.us-west-1.compute.internal Ready worker 25h v1.17.1
ip-10-0-138-58.us-west-1.compute.internal Ready master 25h v1.17.1
ip-10-0-139-15.us-west-1.compute.internal Ready worker 25h v1.17.1
ip-10-0-143-23.us-west-1.compute.internal Ready master 25h v1.17.1
ip-10-0-150-105.us-west-1.compute.internal Ready,SchedulingDisabled worker 25h v1.17.1
ip-10-0-159-232.us-west-1.compute.internal Ready master 25h v1.17.1
5. verify the limits defined in the config.yaml is applied to the nodes.
pruan@fedora-vm ~/workspace/testcases/1794493 $ oc debug node/ip-10-0-137-163.us-west-1.compute.internal
Starting pod/ip-10-0-137-163us-west-1computeinternal-debug ...
To use host binaries, run `chroot /host`
Pod IP: 10.0.137.163
If you don't see a command prompt, try pressing enter.
sh-4.2# chroot /host
sh-4.4# cat /etc/crio/crio.conf | grep limit
pids_limit = 2048
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2020:0581 |