One of the revision pruner pods shows as OOM killed: [root@jumpserver ~]# oc get pods -A -o wide | grep -iv runn | grep -iv compl NAMESPACE NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES openshift-kube-scheduler revision-pruner-7-master01.example.com 0/1 OOMKilled 0 13d 192.168.3.96 master01.example.com <none> <none> [root@jumpserver ~]# oc describe pod -n openshift-kube-scheduler revision-pruner-7-master01.example.com Name: revision-pruner-7-master01.example.com Namespace: openshift-kube-scheduler Priority: 2000001000 Priority Class Name: system-node-critical Node: master01.example.com/10.0.0.251 Start Time: Wed, 29 Apr 2020 10:29:15 +0200 Labels: app=pruner Annotations: k8s.v1.cni.cncf.io/networks-status: Status: Succeeded IP: 192.168.3.96 IPs: IP: 192.168.3.96 Containers: pruner: Container ID: cri-o://<container id> Image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:<checksum> Image ID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:<checksum> Port: <none> Host Port: <none> Command: cluster-kube-scheduler-operator prune Args: -v=4 --max-eligible-revision=7 --protected-revisions=1,2,3,4,5,6,7 --resource-dir=/etc/kubernetes/static-pod-resources --static-pod-name=kube-scheduler-pod State: Terminated Reason: OOMKilled Exit Code: 0 Started: Wed, 29 Apr 2020 10:29:18 +0200 Finished: Wed, 29 Apr 2020 10:29:18 +0200 Ready: False Restart Count: 0 Environment: <none> Mounts: /etc/kubernetes/ from kubelet-dir (rw) /var/run/secrets/kubernetes.io/serviceaccount from installer-sa-token-aaaaa (ro) Conditions: Type Status Initialized True Ready False ContainersReady False PodScheduled True Volumes: kubelet-dir: Type: HostPath (bare host directory volume) Path: /etc/kubernetes/ HostPathType: installer-sa-token-aaaaa: Type: Secret (a volume populated by a Secret) SecretName: installer-sa-token-aaaaa Optional: false QoS Class: BestEffort Node-Selectors: <none> Tolerations: Events: <none> --------------- root@jumpserver ~]# oc logs -n openshift-kube-scheduler revision-pruner-7-master01.example.com unable to retrieve container logs for cri-o://<container id>[root@jumpserver ~]# [root@jumpserver ~]# [root@jumpserver ~]# Could be related to any of the following so please close as duplicate if that's the case: https://bugzilla.redhat.com/show_bug.cgi?id=1800609 ; https://bugzilla.redhat.com/show_bug.cgi?id=1799079 ; https://bugzilla.redhat.com/show_bug.cgi?id=1792501 ; https://github.com/openshift/origin/pull/24596
Yes, these look like similar cases to aforementioned, but I'll let the node team give the final call if it's a duplicate.
*** This bug has been marked as a duplicate of bug 1800609 ***