Bug 1930960 - After a disaster recovery pods a stuck in "NodeAffinity" state and not running
Summary: After a disaster recovery pods a stuck in "NodeAffinity" state and not running
Alias: None
Product: OpenShift Container Platform
Classification: Red Hat
Component: Node
Version: 4.5
Hardware: All
OS: Linux
Target Milestone: ---
: 4.6.z
Assignee: Elana Hashman
QA Contact: Sunil Choudhary
Depends On: 1868645
TreeView+ depends on / blocked
Reported: 2021-02-19 21:23 UTC by OpenShift BugZilla Robot
Modified: 2021-03-25 04:45 UTC (History)
14 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Cause: Node is marked as Ready and admits pods before it has a chance to sync. Consequence: Pod status may go out of sync, sometimes many are stuck in NodeAffinity, at node startup for a node that is not cordoned. Fix: Do not mark node as Ready until Node has synced with API servers at least once. Result: Pods should not get stuck in NodeAffinity after e.g. a cold cluster restart.
Clone Of:
Last Closed: 2021-03-25 04:45:13 UTC
Target Upstream Version:

Attachments (Terms of Use)

System ID Private Priority Status Summary Last Updated
Github openshift kubernetes pull 582 0 None open [release-4.6] Bug 1930960: UPSTREAM: 94087: kubelet: node sync at least once 2021-02-19 21:24:13 UTC
Red Hat Product Errata RHBA-2021:0825 0 None None None 2021-03-25 04:45:21 UTC

Comment 5 errata-xmlrpc 2021-03-25 04:45:13 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory (OpenShift Container Platform 4.6.22 bug fix update), and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.


Note You need to log in before you can comment on or make changes to this bug.