Bug 2006944 - mgr pod with status running got stuck into init without any disruptive action
Summary: mgr pod with status running got stuck into init without any disruptive action
Keywords:
Status: CLOSED INSUFFICIENT_DATA
Alias: None
Product: Red Hat OpenShift Data Foundation
Classification: Red Hat Storage
Component: rook
Version: 4.9
Hardware: Unspecified
OS: Unspecified
unspecified
medium
Target Milestone: ---
: ---
Assignee: Travis Nielsen
QA Contact: Elad
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2021-09-22 17:12 UTC by Aman Agrawal
Modified: 2023-08-09 17:03 UTC (History)
7 users (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed: 2021-10-11 18:25:00 UTC
Embargoed:


Attachments (Terms of Use)

Comment 4 Travis Nielsen 2021-09-27 20:15:03 UTC
The containers are all stuck in PodInitializing state while the overall pod phase is "Pending". 
Can we get the "oc pod describe <mgr-pod>" to see if there are more details on the pending status? This usually means a volume cannot be mounted. Or is a connection to the cluster available to take a closer look?

status:
  conditions:
  - lastProbeTime: null
    lastTransitionTime: "2021-09-22T13:53:21Z"
    message: 'containers with incomplete status: [chown-container-data-dir]'
    reason: ContainersNotInitialized
    status: "False"
    type: Initialized
  - lastProbeTime: null
    lastTransitionTime: "2021-09-22T13:53:21Z"
    message: 'containers with unready status: [mgr log-collector]'
    reason: ContainersNotReady
    status: "False"
    type: Ready
  - lastProbeTime: null
    lastTransitionTime: "2021-09-22T13:53:21Z"
    message: 'containers with unready status: [mgr log-collector]'
    reason: ContainersNotReady
    status: "False"
    type: ContainersReady
  - lastProbeTime: null
    lastTransitionTime: "2021-09-22T13:53:21Z"
    status: "True"
    type: PodScheduled
  containerStatuses:
  - image: quay.io/rhceph-dev/rhceph@sha256:0d7d408ccfb1e0f7dd9b846af0724005487b00171b10e4a884d360fdc1dd0ad6
    imageID: ""
    lastState: {}
    name: log-collector
    ready: false
    restartCount: 0
    started: false
    state:
      waiting:
        reason: PodInitializing
  - image: quay.io/rhceph-dev/rhceph@sha256:0d7d408ccfb1e0f7dd9b846af0724005487b00171b10e4a884d360fdc1dd0ad6
    imageID: ""
    lastState: {}
    name: mgr
    ready: false
    restartCount: 0
    started: false
    state:
      waiting:
        reason: PodInitializing
  hostIP: 10.1.160.184
  initContainerStatuses:
  - image: quay.io/rhceph-dev/rhceph@sha256:0d7d408ccfb1e0f7dd9b846af0724005487b00171b10e4a884d360fdc1dd0ad6
    imageID: ""
    lastState: {}
    name: chown-container-data-dir
    ready: false
    restartCount: 0
    state:
      waiting:
        reason: PodInitializing
  phase: Pending

Comment 8 Travis Nielsen 2021-10-11 18:25:00 UTC
Let's reopen if you get another repro


Note You need to log in before you can comment on or make changes to this bug.