Bug 2006944

Summary: mgr pod with status running got stuck into init without any disruptive action
Product: [Red Hat Storage] Red Hat OpenShift Data Foundation Reporter: Aman Agrawal <amagrawa>
Component: rookAssignee: Travis Nielsen <tnielsen>
Status: CLOSED INSUFFICIENT_DATA QA Contact: Elad <ebenahar>
Severity: medium Docs Contact:
Priority: unspecified    
Version: 4.9CC: aos-bugs, madam, muagarwa, nthomas, ocs-bugs, odf-bz-bot, sostapov
Target Milestone: ---   
Target Release: ---   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2021-10-11 18:25:00 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Comment 4 Travis Nielsen 2021-09-27 20:15:03 UTC
The containers are all stuck in PodInitializing state while the overall pod phase is "Pending". 
Can we get the "oc pod describe <mgr-pod>" to see if there are more details on the pending status? This usually means a volume cannot be mounted. Or is a connection to the cluster available to take a closer look?

status:
  conditions:
  - lastProbeTime: null
    lastTransitionTime: "2021-09-22T13:53:21Z"
    message: 'containers with incomplete status: [chown-container-data-dir]'
    reason: ContainersNotInitialized
    status: "False"
    type: Initialized
  - lastProbeTime: null
    lastTransitionTime: "2021-09-22T13:53:21Z"
    message: 'containers with unready status: [mgr log-collector]'
    reason: ContainersNotReady
    status: "False"
    type: Ready
  - lastProbeTime: null
    lastTransitionTime: "2021-09-22T13:53:21Z"
    message: 'containers with unready status: [mgr log-collector]'
    reason: ContainersNotReady
    status: "False"
    type: ContainersReady
  - lastProbeTime: null
    lastTransitionTime: "2021-09-22T13:53:21Z"
    status: "True"
    type: PodScheduled
  containerStatuses:
  - image: quay.io/rhceph-dev/rhceph@sha256:0d7d408ccfb1e0f7dd9b846af0724005487b00171b10e4a884d360fdc1dd0ad6
    imageID: ""
    lastState: {}
    name: log-collector
    ready: false
    restartCount: 0
    started: false
    state:
      waiting:
        reason: PodInitializing
  - image: quay.io/rhceph-dev/rhceph@sha256:0d7d408ccfb1e0f7dd9b846af0724005487b00171b10e4a884d360fdc1dd0ad6
    imageID: ""
    lastState: {}
    name: mgr
    ready: false
    restartCount: 0
    started: false
    state:
      waiting:
        reason: PodInitializing
  hostIP: 10.1.160.184
  initContainerStatuses:
  - image: quay.io/rhceph-dev/rhceph@sha256:0d7d408ccfb1e0f7dd9b846af0724005487b00171b10e4a884d360fdc1dd0ad6
    imageID: ""
    lastState: {}
    name: chown-container-data-dir
    ready: false
    restartCount: 0
    state:
      waiting:
        reason: PodInitializing
  phase: Pending

Comment 8 Travis Nielsen 2021-10-11 18:25:00 UTC
Let's reopen if you get another repro