Note: This bug is displayed in read-only format because the product is no longer active in Red Hat Bugzilla.

Bug 1848450

Summary: deployment prometheus-operator is not ready- OCP 4.2.25
Product: OpenShift Container Platform Reporter: mchebbi <mchebbi>
Component: MonitoringAssignee: Sergiusz Urbaniak <surbania>
Status: CLOSED NOTABUG QA Contact: Junqi Zhao <juzhao>
Severity: high Docs Contact:
Priority: unspecified    
Version: 4.2.0CC: alegrand, anpicker, dhansen, erooth, kakkoyun, lcosic, mloibl, pkrupa, scuppett, surbania
Target Milestone: ---   
Target Release: 4.6.0   
Hardware: Unspecified   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2020-06-19 07:01:00 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Description mchebbi@redhat.com 2020-06-18 11:34:56 UTC
The customer did an upgrade of his cluster from 4.2.25 => 4.3.18 but he gets seral operators not updated. I have fixed the machine-config operator but still the prometheus-operator is not ready  and the monitoring operator degraded

NAME                                       VERSION   AVAILABLE   PROGRESSING   DEGRADED   SINCE
dns                                        4.2.25    True        True          False      160d
monitoring                                 4.2.25    False       True          True       14d
network                                    4.2.25    True        True          False      160d
sl-uosbast1t:~ # oc project openshift-monitoring
Now using project "openshift-monitoring" on server "https://api.ocp.corp.wan:6443".
sl-uosbast1t:~ # oc get pods | grep operator
cluster-monitoring-operator-bc89787ff-fptbl   1/1     Running   0          82d
prometheus-operator-f8fc5b975-fh8vg           1/1     Running   0          82d
sl-uosbast1t:~ # oc get deployments
NAME                          READY   UP-TO-DATE   AVAILABLE   AGE
cluster-monitoring-operator   1/1     1            1           160d
grafana                       1/1     1            1           160d
kube-state-metrics            1/1     1            1           160d
openshift-state-metrics       1/1     1            1           160d
prometheus-adapter            2/2     2            2           160d
prometheus-operator           0/1     1            0           160d
telemeter-client              0/1     1            0           61d

 conditions:
    - lastTransitionTime: 2020-06-16T13:50:21Z
      message: Rolling out the stack.
      reason: RollOutInProgress
      status: "True"
      type: Progressing
    - lastTransitionTime: 2020-06-16T09:04:25Z
      message: 'Failed to rollout the stack. Error: running task Updating Prometheus Operator failed: reconciling Prometheus Operator Deployment failed: updating deployment object failed: waiting for DeploymentRollout of prometheus-operator: deployment prometheus-operator is not ready. status: (replicas: 1, updated: 1, ready: 0, unavailable: 1)'
      reason: UpdatingPrometheusOperatorFailed
      status: "True"
      type: Degraded
    - lastTransitionTime: 2020-06-16T13:50:21Z
      message: Rollout of the monitoring stack is in progress. Please wait until it
        finishes.
      reason: RollOutInProgress
      status: "True"
      type: Upgradeable


All needed information are available through this link : https://bit.ly/2YdDGZi

Comment 1 Stephen Cuppett 2020-06-18 14:44:30 UTC
Setting target release to the active development branch (4.6.0). For fixes, if any, requested/required on previous versions, cloned BZs targeting those release z-streams will be created.

Comment 2 Pawel Krupa 2020-06-19 07:01:00 UTC
You have 6 nodes:
3 of those are master nodes which are tainted by default and pods are not allowed to be scheduled there
3 of those (worker nodes) don't have enough CPU for new nodes.
You need to increase worker nodes resources or add new nodes.

Comment 3 Pawel Krupa 2020-06-19 07:01:27 UTC
*** Bug 1848833 has been marked as a duplicate of this bug. ***