Description of problem: Upgrading 4.1.0 GA to 4.1.1 the network operator got stuck in progressing with these messages at the bottom of the log: 2019/06/11 19:36:58 Updated ClusterOperator with status: conditions: - lastTransitionTime: "2019-06-11T11:53:34Z" status: "False" type: Degraded - lastTransitionTime: "2019-06-11T18:35:35Z" message: |- DaemonSet "openshift-multus/multus" is not available (awaiting 1 nodes) DaemonSet "openshift-sdn/ovs" is not available (awaiting 1 nodes) DaemonSet "openshift-sdn/sdn" is not available (awaiting 1 nodes) reason: Deploying status: "True" type: Progressing - lastTransitionTime: "2019-06-11T11:54:14Z" status: "True" type: Available extension: null versions: - name: operator version: 4.1.1 # oc get all -n openshift-sdn NAME READY STATUS RESTARTS AGE pod/ovs-5xwcf 1/1 Running 1 76m pod/ovs-ck2r5 1/1 Running 1 79m pod/ovs-lbzdp 1/1 Running 0 80m pod/ovs-lch8t 1/1 Running 1 78m pod/ovs-sx88h 1/1 Running 1 80m pod/ovs-z4dq5 1/1 Running 1 77m pod/ovs-znk4z 1/1 Running 1 78m pod/sdn-2kq6q 1/1 Running 2 81m pod/sdn-4ftks 1/1 Running 2 79m pod/sdn-4xhsp 1/1 Running 0 80m pod/sdn-6zj9r 1/1 Running 2 80m pod/sdn-79zws 1/1 Running 2 80m pod/sdn-7g5pl 1/1 Running 1 79m pod/sdn-controller-lsjcv 1/1 Running 1 80m pod/sdn-controller-skqrq 1/1 Running 1 79m pod/sdn-controller-wzzc2 1/1 Running 1 80m pod/sdn-prkbl 1/1 Running 2 80m NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE service/sdn ClusterIP None <none> 9101/TCP 7h45m NAME DESIRED CURRENT READY UP-TO-DATE AVAILABLE NODE SELECTOR AGE daemonset.apps/ovs 7 7 7 7 7 beta.kubernetes.io/os=linux 7h45m daemonset.apps/sdn 7 7 7 7 7 beta.kubernetes.io/os=linux 7h45m daemonset.apps/sdn-controller 3 3 3 3 3 node-role.kubernetes.io/master= 7h45m # oc get clusteroperators NAME VERSION AVAILABLE PROGRESSING DEGRADED SINCE authentication 4.1.1 True False False 7h26m cloud-credential 4.1.1 True False False 7h45m cluster-autoscaler 4.1.1 True False False 7h45m console 4.1.1 True False False 7h36m dns 4.1.1 True False False 7h45m image-registry 4.1.1 True False False 64m ingress 4.1.1 True False False 7h38m kube-apiserver 4.1.1 True False False 7h42m kube-controller-manager 4.1.1 True False False 7h42m kube-scheduler 4.1.1 True False False 7h42m machine-api 4.1.1 True False False 7h45m machine-config 4.1.1 True False False 7h44m marketplace 4.1.1 True False False 62m monitoring 4.1.1 True False False 60m network 4.1.1 True True False 7h45m node-tuning 4.1.1 True False False 94m openshift-apiserver 4.1.1 True False False 62m openshift-controller-manager 4.1.1 True False False 7h44m openshift-samples 4.1.1 True False False 82m operator-lifecycle-manager 4.1.1 True False False 7h44m operator-lifecycle-manager-catalog 4.1.1 True False False 7h44m service-ca 4.1.1 True False False 7h45m service-catalog-apiserver 4.1.1 True False False 7h41m service-catalog-controller-manager 4.1.1 True False False 7h41m storage 4.1.1 True False False 94m Version-Release number of selected component (if applicable): 4.1.1 final How reproducible: Unknown. 1/1 Steps to Reproduce: 1. Install 4.1.0 GA 2. Change channel to prerelease-4.1 3. oc adm upgrade --to=4.1.1 Actual results: See description I will link must-gather
apiVersion: config.openshift.io/v1 kind: ClusterOperator metadata: creationTimestamp: "2019-06-11T11:53:33Z" generation: 1 name: network resourceVersion: "218327" selfLink: /apis/config.openshift.io/v1/clusteroperators/network uid: 89c6ba43-8c3f-11e9-9e37-0687d8271f94 spec: {} status: conditions: - lastTransitionTime: "2019-06-11T11:53:34Z" status: "False" type: Degraded - lastTransitionTime: "2019-06-11T18:35:35Z" message: |- DaemonSet "openshift-multus/multus" is not available (awaiting 1 nodes) DaemonSet "openshift-sdn/ovs" is not available (awaiting 1 nodes) DaemonSet "openshift-sdn/sdn" is not available (awaiting 1 nodes) reason: Deploying status: "True" type: Progressing - lastTransitionTime: "2019-06-11T11:54:14Z" status: "True" type: Available extension: null versions: - name: operator version: 4.1.1
I think this is a dup of bug 1718375. But folks who actually understand it better should feel free to reopen if I'm wrong ;). *** This bug has been marked as a duplicate of bug 1718375 ***