Bug 1592303
| Summary: | After running redeploy-certificates.yml playbook in OCP 3.9 webconsole stop working. | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Product: | OpenShift Container Platform | Reporter: | Joel Rosental R. <jrosenta> | ||||||||
| Component: | Installer | Assignee: | Vadim Rutkovsky <vrutkovs> | ||||||||
| Status: | CLOSED ERRATA | QA Contact: | Yadan Pei <yapei> | ||||||||
| Severity: | high | Docs Contact: | |||||||||
| Priority: | unspecified | ||||||||||
| Version: | 3.9.0 | CC: | aos-bugs, aship, dcaldwel, dmoessne, farandac, fshaikh, gsapienz, jokerman, jolee, mifiedle, mmccomas, nate.childers, oarribas, openshift-bugs-escalate, rsandu, spagno, stwalter, vlaad, vrutkovs, wmeng, yanpzhan, yapei | ||||||||
| Target Milestone: | --- | ||||||||||
| Target Release: | 3.9.z | ||||||||||
| Hardware: | Unspecified | ||||||||||
| OS: | Linux | ||||||||||
| Whiteboard: | |||||||||||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |||||||||
| Doc Text: | Story Points: | --- | |||||||||
| Clone Of: | |||||||||||
| : | 1596557 1667981 (view as bug list) | Environment: | |||||||||
| Last Closed: | 2018-09-22 04:53:09 UTC | Type: | Bug | ||||||||
| Regression: | --- | Mount Type: | --- | ||||||||
| Documentation: | --- | CRM: | |||||||||
| Verified Versions: | Category: | --- | |||||||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||||||
| Embargoed: | |||||||||||
| Bug Depends On: | 1596233 | ||||||||||
| Bug Blocks: | 1596557, 1623987, 1667981 | ||||||||||
| Attachments: |
|
||||||||||
|
Description
Joel Rosental R.
2018-06-18 12:14:32 UTC
The playbook should also take in consideration if a new named certificate is provided for the public url, so the secret is created using it instead of the selfsigned one. PR for master: https://github.com/openshift/openshift-ansible/pull/8891 Hi, I am the critical Situation Manager for EMEA and would like to ask please on the ETA of this Fix.. If this will be scheduled in the next Openshift-ansible errata when will that date be? PR for 3.9 merged https://github.com/openshift/openshift-ansible/pull/9005 Tried with fix in pr9005, after run "ansible-playbook -i /path/to/inventory /usr/share/ansible/openshift-ansible/playbooks/redeploy-certificates.yml", web pod could not running due to lacking of secret "webconsole-serving-cert"
# oc get pod
NAME READY STATUS RESTARTS AGE
webconsole-6b74f5c578-n8mdg 0/1 ContainerCreating 0 6m
# oc describe pod webconsole-6b74f5c578-n8mdg
Name: webconsole-6b74f5c578-n8mdg
Namespace: openshift-web-console
Node: qe-juzhao-39-qeos-1-master-etcd-1/172.16.120.82
Start Time: Thu, 28 Jun 2018 06:49:50 -0400
Labels: app=openshift-web-console
pod-template-hash=2630917134
webconsole=true
Annotations: openshift.io/scc=restricted
Status: Pending
IP:
Controlled By: ReplicaSet/webconsole-6b74f5c578
Containers:
webconsole:
Container ID:
Image: registry.reg-aws.openshift.com:443/openshift3/ose-web-console:v3.9.31
Image ID:
Port: 8443/TCP
Command:
/usr/bin/origin-web-console
--audit-log-path=-
-v=0
--config=/var/webconsole-config/webconsole-config.yaml
State: Waiting
Reason: ContainerCreating
Ready: False
Restart Count: 0
Requests:
cpu: 100m
memory: 100Mi
Liveness: exec [/bin/sh -c if [[ ! -f /tmp/webconsole-config.hash ]]; then \
md5sum /var/webconsole-config/webconsole-config.yaml > /tmp/webconsole-config.hash; \
elif [[ $(md5sum /var/webconsole-config/webconsole-config.yaml) != $(cat /tmp/webconsole-config.hash) ]]; then \
echo 'webconsole-config.yaml has changed.'; \
exit 1; \
fi && curl -k -f https://0.0.0.0:8443/console/] delay=0s timeout=1s period=10s #success=1 #failure=3
Readiness: http-get https://:8443/healthz delay=0s timeout=1s period=10s #success=1 #failure=3
Environment: <none>
Mounts:
/var/run/secrets/kubernetes.io/serviceaccount from webconsole-token-26l96 (ro)
/var/serving-cert from serving-cert (rw)
/var/webconsole-config from webconsole-config (rw)
Conditions:
Type Status
Initialized True
Ready False
PodScheduled True
Volumes:
serving-cert:
Type: Secret (a volume populated by a Secret)
SecretName: webconsole-serving-cert
Optional: false
webconsole-config:
Type: ConfigMap (a volume populated by a ConfigMap)
Name: webconsole-config
Optional: false
webconsole-token-26l96:
Type: Secret (a volume populated by a Secret)
SecretName: webconsole-token-26l96
Optional: false
QoS Class: Burstable
Node-Selectors: node-role.kubernetes.io/master=true
Tolerations: node.kubernetes.io/memory-pressure:NoSchedule
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 1m default-scheduler Successfully assigned webconsole-6b74f5c578-n8mdg to qe-juzhao-39-qeos-1-master-etcd-1
Normal SuccessfulMountVolume 1m kubelet, qe-juzhao-39-qeos-1-master-etcd-1 MountVolume.SetUp succeeded for volume "webconsole-config"
Normal SuccessfulMountVolume 1m kubelet, qe-juzhao-39-qeos-1-master-etcd-1 MountVolume.SetUp succeeded for volume "webconsole-token-26l96"
Warning FailedMount 27s (x8 over 1m) kubelet, qe-juzhao-39-qeos-1-master-etcd-1 MountVolume.SetUp failed for volume "serving-cert" : secrets "webconsole-serving-cert" not found
[root@qe-juzhao-39-qeos-1-master-etcd-1 ~]# oc get secret
NAME TYPE DATA AGE
builder-dockercfg-4xx56 kubernetes.io/dockercfg 1 9h
builder-token-nx57t kubernetes.io/service-account-token 4 9h
builder-token-zrr24 kubernetes.io/service-account-token 4 9h
default-dockercfg-fjvpm kubernetes.io/dockercfg 1 9h
default-token-4gsx6 kubernetes.io/service-account-token 4 9h
default-token-ssrnq kubernetes.io/service-account-token 4 9h
deployer-dockercfg-m2c92 kubernetes.io/dockercfg 1 9h
deployer-token-4pd56 kubernetes.io/service-account-token 4 9h
deployer-token-z2c68 kubernetes.io/service-account-token 4 9h
webconsole-dockercfg-bg5hc kubernetes.io/dockercfg 1 9h
webconsole-token-26l96 kubernetes.io/service-account-token 4 9h
webconsole-token-c8m6j kubernetes.io/service-account-token 4 9h
Please attach the output of `ansible-playbook -vvv` and `oc get ev -n openshift-web-console` output Created attachment 1455244 [details]
oc get ev -n openshift-web-console
Created attachment 1455251 [details]
`ansible-playbook -vvv` log
Created attachment 1455280 [details]
Controllers
Controllers are not recreating this secret for some reason, see log attached
Created a PR https://github.com/openshift/openshift-ansible/pull/9012 to workaround the issue with controllers (bug #1596233) Successfully tested with the hotfix build in comment 15. Prior to running the playbook the web console pod was stuck in ContainerCreating as described in comment 8. After running the playbook, the web console pod was Running and the web console was accessible via browser. Not marking this Verified as of now - it needs to be tested in an official puddle. Fix is available in openshift-ansible-3.9.37-1 openshift v3.9.37 kubernetes v1.9.1+a0ce1bc657 openshift-ansible-3.9.37-1.git.0.51fbd81.el7.noarch.rpm Using command: "ansible-playbook -i /path/to/inventory /usr/share/ansible/openshift-ansible/playbooks/redeploy-certificates.yml" Redeploy certificates on OCP 3.9 env with above package version. the web console pod and cert are redeployed and web console could be accessed successfully. The bug has been fixed, so move it to Verified. The PR for 3.9 merged https://github.com/openshift/openshift-ansible/pull/9005 doesn't fix if you want to re-deploy only master certificates I created the PR https://github.com/openshift/openshift-ansible/pull/9713 regards https://github.com/openshift/openshift-ansible/pull/9713 The pr is not merged yet. (In reply to Yanping Zhang from comment #22) > https://github.com/openshift/openshift-ansible/pull/9713 > The pr is not merged yet. This PR is not required, https://github.com/openshift/openshift-ansible/pull/9012 is sufficient to get web console working after certs were redeployed. Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2018:2658 |