Bug 1635254
Summary: | [3.11] uninstall will remove the all the docker related files and reinstall on the same cluster fails | |||
---|---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | Jatan Malde <jmalde> | |
Component: | Installer | Assignee: | Russell Teague <rteague> | |
Installer sub component: | openshift-ansible | QA Contact: | Qin Ping <piqin> | |
Status: | CLOSED ERRATA | Docs Contact: | ||
Severity: | high | |||
Priority: | low | CC: | aos-bugs, gpei, jokerman, klaas, mgugino, mmccomas, pdwyer, piqin, sdodson, sgaikwad, tripletrk | |
Version: | 3.11.0 | |||
Target Milestone: | --- | |||
Target Release: | 3.11.z | |||
Hardware: | x86_64 | |||
OS: | Linux | |||
Whiteboard: | ||||
Fixed In Version: | Doc Type: | Bug Fix | ||
Doc Text: |
Cause: All docker related packages are not removed during uninstall
Consequence: Docker is not reinstalled properly during install causing docker cli tasks to fail.
Fix: Added all related docker packages to uninstall
Result: Reinstall succeeds after running uninstall playbook
|
Story Points: | --- | |
Clone Of: | ||||
: | 1655684 (view as bug list) | Environment: | ||
Last Closed: | 2019-02-20 14:11:01 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | ||||
Bug Blocks: | 1655684 |
Description
Jatan Malde
2018-10-02 13:15:06 UTC
The context of the uninstall playbooks are to "Remove as little as possible after failing to install openshift to enable retrying the install" Some items during install might not be idempotent due to the large complexity of installing openshift. We are not making any attempt to restore a host to any condition other than being ready for another install attempt of openshift. Optionally, you can choose to uninstall docker in a heavy handed manner. Docker storage setup is a prerequisite before running any playbooks, thus we don't know what the end user's docker configuration is because we didn't set it up. These playbooks are not meant for a production cluster. You can use them for POCs and iterating on a test set of machines if you are trying to get your initial install process figured out. I ran into this exact scenario. The following workaround worked for me: "yum -y reinstall docker-common", which reinstalled the /etc/sysconfig/docker file. PR Created in 3.11: https://github.com/openshift/openshift-ansible/pull/10810 In openshift-ansible-3.11.60-1 and later. 1. Reproduce the failure # rpm -qa |grep openshift-ansible openshift-ansible-3.11.56-1.git.0.59f0535.el7.noarch.rpm After the finished playbook uninstall.yml with enabled openshift_uninstall_docker, I check with command: # rpm -qa |grep docker docker-1.13.1-88.git07f3374.el7.x86_64 Re-install OCP with commands: # ansible-playbook -i inventory.yml playbooks/prerequisites.yml And get error messages from docker service: # systemctl status docker Jan 25 01:51:30 xxx dockerd-current[69936]: /usr/bin/docker-containerd: line 2: /etc/sysconfig/docker: No such file or directory 2. Verify the Patch Check in the uninstalled enviroment # rpm -qa |grep openshift-ansible openshift-ansible-3.11.73-1.git.0.89d3763.el7.noarch.rpm Re-install OCP and get the same error: # systemctl status docker Jan 25 03:23:35 xxx dockerd-current[64749]: /usr/bin/docker-containerd: line 2: /etc/sysconfig/docker: No such file or directory # rpm -qf /etc/sysconfig/docker docker-common-1.13.1-88.git07f3374.el7.x86_64 3. Result: Failed, because installer need to uninstall the rpm of docker-common Need to update the task referenced in https://github.com/openshift/openshift-ansible/pull/10810 to consider all installed docker packages, for example docker-common. Verified with: openshift-ansible-playbooks-3.11.82-1.git.0.f29227a.el7.noarch.rpm Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2019:0326 |