Bug 1505537
Summary: | Installer hangs at "Wait for master controller service to start on first master" | ||
---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | Vikas Laad <vlaad> |
Component: | Installer | Assignee: | Russell Teague <rteague> |
Status: | CLOSED ERRATA | QA Contact: | Vikas Laad <vlaad> |
Severity: | high | Docs Contact: | |
Priority: | high | ||
Version: | 3.7.0 | CC: | aos-bugs, jokerman, mmccomas, vlaad, xtian |
Target Milestone: | --- | ||
Target Release: | 3.7.0 | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | Bug Fix | |
Doc Text: |
A bug in Ansible was causing the pause module to hang when running playbooks in a background process. The pause was not necessary as the master controller service does not need to be stagger started. The tasks were refactored to remove the pause and start all master controller services at the same time.
|
Story Points: | --- |
Clone Of: | Environment: | ||
Last Closed: | 2017-11-28 22:18:47 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Vikas Laad
2017-10-23 19:54:57 UTC
root@ip-172-31-14-171: ~ # systemctl status atomic-openshift-master-controllers.service ● atomic-openshift-master-controllers.service Loaded: loaded (/etc/systemd/system/atomic-openshift-master-controllers.service; enabled; vendor preset: disabled) Active: active (running) since Mon 2017-10-23 18:11:02 UTC; 1h 47min ago Main PID: 32327 (runc) Memory: 3.4M CGroup: /system.slice/atomic-openshift-master-controllers.service └─32327 /bin/runc --systemd-cgroup run atomic-openshift-master-controllers Oct 23 19:56:31 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:56:31.237717 32340 reflector.go:343] github.com/openshift/origin/vendor/k8s.io/kubernetes/pkg/client/informers/informers_g...en compacted Oct 23 19:56:37 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:56:37.247487 32340 reflector.go:343] github.com/openshift/origin/vendor/k8s.io/kubernetes/pkg/client/informers/informers_g...en compacted Oct 23 19:56:37 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:56:37.247545 32340 reflector.go:343] github.com/openshift/origin/vendor/k8s.io/kubernetes/pkg/client/informers/informers_g...en compacted Oct 23 19:56:53 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:56:53.174642 32340 reflector.go:343] github.com/openshift/origin/vendor/k8s.io/kubernetes/pkg/client/informers/informers_g...en compacted Oct 23 19:57:06 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:57:06.195799 32340 reflector.go:343] github.com/openshift/origin/vendor/k8s.io/kubernetes/pkg/client/informers/informers_g...en compacted Oct 23 19:57:33 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:57:33.240900 32340 reflector.go:343] github.com/openshift/origin/pkg/build/generated/informers/internalversion/factory.go:...en compacted Oct 23 19:58:07 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:58:07.097670 32340 reflector.go:343] github.com/openshift/origin/vendor/k8s.io/kubernetes/pkg/client/informers/informers_g...en compacted Oct 23 19:58:07 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:58:07.097682 32340 reflector.go:343] github.com/openshift/origin/vendor/k8s.io/kubernetes/pkg/client/informers/informers_g...en compacted Oct 23 19:58:15 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:58:15.110434 32340 reflector.go:343] github.com/openshift/origin/vendor/k8s.io/kubernetes/pkg/client/informers/informers_g...en compacted Oct 23 19:58:34 ip-172-31-14-171.us-west-2.compute.internal atomic-openshift-master-controllers[32327]: W1023 19:58:34.141808 32340 reflector.go:343] github.com/openshift/origin/vendor/k8s.io/kubernetes/pkg/client/informers/informers_g...en compacted Hint: Some lines were ellipsized, use -l to show in full. The task you've referenced is an unconditional 15 second pause and I don't see from your logs that the run aborted in anyway. Did you terminate it before the 15 second pause period? Playbook is still stuck at that task after almost 2 hours root 12662 1497 0 18:55 pts/1 00:00:38 /usr/bin/python2 /usr/bin/ansible-playbook -i inv openshift-ansible/playbooks/byo/config.yml root 18567 12662 0 18:59 pts/1 00:00:00 /usr/bin/python2 /usr/bin/ansible-playbook -i inv openshift-ansible/playbooks/byo/config.yml this is current time on that machine root@ip-172-31-0-206: ~/openshift-ansible # date Mon Oct 23 20:47:31 UTC 2017 The attached logs appear to be truncated. This issue may be related to Ansible 2.4 usage of 'pause' in non-interactive shells. We are removing the pauses since they may no longer be necessary. The task noted above is the first instance of using 'pause' as recorded in the log so we may see this issue again on the next pause. Continuing to investigate and attempting to reproduce. (In reply to Russell Teague from comment #7) You are right, I was running the playbook in background ansible-playbook -i inv openshift-ansible/playbooks/byo/config.yml -vvv > ansible.log & Upstream Ansible issue: https://github.com/ansible/ansible/issues/32142 $ git tag --contains a07eba1e05cf3f52bf247afbec514d0af6629953 openshift-ansible-3.7.0-0.179.0 verified in following version, playbook completed without any problem when run in background. openshift v3.7.0-0.184.0 Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2017:3188 |