Created attachment 1481221 [details] oc get csr -oyaml Description of problem: I open it following https://bugzilla.redhat.com/show_bug.cgi?id=1623204#c8. Installer fails on task [Approve node certificates when bootstrapping] with error "Could not find csr for nodes". Version-Release number of the following components: both with openshift-ansible-3.10.41-1.git.0.fd15dd7.el7.noarch (rpm) and openshift-ansible-openshift-ansible-3.10.43-1 (git release) ansible-2.6.3-1.el7.noarch How reproducible: always Steps to Reproduce: 3 masters + 3 infra + 1 node + 1 lb ansible-playbook -i /etc/ansible/hosts playbooks/deploy_cluster.yml Actual results: PLAY [Approve any pending CSR requests from inventory nodes] ****************************************************************************************************************************** TASK [Dump all candidate bootstrap hostnames] ********************************************************************************************************************************************* Thursday 06 September 2018 09:55:21 +0200 (0:00:00.398) 0:27:56.128 **** ok: [red-ose-master01.internal.ose.extrasys.it] => { "msg": [ "red-ose-master01.internal.ose.extrasys.it", "red-ose-master02.internal.ose.extrasys.it", "red-ose-master03.internal.ose.extrasys.it", "red-ose-node01.internal.ose.extrasys.it", "red-ose-infra01.internal.ose.extrasys.it", "red-ose-infra02.internal.ose.extrasys.it", "red-ose-infra03.internal.ose.extrasys.it" ] } TASK [Find all hostnames for bootstrapping] *********************************************************************************************************************************************** Thursday 06 September 2018 09:55:21 +0200 (0:00:00.084) 0:27:56.212 **** ok: [red-ose-master01.internal.ose.extrasys.it] TASK [Dump the bootstrap hostnames] ******************************************************************************************************************************************************* Thursday 06 September 2018 09:55:22 +0200 (0:00:00.240) 0:27:56.453 **** ok: [red-ose-master01.internal.ose.extrasys.it] => { "msg": [ "red-ose-master01", "red-ose-master02", "red-ose-master03", "red-ose-node01", "red-ose-infra01", "red-ose-infra02", "red-ose-infra03" ] } TASK [Approve node certificates when bootstrapping] *************************************************************************************************************************************** Thursday 06 September 2018 09:55:22 +0200 (0:00:00.085) 0:27:56.539 **** FAILED - RETRYING: Approve node certificates when bootstrapping (30 retries left). FAILED - RETRYING: Approve node certificates when bootstrapping (29 retries left). FAILED - RETRYING: Approve node certificates when bootstrapping (1 retries left). fatal: [red-ose-master01.internal.ose.extrasys.it]: FAILED! => {"attempts": 30, "changed": false, "msg": "Cound not find csr for nodes: red-ose-master01, red-ose-master02, red-ose-master0 3", "state": "unknown"} PLAY RECAP ******************************************************************************************************************************************************************************** localhost : ok=13 changed=0 unreachable=0 failed=0 red-ose-infra01.internal.ose.extrasys.it : ok=119 changed=65 unreachable=0 failed=0 red-ose-infra02.internal.ose.extrasys.it : ok=119 changed=65 unreachable=0 failed=0 red-ose-infra03.internal.ose.extrasys.it : ok=119 changed=65 unreachable=0 failed=0 red-ose-int-haproxy01.internal.ose.extrasys.it : ok=33 changed=4 unreachable=0 failed=0 red-ose-master01.internal.ose.extrasys.it : ok=432 changed=205 unreachable=0 failed=1 red-ose-master02.internal.ose.extrasys.it : ok=293 changed=147 unreachable=0 failed=0 red-ose-master03.internal.ose.extrasys.it : ok=293 changed=147 unreachable=0 failed=0 red-ose-node01.internal.ose.extrasys.it : ok=119 changed=65 unreachable=0 failed=0 INSTALLER STATUS ************************************************************************************************************************************************************************** Initialization : Complete (0:00:49) Health Check : Complete (0:03:02) Node Bootstrap Preparation : Complete (0:14:59) etcd Install : Complete (0:01:00) Load Balancer Install : Complete (0:00:16) Master Install : Complete (0:06:07) Master Additional Install : Complete (0:01:32) Node Join : In Progress (0:02:55) Failure summary: 1. Hosts: red-ose-master01.internal.ose.extrasys.it Play: Approve any pending CSR requests from inventory nodes Task: Approve node certificates when bootstrapping Message: Cound not find csr for nodes: red-ose-master01, red-ose-master02, red-ose-master03 Expected results: Installer completes successfully. Additional info: in attachment I suspect this could be also related with the generated name in csr and how this should match with node hostnames and DNS entries. - On DNS I have nodename.internal.ose.extrasys.it entries. All nodes involved correctly resolve each other hostname (both fqdn and short) - hostname -f on nodes gives "nodename"(e.g.red-ose-master01). Should it be nodename.internal.ose.extrasys.it? - in /etc/hosts I have the entry "IP nodename" (with no fqdn) Is that configuration right?
Created attachment 1481222 [details] inventory file
Created attachment 1481223 [details] oc get nodes
Created attachment 1481224 [details] oc get csr
I am also experiencing this even when falling back to a single node setup. As this is targeted to be fixed for 3.11, any chance that I can bring up a cluster with 3.10?
Latest code in master and 3.10 branches has added additional debug output. Please recreate problem using ansible-playbook -vvv (3 v's, not 2) and attach output.
Added customer ticket to BZ, as they are experiencing this same issue, as well. Please let me know if any attachments or further information is needed.
*** Bug 1627623 has been marked as a duplicate of this bug. ***
I had the same problem and this commit from Michael Gugino fixed it: https://github.com/openshift/openshift-ansible/pull/10033 thanks (mgugino)
The comments on this bug contain two unique problems. The original reporter is observing difference in the nodenames and the bootstrap names. Latter comments, some of which are private, are more similar to https://bugzilla.redhat.com/show_bug.cgi?id=1625873
(In reply to Michael Gugino from comment #6) > Latest code in master and 3.10 branches has added additional debug output. > > Please recreate problem using ansible-playbook -vvv (3 v's, not 2) and > attach output. I recreated the problem with openshift-ansible-openshift-ansible-3.10.47-1 git release with 1 master + 1 infra + 1 node. Please see attached the -vvv execution and failure. The installation is performed on Rhel images somehow customized with cloud-init. I'll try on brand new qcow to avoid misconfiguration even in nodenames/hostnames/bootstrap names.
Created attachment 1483324 [details] ansible execution -vvv
(In reply to Serena Cortopassi from comment #18) > (In reply to Michael Gugino from comment #6) > > Latest code in master and 3.10 branches has added additional debug output. > > > > Please recreate problem using ansible-playbook -vvv (3 v's, not 2) and > > attach output. > > I recreated the problem with openshift-ansible-openshift-ansible-3.10.47-1 > git release with 1 master + 1 infra + 1 node. > > Please see attached the -vvv execution and failure. > > The installation is performed on Rhel images somehow customized with > cloud-init. I'll try on brand new qcow to avoid misconfiguration even in > nodenames/hostnames/bootstrap names. This run looks like it's the issue that master is not ready yet: runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:docker: network plugin is not ready: cni config uninitialized We have another patch out for this problem, already shipped in master, waiting on QE to verify before we backport to 3.10.
(In reply to Michael Gugino from comment #20) > This run looks like it's the issue that master is not ready yet: runtime > network not ready: NetworkReady=false reason:NetworkPluginNotReady > message:docker: network plugin is not ready: cni config uninitialized > > We have another patch out for this problem, already shipped in master, > waiting on QE to verify before we backport to 3.10. I can confirm what you say: [awx@red-ose-master01 ~]$ oc get nodes NAME STATUS ROLES AGE VERSION red-ose-master01 NotReady master 2h v1.10.0+b81c8f8 [awx@red-ose-master01 ~]$ sudo journalctl -xe -u atomic-openshift-node ... Unable to update cni config: No networks found in /etc/cni/net.d ... The folder /etc/cni/net.d is empty indeed. This behaviour does not happen to everyone, I am on RH OpenStack 10 and I see a lot of people are affected in that case. Do you know if it is something related to nodenames or OS version,.., or what? Could you please link the patch shipped? Thanks
(In reply to Serena Cortopassi from comment #21) > (In reply to Michael Gugino from comment #20) > > > This run looks like it's the issue that master is not ready yet: runtime > > network not ready: NetworkReady=false reason:NetworkPluginNotReady > > message:docker: network plugin is not ready: cni config uninitialized > > > > We have another patch out for this problem, already shipped in master, > > waiting on QE to verify before we backport to 3.10. > > I can confirm what you say: > [awx@red-ose-master01 ~]$ oc get nodes > NAME STATUS ROLES AGE VERSION > red-ose-master01 NotReady master 2h v1.10.0+b81c8f8 > > [awx@red-ose-master01 ~]$ sudo journalctl -xe -u atomic-openshift-node > ... > Unable to update cni config: No networks found in /etc/cni/net.d > ... > > The folder /etc/cni/net.d is empty indeed. > > This behaviour does not happen to everyone, I am on RH OpenStack 10 and I > see a lot of people are affected in that case. > Do you know if it is something related to nodenames or OS version,.., or > what? > > Could you please link the patch shipped? > > Thanks Here is the patch for 3.10, not shipped yet. https://github.com/openshift/openshift-ansible/pull/10055 That patch links to the original in master/3.11.
The comments on this bug contain more than one issues, just like analysis in comment 17. Here only provide one fix PR in comment 22, it is similar to https://bugzilla.redhat.com/show_bug.cgi?id=1628964, QE already reproduced it and run verification for the fix PR. More detailed refer to https://bugzilla.redhat.com/show_bug.cgi?id=1628964#c9. If the PR do not fix your issue, suggest to open a separate bug for tracking.
Is there a workaround for this until rpm is released?
This same issue (empty /etc/cni/net.d ) also causing issues at "openshift_web_console : Apply the web console template file]" TASK. I have upgraded to "openshift-ansible-playbooks-3.10.51-1.git.0.44a646c.el7.noarch" which was released today. We pass "Could not find csr for nodes - Bootstrapping task" TASK but now: =============================================================================== TASK [openshift_web_console : Apply the web console template file] ************* changed: [master.os.serra.local] TASK [openshift_web_console : Remove temp directory] *************************** ok: [master.os.serra.local] TASK [openshift_web_console : Pause for the web console deployment to start] *** skipping: [master.os.serra.local] TASK [openshift_web_console : include_tasks] *********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_web_console/tasks/start.yml for master.os.serra.local TASK [openshift_web_console : Verify that the console is running] ************** FAILED - RETRYING: Verify that the console is running (60 retries left). FAILED - RETRYING: Verify that the console is running (59 retries left). FAILED - RETRYING: Verify that the console is running (58 retries left). FAILED - RETRYING: Verify that the console is running (57 retries left). FAILED - RETRYING: Verify that the console is running (56 retries left). FAILED - RETRYING: Verify that the console is running (55 retries left). FAILED - RETRYING: Verify that the console is running (54 retries left). FAILED - RETRYING: Verify that the console is running (53 retries left). FAILED - RETRYING: Verify that the console is running (52 retries left). FAILED - RETRYING: Verify that the console is running (51 retries left). FAILED - RETRYING: Verify that the console is running (50 retries left). FAILED - RETRYING: Verify that the console is running (49 retries left). FAILED - RETRYING: Verify that the console is running (48 retries left). FAILED - RETRYING: Verify that the console is running (47 retries left). FAILED - RETRYING: Verify that the console is running (46 retries left). FAILED - RETRYING: Verify that the console is running (45 retries left). FAILED - RETRYING: Verify that the console is running (44 retries left). FAILED - RETRYING: Verify that the console is running (43 retries left). FAILED - RETRYING: Verify that the console is running (42 retries left). FAILED - RETRYING: Verify that the console is running (41 retries left). FAILED - RETRYING: Verify that the console is running (40 retries left). FAILED - RETRYING: Verify that the console is running (39 retries left). FAILED - RETRYING: Verify that the console is running (38 retries left). FAILED - RETRYING: Verify that the console is running (37 retries left). FAILED - RETRYING: Verify that the console is running (36 retries left). FAILED - RETRYING: Verify that the console is running (35 retries left). FAILED - RETRYING: Verify that the console is running (34 retries left). FAILED - RETRYING: Verify that the console is running (33 retries left). FAILED - RETRYING: Verify that the console is running (32 retries left). FAILED - RETRYING: Verify that the console is running (31 retries left). FAILED - RETRYING: Verify that the console is running (30 retries left). FAILED - RETRYING: Verify that the console is running (29 retries left). FAILED - RETRYING: Verify that the console is running (28 retries left). FAILED - RETRYING: Verify that the console is running (27 retries left). FAILED - RETRYING: Verify that the console is running (26 retries left). FAILED - RETRYING: Verify that the console is running (25 retries left). FAILED - RETRYING: Verify that the console is running (24 retries left). FAILED - RETRYING: Verify that the console is running (23 retries left). FAILED - RETRYING: Verify that the console is running (22 retries left). FAILED - RETRYING: Verify that the console is running (21 retries left). FAILED - RETRYING: Verify that the console is running (20 retries left). FAILED - RETRYING: Verify that the console is running (19 retries left). FAILED - RETRYING: Verify that the console is running (18 retries left). FAILED - RETRYING: Verify that the console is running (17 retries left). FAILED - RETRYING: Verify that the console is running (16 retries left). FAILED - RETRYING: Verify that the console is running (15 retries left). FAILED - RETRYING: Verify that the console is running (14 retries left). FAILED - RETRYING: Verify that the console is running (13 retries left). FAILED - RETRYING: Verify that the console is running (12 retries left). FAILED - RETRYING: Verify that the console is running (11 retries left). FAILED - RETRYING: Verify that the console is running (10 retries left). FAILED - RETRYING: Verify that the console is running (9 retries left). FAILED - RETRYING: Verify that the console is running (8 retries left). FAILED - RETRYING: Verify that the console is running (7 retries left). FAILED - RETRYING: Verify that the console is running (6 retries left). FAILED - RETRYING: Verify that the console is running (5 retries left). FAILED - RETRYING: Verify that the console is running (4 retries left). FAILED - RETRYING: Verify that the console is running (3 retries left). FAILED - RETRYING: Verify that the console is running (2 retries left). FAILED - RETRYING: Verify that the console is running (1 retries left). fatal: [master.os.serra.local]: FAILED! => {"attempts": 60, "changed": false, "failed": true, "results": {"cmd": "/usr/bin/oc get deployment webconsole -o json -n openshift-web-console", "results": [{"apiVersion": "extensions/v1beta1", "kind": "Deployment", "metadata": {"annotations": {"deployment.kubernetes.io/revision": "1", "kubectl.kubernetes.io/last-applied-configuration": "{\"apiVersion\":\"apps/v1beta1\",\"kind\":\"Deployment\",\"metadata\":{\"annotations\":{},\"labels\":{\"app\":\"openshift-web-console\",\"webconsole\":\"true\"},\"name\":\"webconsole\",\"namespace\":\"openshift-web-console\"},\"spec\":{\"replicas\":1,\"strategy\":{\"rollingUpdate\":{\"maxUnavailable\":\"100%\"},\"type\":\"RollingUpdate\"},\"template\":{\"metadata\":{\"labels\":{\"app\":\"openshift-web-console\",\"webconsole\":\"true\"},\"name\":\"webconsole\"},\"spec\":{\"containers\":[{\"command\":[\"/usr/bin/origin-web-console\",\"--audit-log-path=-\",\"-v=0\",\"--config=/var/webconsole-config/webconsole-config.yaml\"],\"image\":\"docker.io/openshift/origin-web-console:v3.10.0\",\"imagePullPolicy\":\"IfNotPresent\",\"livenessProbe\":{\"exec\":{\"command\":[\"/bin/sh\",\"-c\",\"if [[ ! -f /tmp/webconsole-config.hash ]]; then \\\\\\n md5sum /var/webconsole-config/webconsole-config.yaml \\u003e /tmp/webconsole-config.hash; \\\\\\nelif [[ $(md5sum /var/webconsole-config/webconsole-config.yaml) != $(cat /tmp/webconsole-config.hash) ]]; then \\\\\\n echo 'webconsole-config.yaml has changed.'; \\\\\\n exit 1; \\\\\\nfi \\u0026\\u0026 curl -k -f https://0.0.0.0:8443/console/\"]}},\"name\":\"webconsole\",\"ports\":[{\"containerPort\":8443}],\"readinessProbe\":{\"httpGet\":{\"path\":\"/healthz\",\"port\":8443,\"scheme\":\"HTTPS\"}},\"resources\":{\"requests\":{\"cpu\":\"100m\",\"memory\":\"100Mi\"}},\"volumeMounts\":[{\"mountPath\":\"/var/serving-cert\",\"name\":\"serving-cert\"},{\"mountPath\":\"/var/webconsole-config\",\"name\":\"webconsole-config\"}]}],\"nodeSelector\":{\"node-role.kubernetes.io/master\":\"true\"},\"serviceAccountName\":\"webconsole\",\"volumes\":[{\"name\":\"serving-cert\",\"secret\":{\"defaultMode\":288,\"secretName\":\"webconsole-serving-cert\"}},{\"configMap\":{\"defaultMode\":288,\"name\":\"webconsole-config\"},\"name\":\"webconsole-config\"}]}}}}\n"}, "creationTimestamp": "2018-10-10T17:47:02Z", "generation": 1, "labels": {"app": "openshift-web-console", "webconsole": "true"}, "name": "webconsole", "namespace": "openshift-web-console", "resourceVersion": "5834", "selfLink": "/apis/extensions/v1beta1/namespaces/openshift-web-console/deployments/webconsole", "uid": "7e6039f6-ccb4-11e8-a3b9-525400380b0e"}, "spec": {"progressDeadlineSeconds": 600, "replicas": 1, "revisionHistoryLimit": 2, "selector": {"matchLabels": {"app": "openshift-web-console", "webconsole": "true"}}, "strategy": {"rollingUpdate": {"maxSurge": "25%", "maxUnavailable": "100%"}, "type": "RollingUpdate"}, "template": {"metadata": {"creationTimestamp": null, "labels": {"app": "openshift-web-console", "webconsole": "true"}, "name": "webconsole"}, "spec": {"containers": [{"command": ["/usr/bin/origin-web-console", "--audit-log-path=-", "-v=0", "--config=/var/webconsole-config/webconsole-config.yaml"], "image": "docker.io/openshift/origin-web-console:v3.10.0", "imagePullPolicy": "IfNotPresent", "livenessProbe": {"exec": {"command": ["/bin/sh", "-c", "if [[ ! -f /tmp/webconsole-config.hash ]]; then \\\n md5sum /var/webconsole-config/webconsole-config.yaml > /tmp/webconsole-config.hash; \\\nelif [[ $(md5sum /var/webconsole-config/webconsole-config.yaml) != $(cat /tmp/webconsole-config.hash) ]]; then \\\n echo 'webconsole-config.yaml has changed.'; \\\n exit 1; \\\nfi && curl -k -f https://0.0.0.0:8443/console/"]}, "failureThreshold": 3, "periodSeconds": 10, "successThreshold": 1, "timeoutSeconds": 1}, "name": "webconsole", "ports": [{"containerPort": 8443, "protocol": "TCP"}], "readinessProbe": {"failureThreshold": 3, "httpGet": {"path": "/healthz", "port": 8443, "scheme": "HTTPS"}, "periodSeconds": 10, "successThreshold": 1, "timeoutSeconds": 1}, "resources": {"requests": {"cpu": "100m", "memory": "100Mi"}}, "terminationMessagePath": "/dev/termination-log", "terminationMessagePolicy": "File", "volumeMounts": [{"mountPath": "/var/serving-cert", "name": "serving-cert"}, {"mountPath": "/var/webconsole-config", "name": "webconsole-config"}]}], "dnsPolicy": "ClusterFirst", "nodeSelector": {"node-role.kubernetes.io/master": "true"}, "restartPolicy": "Always", "schedulerName": "default-scheduler", "securityContext": {}, "serviceAccount": "webconsole", "serviceAccountName": "webconsole", "terminationGracePeriodSeconds": 30, "volumes": [{"name": "serving-cert", "secret": {"defaultMode": 288, "secretName": "webconsole-serving-cert"}}, {"configMap": {"defaultMode": 288, "name": "webconsole-config"}, "name": "webconsole-config"}]}}}, "status": {"conditions": [{"lastTransitionTime": "2018-10-10T17:47:02Z", "lastUpdateTime": "2018-10-10T17:47:02Z", "message": "Deployment has minimum availability.", "reason": "MinimumReplicasAvailable", "status": "True", "type": "Available"}, {"lastTransitionTime": "2018-10-10T17:57:03Z", "lastUpdateTime": "2018-10-10T17:57:03Z", "message": "ReplicaSet \"webconsole-55c4d867f\" has timed out progressing.", "reason": "ProgressDeadlineExceeded", "status": "False", "type": "Progressing"}], "observedGeneration": 1, "replicas": 1, "unavailableReplicas": 1, "updatedReplicas": 1}}], "returncode": 0}, "state": "list"} ...ignoring TASK [openshift_web_console : Check status in the openshift-web-console namespace] *** changed: [master.os.serra.local] TASK [openshift_web_console : debug] ******************************************* ok: [master.os.serra.local] => { "msg": [ "In project openshift-web-console on server https://master.os.serra.local:8443", "", "svc/webconsole - 172.30.50.83:443 -> 8443", " deployment/webconsole deploys docker.io/openshift/origin-web-console:v3.10.0", " deployment #1 running for 10 minutes - 0/1 pods", "", "View details with 'oc describe <resource>/<name>' or list everything with 'oc get all'." ] } TASK [openshift_web_console : Get pods in the openshift-web-console namespace] *** changed: [master.os.serra.local] TASK [openshift_web_console : debug] ******************************************* ok: [master.os.serra.local] => { "msg": [ "NAME READY STATUS RESTARTS AGE IP NODE", "webconsole-55c4d867f-t9qw6 0/1 Pending 0 10m <none> <none>" ] } TASK [openshift_web_console : Get events in the openshift-web-console namespace] *** changed: [master.os.serra.local] TASK [openshift_web_console : debug] ******************************************* ok: [master.os.serra.local] => { "msg": [ "LAST SEEN FIRST SEEN COUNT NAME KIND SUBOBJECT TYPE REASON SOURCE MESSAGE", "18s 10m 37 webconsole-55c4d867f-t9qw6.155c50721e9f61cd Pod Warning FailedScheduling default-scheduler 0/2 nodes are available: 2 node(s) were not ready.", "10m 10m 1 webconsole-55c4d867f.155c50721ea9bff6 ReplicaSet Normal SuccessfulCreate replicaset-controller Created pod: webconsole-55c4d867f-t9qw6", "10m 10m 1 webconsole.155c5071d5c31064 Deployment Normal ScalingReplicaSet deployment-controller Scaled up replica set webconsole-55c4d867f to 1" ] } TASK [openshift_web_console : Get console pod logs] **************************** changed: [master.os.serra.local] TASK [openshift_web_console : debug] ******************************************* ok: [master.os.serra.local] => { "msg": [] } TASK [openshift_web_console : Report console errors] *************************** fatal: [master.os.serra.local]: FAILED! => {"changed": false, "failed": true, "msg": "Console install failed."} to retry, use: --limit @/usr/share/ansible/openshift-ansible/playbooks/deploy_cluster.retry PLAY RECAP ********************************************************************* localhost : ok=13 changed=0 unreachable=0 failed=0 master.os.serra.local : ok=491 changed=126 unreachable=0 failed=1 node1.os.serra.local : ok=35 changed=2 unreachable=0 failed=0 INSTALLER STATUS *************************************************************** Initialization : Complete (0:00:25) Health Check : Complete (0:01:14) etcd Install : Complete (0:00:49) Node Bootstrap Preparation : Complete (0:00:00) Master Install : Complete (0:04:07) Master Additional Install : Complete (0:01:45) Node Join : Complete (0:00:14) Hosted Install : Complete (0:00:49) Web Console Install : In Progress (0:10:44) This phase can be restarted by running: playbooks/openshift-web-console/config.yml Failure summary: 1. Hosts: master.os.serra.local Play: Web Console Task: Report console errors Message: Console install failed. LATEST PACKAGE JUST INSTALLED: ============================= rpm -qi openshift-ansible-playbooks Name : openshift-ansible-playbooks Version : 3.10.51 Release : 1.git.0.44a646c.el7 Architecture: noarch Install Date: Wed 10 Oct 2018 06:36:43 PM BST Group : Unspecified Size : 442696 License : ASL 2.0 Signature : RSA/SHA1, Thu 27 Sep 2018 10:01:36 AM BST, Key ID c34c5bd42f297ecc Source RPM : openshift-ansible-3.10.51-1.git.0.44a646c.el7.src.rpm Build Date : Wed 26 Sep 2018 05:56:24 PM BST Build Host : c1be.rdu2.centos.org Relocations : (not relocatable) Packager : CBS <cbs> Vendor : CentOS URL : https://github.com/openshift/openshift-ansible Summary : Openshift and Atomic Enterprise Ansible Playbooks Description : Openshift and Atomic Enterprise Ansible Playbooks.
(In reply to Omer SEN from comment #39) > Is there a workaround for this until rpm is released? Omer, please file a new case or bugzilla for this issue.
(In reply to Michael Gugino from comment #41) > (In reply to Omer SEN from comment #39) > > Is there a workaround for this until rpm is released? > > Omer, please file a new case or bugzilla for this issue. https://bugzilla.redhat.com/show_bug.cgi?id=1638120 was filled for this new issue
You can work around this issue by manually setting the hostnames on each system to the FQDN. Ie, hostnamectl set-hostname your.full.hostname.com. Also, make sure to use the FQDNs in /etc/ansible/hosts. Take note, that if the hostname is being set from DHCP, you'll have to set PEERDNS=no on the interface to keep NetworkManager from reverting your changes. -Nick
Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2018:2709