+++ This bug was initially created as a clone of Bug #1781333 +++ There'a an assumption in the installer that "baremetal" and "provisioning" are the names of the two bridges for the external and provisinoing networks, respectively. These are hardcoded in the terraform variables despite them being tunable in the platform config.
Ah sorry, this commit is in release-4.3 branch. GitHub doesn't show it being tagged in that branch...I'm not sure why it detected that this change wasn't already there. https://github.com/openshift/installer/commit/1c8c84fb1de45ea3c4a59258b622a99459505db1
Hi, I've Provisioned build [1] via Jenkins. After that i've changed baremetal network to blabaremetal & adjusted all the machines accordingly. From that point on - I've continued with the manual steps [2]. After cloning [3], I've changed every relevant place from baremetal to blabaremetal, Once I got to 06_create_cluster.sh the script failed on [4]. Since the fix is for this specific var installConfig.Config.Platform.BareMetal.ExternalBridge, I can only assume it is still hard coded. Nataf [1] registry.svc.ci.openshift.org/ocp/release:4.3.0-0.nightly-2019-12-29-173422 [2] https://gitlab.cee.redhat.com/ocp-edge-qe/ocp-edge-demo/blob/master/docs/manual-steps.md [3] https://github.com/openshift-metal3/dev-scripts.git [4]i\f\t\.\c\o\m* ]] + '[' n == y ']' ++ dig +noall +answer api.ostest.test.metalkube.org ++ awk '{print $NF}' + API_VIP= ++ dig +noall +answer test.apps.ostest.test.metalkube.org ++ awk '{print $NF}' + INGRESS_VIP= + '[' '!' -f ocp/install-config.yaml ']' + create_cluster ocp + local assets_dir + assets_dir=ocp + export TF_LOG=DEBUG + TF_LOG=DEBUG + cp ocp/install-config.yaml ocp/install-config.yaml.tmp + ocp/openshift-baremetal-install --dir ocp --log-level=debug create manifests level=debug msg="OpenShift Installer unreleased-master-2201-g93c78d09ed9e2badb4bf5dab708152fe6b3b6602-dirty" level=debug msg="Built from commit 93c78d09ed9e2badb4bf5dab708152fe6b3b6602" level=debug msg="Fetching Master Machines..." level=debug msg="Loading Master Machines..." level=debug msg=" Loading Cluster ID..." level=debug msg=" Loading Install Config..." level=debug msg=" Loading SSH Key..." level=debug msg=" Loading Base Domain..." level=debug msg=" Loading Platform..." level=debug msg=" Loading Cluster Name..." level=debug msg=" Loading Base Domain..." level=debug msg=" Loading Platform..." level=debug msg=" Loading Pull Secret..." level=debug msg=" Loading Platform..." level=fatal msg="failed to fetch Master Machines: failed to load asset \"Install Config\": invalid \"install-config.yaml\" file: [platform.baremetal.apiVIP: Invalid value: \"DNS lookup failure: lookup api.ostest.test.metalkube.org on 192.168.123.1:53: no such host\": \"DNS lookup failure: lookup api.ostest.test.metalkube.org on 192.168.123.1:53: no such host\" is not a valid IP, platform.baremetal.apiVIP: Invalid value: \"DNS lookup failure: lookup api.ostest.test.metalkube.org on 192.168.123.1:53: no such host\": the virtual IP is expected to be in 192.168.123.0/24 subnet, platform.baremetal.ingressVIP: Invalid value: \"DNS lookup failure: lookup test.apps.ostest.test.metalkube.org on 192.168.123.1:53: no such host\": \"DNS lookup failure: lookup test.apps.ostest.test.metalkube.org on 192.168.123.1:53: no such host\" is not a valid IP, platform.baremetal.ingressVIP: Invalid value: \"DNS lookup failure: lookup test.apps.ostest.test.metalkube.org on 192.168.123.1:53: no such host\": the virtual IP is expected to be in 192.168.123.0/24 subnet, platform.baremetal.externalBridge: Invalid value: \"baremetal\": could not find interface \"baremetal\", valid interfaces are blabaremetal, lo, provisioning]"
This has been fixed and we are aware of people including those in our team who are using this successfully. Moving back to ON_QA.