Note: This bug is displayed in read-only format because the product is no longer active in Red Hat Bugzilla.

Bug 1861954

Summary: Openshift install failed to create machine: unable to get resource pool for <nil>
Product: OpenShift Container Platform Reporter: Luciano R <lscorsin>
Component: InstallerAssignee: aos-install
Installer sub component: openshift-installer QA Contact: Gaoyun Pei <gpei>
Status: CLOSED DUPLICATE Docs Contact:
Severity: high    
Priority: unspecified CC: adahiya
Version: 4.5   
Target Milestone: ---   
Target Release: ---   
Hardware: x86_64   
OS: Unspecified   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2020-07-30 03:04:38 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Description Luciano R 2020-07-30 02:24:12 UTC
Description of problem:
Vsphere IPI installation fail with some operators stuck. The control-plane is up and the following status is being freezed with Cluster Operators

NAME                                       VERSION   AVAILABLE   PROGRESSING   DEGRADED   SINCE
authentication                                       Unknown     Unknown       True       67m
cloud-credential                           4.5.4     True        False         False      77m
cluster-autoscaler                         4.5.4     True        False         False      59m
config-operator                            4.5.4     True        False         False      60m
console                                    4.5.4     Unknown     True          False      60m
csi-snapshot-controller
dns                                        4.5.4     True        False         False      65m
etcd                                       4.5.4     True        False         False      64m
image-registry                             4.5.4     True        False         False      60m
ingress                                              False       True          True       61m
insights                                   4.5.4     True        False         False      60m
kube-apiserver                             4.5.4     True        False         False      62m
kube-controller-manager                    4.5.4     True        False         False      65m
kube-scheduler                             4.5.4     True        False         False      65m
kube-storage-version-migrator              4.5.4     False       False         False      67m
machine-api                                4.5.4     True        False         False      57m
machine-approver                           4.5.4     True        False         False      63m
machine-config                             4.5.4     True        False         False      58m
marketplace                                4.5.4     True        False         False      60m
monitoring                                           False       True          True       55m
network                                    4.5.4     True        False         False      67m
node-tuning                                4.5.4     True        False         False      67m
openshift-apiserver                        4.5.4     True        False         False      61m
openshift-controller-manager               4.5.4     True        False         False      60m
openshift-samples                          4.5.4     True        False         False      60m
operator-lifecycle-manager                 4.5.4     True        False         False      66m
operator-lifecycle-manager-catalog         4.5.4     True        False         False      66m
operator-lifecycle-manager-packageserver   4.5.4     True        False         False      61m
service-ca                                 4.5.4     True        False         False      67m
storage                                    4.5.4     True        False         False      60m

When checking the logs of machine-api-controllers we noticed the following lines:
W0730 02:16:15.994460       1 controller.go:315] prd-xgtmg-worker-v8qzn: failed to create machine: unable to get resource pool for <nil>: default resource pool resolves to multiple instances, please specify

How reproducible:
Every time.

Steps to Reproduce:
1. Just install the cluster using install-config generated by installer (we are using installer v 4.5.4)

Actual results:
The installation fails. Only the master nodes are up. Worker nodes not being created.


Expected results:
Installation finished or at least all cluster operators running.

Comment 1 Abhinav Dahiya 2020-07-30 03:04:38 UTC

*** This bug has been marked as a duplicate of bug 1852545 ***