Bug 1750269
| Summary: | vSphere UPI: failed to initialize the cluster: Some cluster operators are still updating: authentication, console | ||
|---|---|---|---|
| Product: | OpenShift Container Platform | Reporter: | Vijay Avuthu <vavuthu> |
| Component: | Networking | Assignee: | Casey Callendrello <cdc> |
| Networking sub component: | openshift-sdn | QA Contact: | zhaozhanqi <zzhao> |
| Status: | CLOSED WORKSFORME | Docs Contact: | |
| Severity: | high | ||
| Priority: | unspecified | CC: | aos-bugs, lsm5, mfojtik, ratamir, rphillips, sbatsche, scuppett, sttts |
| Version: | 4.2.0 | ||
| Target Milestone: | --- | ||
| Target Release: | 4.2.0 | ||
| Hardware: | Unspecified | ||
| OS: | Unspecified | ||
| Whiteboard: | |||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |
| Doc Text: | Story Points: | --- | |
| Clone Of: | Environment: | ||
| Last Closed: | 2019-09-11 00:20:17 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Vijay Avuthu
2019-09-09 08:00:58 UTC
It seems that kube-apiserver pod fails on being created because networking is not available at one of the nodes, here's the reported error:
```
'Failed create pod sandbox: rpc error: code = Unknown desc = failed to
create pod network sandbox k8s_installer-5-control-plane-1_openshift-kube-apiserver_900db7f3-d2ce-11e9-8fc8-005056be0641_0(6eb4a350ef81b980482f853dc2585bcac49e5b395ab03fb7472c0736833d91e3):
Multus: Err adding pod to network "openshift-sdn": Multus: error in invoke Delegate
add - "openshift-sdn": failed to send CNI request: Post http://dummy/: dial unix
/var/run/openshift-sdn/cniserver/socket: connect: connection refused'
```
sdn-7llq6 is failing with "rm: cannot remove '/etc/cni/net.d/80-openshift-network.conf': Permission denied" How!?!? Is this cluster still up? Can we get the node journal? (In reply to Casey Callendrello from comment #4) > Is this cluster still up? Can we get the node journal? Cluster is not available. It was removed after collecting all logs. Unfortunately must-gather doesn't actually gather everything we need to debug this. Please try and reproduce, and keep the cluster up. I suspect this is a selinux issue. Running `ls -Z /etc/cni/net.d/80-openshift-network.conf` on all the nodes would tell us if different selinux permissions are being used. > I have tried 3 times to reproduce the issue but it was successful every time. Below are builds used for installation
1st attempt: 4.2.0-0.ci-2019-09-10-121820
2nd attempt : 4.2.0-0.ci-2019-09-09-021340 ( same build where I faced issue previously )
3rd attempt: 4.2.0-0.ci-2019-09-09-021340
>
$ date;oc --kubeconfig /home/vavuthu/VJ/installations/clusterdirs/qe1/auth/kubeconfig get ClusterOperator
Wed Sep 11 00:27:32 IST 2019
NAME VERSION AVAILABLE PROGRESSING DEGRADED SINCE
authentication 4.2.0-0.ci-2019-09-09-021340 True False False 3m4s
cloud-credential 4.2.0-0.ci-2019-09-09-021340 True False False 40m
cluster-autoscaler 4.2.0-0.ci-2019-09-09-021340 True False False 12m
console 4.2.0-0.ci-2019-09-09-021340 True False False 6m38s
dns 4.2.0-0.ci-2019-09-09-021340 True False False 28m
image-registry 4.2.0-0.ci-2019-09-09-021340 True False False 17m
>
|