Bug 1873275
Summary: | kubectl example for oc new-project results in a pod stuck in ImagePullBackoff | ||
---|---|---|---|
Product: | OpenShift Container Platform | Reporter: | Mike Fiedler <mifiedle> |
Component: | oc | Assignee: | Maciej Szulik <maszulik> |
Status: | CLOSED ERRATA | QA Contact: | zhou ying <yinzhou> |
Severity: | medium | Docs Contact: | |
Priority: | medium | ||
Version: | 4.6 | CC: | aos-bugs, eminguez, jokerman, mfojtik |
Target Milestone: | --- | ||
Target Release: | 4.6.0 | ||
Hardware: | Unspecified | ||
OS: | Unspecified | ||
Whiteboard: | |||
Fixed In Version: | Doc Type: | If docs needed, set a value | |
Doc Text: | Story Points: | --- | |
Clone Of: | Environment: | ||
Last Closed: | 2020-10-27 16:35:33 UTC | Type: | Bug |
Regression: | --- | Mount Type: | --- |
Documentation: | --- | CRM: | |
Verified Versions: | Category: | --- | |
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
Cloudforms Team: | --- | Target Upstream Version: | |
Embargoed: |
Description
Mike Fiedler
2020-08-27 18:25:49 UTC
Confirmed with latest version , Image pull issue has fixed , but the pod still couldn't running: [root@dhcp-140-138 ~]# ./kubectl create deployment hello-node --image=k8s.gcr.io/echoserver:1.4 deployment.apps/hello-node created [root@dhcp-140-138 ~]# ./kubectl get po NAME READY STATUS RESTARTS AGE hello-node-7567d9fdc9-752dz 0/1 CrashLoopBackOff 3 86s [root@dhcp-140-138 ~]# ./kubectl describe po/hello-node-7567d9fdc9-752dz Name: hello-node-7567d9fdc9-752dz Namespace: zhouy Priority: 0 Node: ip-10-0-187-49.ap-northeast-2.compute.internal/10.0.187.49 Start Time: Tue, 01 Sep 2020 13:41:03 +0800 Labels: app=hello-node pod-template-hash=7567d9fdc9 Annotations: k8s.v1.cni.cncf.io/network-status: [{ "name": "", "interface": "eth0", "ips": [ "10.129.2.54" ], "default": true, "dns": {} }] k8s.v1.cni.cncf.io/networks-status: [{ "name": "", "interface": "eth0", "ips": [ "10.129.2.54" ], "default": true, "dns": {} }] openshift.io/scc: restricted Status: Running IP: 10.129.2.54 IPs: IP: 10.129.2.54 Controlled By: ReplicaSet/hello-node-7567d9fdc9 Containers: echoserver: Container ID: cri-o://71622c4af6bf6e47d566a69788a4621d424359077b9fe9f525e180ab80bb0c6a Image: k8s.gcr.io/echoserver:1.4 Image ID: k8s.gcr.io/echoserver@sha256:5d99aa1120524c801bc8c1a7077e8f5ec122ba16b6dda1a5d3826057f67b9bcb Port: <none> Host Port: <none> State: Waiting Reason: CrashLoopBackOff Last State: Terminated Reason: Error Exit Code: 1 Started: Tue, 01 Sep 2020 13:41:28 +0800 Finished: Tue, 01 Sep 2020 13:41:28 +0800 Ready: False Restart Count: 2 Environment: <none> Mounts: /var/run/secrets/kubernetes.io/serviceaccount from default-token-tgtfb (ro) Conditions: Type Status Initialized True Ready False ContainersReady False PodScheduled True Volumes: default-token-tgtfb: Type: Secret (a volume populated by a Secret) SecretName: default-token-tgtfb Optional: false QoS Class: BestEffort Node-Selectors: <none> Tolerations: node.kubernetes.io/not-ready:NoExecute for 300s node.kubernetes.io/unreachable:NoExecute for 300s Events: Type Reason Age From Message ---- ------ ---- ---- ------- Normal Scheduled <unknown> Successfully assigned zhouy/hello-node-7567d9fdc9-752dz to ip-10-0-187-49.ap-northeast-2.compute.internal Normal AddedInterface 42s multus Add eth0 [10.129.2.54/23] Normal Pulling 41s kubelet, ip-10-0-187-49.ap-northeast-2.compute.internal Pulling image "k8s.gcr.io/echoserver:1.4" Normal Pulled 33s kubelet, ip-10-0-187-49.ap-northeast-2.compute.internal Successfully pulled image "k8s.gcr.io/echoserver:1.4" in 8.041931132s Normal Created 18s (x3 over 33s) kubelet, ip-10-0-187-49.ap-northeast-2.compute.internal Created container echoserver Normal Started 18s (x3 over 33s) kubelet, ip-10-0-187-49.ap-northeast-2.compute.internal Started container echoserver Normal Pulled 18s (x2 over 33s) kubelet, ip-10-0-187-49.ap-northeast-2.compute.internal Container image "k8s.gcr.io/echoserver:1.4" already present on machine Warning BackOff 4s (x4 over 32s) kubelet, ip-10-0-187-49.ap-northeast-2.compute.internal Back-off restarting failed container [root@dhcp-140-138 ~]# ./kubectl logs -f po/hello-node-7567d9fdc9-752dz 2020/09/01 05:42:39 [emerg] 1#1: mkdir() "/var/lib/nginx/proxy" failed (13: Permission denied) nginx: [emerg] mkdir() "/var/lib/nginx/proxy" failed (13: Permission denied) The issue has fixed: [zhouying@dhcp-140-138 ~]$ oc new-project zhouyt1 Now using project "zhouyt1" on server "https://api.zy09096.qe.devcluster.openshift.com:6443". You can add applications to this project with the 'new-app' command. For example, try: oc new-app rails-postgresql-example to build a new example application in Ruby. Or use kubectl to deploy a simple Kubernetes application: kubectl create deployment hello-node --image=k8s.gcr.io/serve_hostname [zhouying@dhcp-140-138 ~]$ kubectl create deployment hello-node --image=k8s.gcr.io/serve_hostname deployment.apps/hello-node created [zhouying@dhcp-140-138 ~]$ oc get po NAME READY STATUS RESTARTS AGE hello-node-6d9d95bb88-4wmp5 1/1 Running 0 11s *** Bug 1879051 has been marked as a duplicate of this bug. *** Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory (OpenShift Container Platform 4.6 GA Images), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHBA-2020:4196 |