Bug 1628075 - [free-int] invalid username/password when pull ose-pod image
Summary: [free-int] invalid username/password when pull ose-pod image
Keywords:
Status: NEW
Alias: None
Product: OpenShift Online
Classification: Red Hat
Component: Unknown
Version: 3.x
Hardware: Unspecified
OS: Unspecified
unspecified
high
Target Milestone: ---
: ---
Assignee: Abhishek Gupta
QA Contact: Junqi Zhao
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2018-09-12 07:40 UTC by Junqi Zhao
Modified: 2023-05-15 19:01 UTC (History)
1 user (show)

Fixed In Version:
Doc Type: If docs needed, set a value
Doc Text:
Clone Of:
Environment:
Last Closed:
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)

Description Junqi Zhao 2018-09-12 07:40:56 UTC
Description of problem:
free-int cluster, there are 5 node-exporter pods in ContainerCreating status
$ oc get pod -n openshift-monitoring | grep node-exporter
node-exporter-4dsj5                            2/2       Running             2          19d
node-exporter-7qjr2                            2/2       Running             0          4d
node-exporter-8jtk6                            2/2       Running             2          19d
node-exporter-8s78w                            2/2       Running             2          19d
node-exporter-czsrb                            2/2       Running             0          19d
node-exporter-dwj2n                            0/2       ContainerCreating   0          9h
node-exporter-fkskr                            2/2       Running             2          19d
node-exporter-h2q9q                            0/2       ContainerCreating   0          10h
node-exporter-h7ggd                            0/2       ContainerCreating   0          13h
node-exporter-j56ln                            2/2       Running             0          10h
node-exporter-jsxdj                            2/2       Running             2          18d
node-exporter-lvvdl                            2/2       Running             0          19d
node-exporter-lzwt2                            2/2       Running             0          1d
node-exporter-pmscl                            2/2       Running             0          19d
node-exporter-r5mbv                            0/2       ContainerCreating   0          10h
node-exporter-wh426                            0/2       ContainerCreating   0          9h
node-exporter-z8k4p                            2/2       Running             2          19d

describe one pod
$ oc describe pod node-exporter-dwj2n -n openshift-monitoring
Events:
  Type     Reason                  Age                 From                                    Message
  ----     ------                  ----                ----                                    -------
  Warning  FailedCreatePodSandBox  4m (x2604 over 9h)  kubelet, ip-172-31-63-238.ec2.internal  Failed create pod sandbox: rpc error: code = Unknown desc = error creating pod sandbox with name "k8s_node-exporter-dwj2n_openshift-monitoring_a73dae20-b609-11e8-a8f8-0ac586c2eb16_0": Error determining manifest MIME type for docker://registry.reg-aws.openshift.com:443/openshift3/ose-pod:v3.11: unable to retrieve auth token: invalid username/password

Version-Release number of selected component (if applicable):
openshift v3.11.0-0.21.0


How reproducible:
Always

Steps to Reproduce:
1. Check pods under openshift-monitoring
2.
3.

Actual results:
There are 5 node-exporter pods in ContainerCreating status

Expected results:
All node-exporter pods should be ready

Additional info:

Comment 1 Frederic Branczyk 2018-09-12 11:36:48 UTC
I'm seeing this for essentially every daemonset. See

kubectl get ev -n openshift-sdn

kubectl get ev -n openshift-node

Those yield the same errors/events.

Comment 2 Seth Jennings 2018-09-12 13:39:09 UTC
This is likely the result of the docker auth token for registry.reg-aws.openshift.com expiring.

There is a docker config.json local to each node that allows the kubelet to pull the sandbox image (ose-pod).  It seems that one a subset of these nodes, the token has expired.

This is an issue for any long lived cluster that 1) pulls ose-pod from a private registry and 2) registry auth uses tokens that expire

Comment 3 Doctor VWA 2018-11-04 10:34:55 UTC
Same problem here, there's a temporary solution?

In addcition the problem started when I was working on my project, everthing was going normal, so this problema appears. I tried to recreate the project, deleted absolutelly every single thing of the old one and recreated everything, but the bug still there.

Comment 4 Doctor VWA 2018-11-11 11:13:19 UTC
Issue got solved, looks like that was a temporary problem.


Note You need to log in before you can comment on or make changes to this bug.