Bug 1587889
| Summary: | The master-api restart continuously during logging redeploy testing | ||||||
|---|---|---|---|---|---|---|---|
| Product: | OpenShift Container Platform | Reporter: | Anping Li <anli> | ||||
| Component: | Master | Assignee: | Jordan Liggitt <jliggitt> | ||||
| Status: | CLOSED INSUFFICIENT_DATA | QA Contact: | Wang Haoran <haowang> | ||||
| Severity: | medium | Docs Contact: | |||||
| Priority: | medium | ||||||
| Version: | 3.10.0 | CC: | anli, aos-bugs, jokerman, mfojtik, mmccomas | ||||
| Target Milestone: | --- | ||||||
| Target Release: | 3.10.z | ||||||
| Hardware: | Unspecified | ||||||
| OS: | Unspecified | ||||||
| Whiteboard: | |||||||
| Fixed In Version: | Doc Type: | If docs needed, set a value | |||||
| Doc Text: | Story Points: | --- | |||||
| Clone Of: | Environment: | ||||||
| Last Closed: | 2018-06-19 07:35:42 UTC | Type: | Bug | ||||
| Regression: | --- | Mount Type: | --- | ||||
| Documentation: | --- | CRM: | |||||
| Verified Versions: | Category: | --- | |||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||
| Embargoed: | |||||||
| Attachments: |
|
||||||
|
Description
Anping Li
2018-06-06 09:11:05 UTC
From the logs it seems like the healthz panic is not what is causing the restarts as it eventually succeed: I0606 08:20:39.491762 1 wrap.go:42] GET /healthz: (202.482596ms) 500 github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f6d173e0d58, 0xc437741920, 0xc422814b00) /builddir/build/BUILD/atomic-openshift-git-0.0e11ff5/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/apiserver/pkg/server/healthz/healthz.go:130 +0x508 logging error output: "[+]ping ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/start-apiextensions-informers ok\n[+]poststarthook/start-apiextensions-controllers ok\n[+]poststarthook/project.openshift.io-projectcache ok\n[+]poststarthook/project.openshift.io-projectauthorizationcache ok\n[+]poststarthook/security.openshift.io-bootstrapscc ok\n[+]poststarthook/bootstrap-controller ok\n[+]poststarthook/ca-registration ok\n[+]poststarthook/start-kube-aggregator-informers ok\n[+]poststarthook/apiservice-registration-controller ok\n[+]poststarthook/apiservice-status-available-controller ok\n[+]poststarthook/apiservice-openapi-controller ok\n[+]poststarthook/kube-apiserver-autoregistration ok\n[+]autoregister-completion ok\n[-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld\n[+]poststarthook/authorization.openshift.io-ensureopenshift-infra ok\n[+]poststarthook/quota.openshift.io-clusterquotamapping ok\n[+]poststarthook/openshift.io-AdmissionInit ok\n[+]poststarthook/openshift.io-StartInformers ok\n[+]poststarthook/oauth.openshift.io-StartOAuthClientsBootstrapping ok\nhealthz check failed\n" I0606 08:20:45.627769 1 handler.go:159] kube-aggregator: GET "/healthz/ready" satisfied by nonGoRestful I0606 08:20:45.637807 1 pathrecorder.go:253] kube-aggregator: "/healthz/ready" satisfied by NotFoundHandler I0606 08:20:45.638027 1 handler.go:159] kube-apiserver: GET "/healthz/ready" satisfied by nonGoRestful I0606 08:20:45.638196 1 pathrecorder.go:253] kube-apiserver: "/healthz/ready" satisfied by NotFoundHandler I0606 08:20:45.638320 1 handler.go:159] openshift-apiserver: GET "/healthz/ready" satisfied by nonGoRestful I0606 08:20:45.638427 1 pathrecorder.go:240] openshift-apiserver: "/healthz/ready" satisfied by exact match I0606 08:20:45.638617 1 wrap.go:42] GET /healthz/ready: (11.083987ms) 200 [[kube-probe/1.10+] 10.240.0.23:41800] I0606 08:20:55.586263 1 handler.go:159] kube-aggregator: GET "/healthz/ready" satisfied by nonGoRestful I0606 08:20:55.586580 1 pathrecorder.go:253] kube-aggregator: "/healthz/ready" satisfied by NotFoundHandler I0606 08:20:55.586703 1 handler.go:159] kube-apiserver: GET "/healthz/ready" satisfied by nonGoRestful I0606 08:20:55.586844 1 pathrecorder.go:253] kube-apiserver: "/healthz/ready" satisfied by NotFoundHandler I0606 08:20:55.586963 1 handler.go:159] openshift-apiserver: GET "/healthz/ready" satisfied by nonGoRestful I0606 08:20:55.587082 1 pathrecorder.go:240] openshift-apiserver: "/healthz/ready" satisfied by exact match I0606 08:20:55.587225 1 wrap.go:42] GET /healthz/ready: (1.214201ms) 200 [[kube-probe/1.10+] 10.240.0.23:42206] I0606 08:21:05.583527 1 handler.go:159] kube-aggregator: GET "/healthz/ready" satisfied by nonGoRestful I0606 08:21:05.583691 1 pathrecorder.go:253] kube-aggregator: "/healthz/ready" satisfied by NotFoundHandler I0606 08:21:05.583822 1 handler.go:159] kube-apiserver: GET "/healthz/ready" satisfied by nonGoRestful I0606 08:21:05.583936 1 pathrecorder.go:253] kube-apiserver: "/healthz/ready" satisfied by NotFoundHandler I0606 08:21:05.584048 1 handler.go:159] openshift-apiserver: GET "/healthz/ready" satisfied by nonGoRestful I0606 08:21:05.584169 1 pathrecorder.go:240] openshift-apiserver: "/healthz/ready" satisfied by exact match I0606 08:21:05.584337 1 wrap.go:42] GET /healthz/ready: (1.369246ms) 200 [[kube-probe/1.10+] 10.240.0.23:42268] Can you get logs from the api server container that died? (docker logs). healthz is not panicking, the server just logs a stack trace on any 5xx error what does `oc get --raw /poststarthook/authorization.openshift.io-bootstrapclusterroles` return? The ENV have been deleted. The log is attached. Closing since it cannot be reproduced and we don't have any other info to go on. |